US20090225826A1 - Multi-View Video Coding Method and Device - Google Patents

Multi-View Video Coding Method and Device Download PDF

Info

Publication number
US20090225826A1
US20090225826A1 US12/224,816 US22481607A US2009225826A1 US 20090225826 A1 US20090225826 A1 US 20090225826A1 US 22481607 A US22481607 A US 22481607A US 2009225826 A1 US2009225826 A1 US 2009225826A1
Authority
US
United States
Prior art keywords
view
views
syntax element
specific information
syntax
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/224,816
Inventor
Purvin Bibhas Pandit
Yeping Su
Peng Yin
Cristina Gomila
Jill MacDonald Boyce
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US12/224,816 priority Critical patent/US20090225826A1/en
Assigned to THOMSON LICENSING reassignment THOMSON LICENSING ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: PANDIT, PURVIN BIBHAS, BOYCE, JILL MACDONALD, GOMILA, CRISTIAN, SU, YEPING, YIN, PENG
Publication of US20090225826A1 publication Critical patent/US20090225826A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/577Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/107Selection of coding mode or of prediction mode between spatial and temporal predictive coding, e.g. picture refresh
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

Definitions

  • the present principles relate generally to video encoding and decoding and, more particularly, to methods and apparatus for use in a multi-view video coding system.
  • Multi-view Video Coding MVC
  • ISO/IEC International Organization for Standardization/International Electrotechnical Commission
  • MPEG-4 MPEG-4 Part 10 Advanced Video Coding
  • ITU-T Telecommunication Sector
  • MPEG-4 AVC standard Multi-view Video Coding
  • a method has been proposed to enable efficient random access in multi-view compressed bit streams.
  • a V-picture type and a View Dependency Supplemental Enhancement Information (SEI) message are defined.
  • SEI View Dependency Supplemental Enhancement Information
  • a feature required in the proposed V-picture type is that V-pictures shall have no temporal dependence on other pictures in the same camera and may only be predicted from pictures in other cameras at the same time.
  • the proposed View Dependency Supplemental Enhancement Information message will describe exactly which views a V-picture, as well as the preceding and following sequences of pictures, may depend on. The following are the details of the proposed changes.
  • V-Picture syntax and semantics a particular syntax table relating to the MPEG-4 AVC standard is extended to include a Network Abstraction Layer (NAL) unit type of 14 corresponding to a V-picture. Also, the V-picture type is defined to have the following semantics:
  • NAL Network Abstraction Layer
  • V-picture A coded picture in which all slices reference only slices with the same temporal index (i.e., only slices in other views and not slices in the current view). When a V-picture would be output or displayed, it also causes the decoding process to mark all pictures from the same view which are not IDR-pictures or V-pictures and which precede the V-picture in output order to be marked as “unused for reference”. Each V-picture shall be associated with a View Dependency SEI message occurring in the same NAL.
  • a View Dependency Supplemental Enhancement Information message is defined with the following syntax:
  • view_dependency ( payloadSize ) ⁇ num_seq_reference_views ue(v) seq_reference_view_0 ue(v) seq_reference_view_1 ue(v) ... seq_reference_view_N ue(v) num_pic_reference_views ue(v) pic_reference_view_0 ue(v) pic_reference_view_1 ue(v) ...
  • num_seq_reference_views/num_pic_reference_views denotes the number of potential views that can be used as a reference for the current sequence/picture
  • seq_reference_view_i/pic_reference_view_i denotes the view number for the i th reference view.
  • the picture associated with a View Dependency Supplemental Enhancement Information message shall only reference the specified views described by pic_reference_view_i. Similarly, all subsequent pictures in output order of that view until the next View Dependency Supplemental Enhancement Information message in that view shall only reference the specified views described by seq_reference_view_i.
  • a View Dependency Supplemental Enhancement Information message shall be associated with each Instantaneous Decoding Refresh (IDR) picture and V-picture.
  • At least one drawback of this method is the complexity introduced in the decoder due to the dependency being recursively obtained. Additionally, this method requires that every V-picture carry an SEI message (which is a non-normative part of the MPEG-4 AVC standard), resulting in the dependency being unable to be used for normative behavior such as reference picture selection.
  • an apparatus includes an encoder for encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include view specific information.
  • the view specific information indicates a decoding interdependency between at least some of the at least two views.
  • the method includes encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include view specific information.
  • the view specific information indicates a decoding interdependency between at least some of the at least two views.
  • an apparatus includes a decoder for decoding at least two views corresponding to multi-view video content from a bitstream, wherein the bitstream is decoded to determine view specific information included therein, the view specific information indicating a decoding interdependency between at least some of the at least two views.
  • the method includes decoding at least two views corresponding to multi-view video content from a bitstream, wherein the bitstream is decoded to determine view specific information included therein.
  • the view specific information indicates a decoding interdependency between at least some of the at least two views.
  • an apparatus includes an encoder for encoding at least two views corresponding to multi-view video content by defining as a base view any of the at least two views that, for a decoding thereof, is independent of any other of the at least two views.
  • the method includes encoding at least two views corresponding to multi-view video content by defining as a base view any of the at least two views that, for a decoding thereof, is independent of any other of the at least two views.
  • an apparatus includes a decoder for decoding at least two views corresponding to multi-view video content, wherein the decoder determines which, if any, of the at least two views is a base view that, for a decoding thereof, is independent of any other of the at least two views.
  • the method includes decoding at least two views corresponding to multi-view video content, wherein the decoding step determines which, if any, of the at least two views is a base view that, for a decoding thereof, is independent of any other of the at least two views.
  • an apparatus includes an encoder for encoding at least two views corresponding to multi-view video content by encoding at least one of the at least two views in a resultant bitstream that is syntax compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/international Telecommunication Union, Telecommunication Sector H.264 recommendation, for backwards compatibility therewith.
  • the method includes encoding at least two views corresponding to multi-view video content by encoding at least one of the at least two views in a resultant bitstream that is syntax compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation, for backwards compatibility therewith.
  • an apparatus includes a decoder for decoding at least two views corresponding to multi-view video content, wherein at least one of the at least two views is included in a bitstream that is syntax compliant with the International Organization for Standardization/international Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation, for backwards compatibility therewith.
  • the method includes decoding at least two views corresponding to multi-view video content, wherein at least one of the at least two views is included in a bitstream that is syntax compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation, for backwards compatibility therewith.
  • an apparatus includes an encoder for encoding at least one of at least two views corresponding to multi-view video content by selecting between one of two-pre-defined slice types.
  • the method includes encoding at least one of at least two views corresponding to multi-view video content by selecting between one of two-pre-defined slice types.
  • an apparatus includes a decoder for decoding at least one of at least two views corresponding to multi-view video content by determining between one of two-pre-defined slice types.
  • the method includes decoding at least one of at least two views corresponding to multi-view video content by determining between one of two-pre-defined slice types.
  • an apparatus includes an encoder for encoding at least two views corresponding to multi-view content into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one camera parameter corresponding to at least one of the at least two views.
  • the method includes encoding at least two views corresponding to multi-view content into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one camera parameter corresponding to at least one of the at least two views.
  • an apparatus includes a decoder for decoding at least two views corresponding to multi-view content from a bitstream, wherein the bitstream is decoded to determine at least one camera parameter included therein.
  • the at least one camera parameter corresponds to at least one of the at least two views.
  • the method includes decoding at least two views corresponding to multi-view content from a bitstream, wherein the bitstream is decoded to deter mine at least one camera parameter included therein.
  • the at least one camera parameter corresponds to at least one of the at least two views.
  • an apparatus includes an encoder for encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views.
  • the method includes encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views.
  • an apparatus includes a decoder for decoding at least two views corresponding to multi-view video content from a bitstream, wherein the bitstream is decoded to determine at least one camera parameter for at least one of the at least two views based on at least one syntax element included in the bitstream.
  • the method includes decoding at least two views corresponding to multi-view video content from a bitstream, wherein the bitstream is decoded to determine at least one camera parameter for at least one of the at least two views based on at least one syntax element included in the bitstream.
  • FIG. 1 is a block diagram for an exemplary video encoder to which the present principles may be applied, in accordance with an embodiment of the present principles
  • FIG. 2 is a block diagram for an exemplary video decoder to which the present principles may be applied, in accordance with an embodiment of the present principles;
  • FIG. 3 is a diagram for an inter-view-temporal prediction structure based on the MPEG-4 AVC standard, using hierarchical B pictures, in accordance with an embodiment of the present principles
  • FIG. 4 is a flow diagram for an exemplary method for encoding multiple views of multi-view video content, in accordance with an embodiment of the present principles.
  • FIG. 5 is a flow diagram for an exemplary method for decoding multiple views of multi-view video content, in accordance with an embodiment of the present principles.
  • the present principles are directed to methods and apparatus for use in a multi-view video coding system.
  • processor or “controller” should not be construed to refer exclusively to hardware capable of executing software, and may implicitly include, without limitation, digital signal processor (“DSP”) hardware, read-only memory (“ROM”) for storing software, random access memory (“RAM”), and non-volatile storage.
  • DSP digital signal processor
  • ROM read-only memory
  • RAM random access memory
  • any switches shown in the figures are conceptual only. Their function may be carried out through the operation of program logic, through dedicated logic, through the interaction of program control and dedicated logic, or even manually, the particular technique being selectable by the implementer as more specifically understood from the context.
  • any element expressed as a means for performing a specified function is intended to encompass any way of performing that function including, for example, a) a combination of circuit elements that performs that function or b) software in any form, including, therefore, firmware, microcode or the like, combined with appropriate circuitry for executing that software to perform the function.
  • the present principles as defined by such claims reside in the fact that the functionalities provided by the various recited means are combined and brought together in the manner which the claims call for. It is thus regarded that any means that can provide those functionalities are equivalent to those shown herein.
  • FIG. 1 an exemplary video encoder to which the present principles may be applied is indicated generally by the reference numeral 100 .
  • An input to the video encoder 100 is connected in signal communication with a non-inverting input of a combiner 110 .
  • the output of the combiner 110 is connected in signal communication with a transformer/quantizer 120 .
  • the output of the transformer/quantizer 120 is connected in signal communication with an entropy coder 140 .
  • An output of the entropy coder 140 is available as an output of the encoder 100 .
  • the output of the transformer/quantizer 120 is further connected in signal communication with an inverse transformer/quantizer 150 .
  • An output of the inverse transformer/quantizer 150 is connected in signal communication with an input of a deblock filter 160 .
  • An output of the deblock filter 160 is connected in signal communication with reference picture stores 170 .
  • a first output of the reference picture stores 170 is connected in signal communication with a first input of a motion estimator 180 .
  • the input to the encoder 100 is further connected in signal communication with a second input of the motion estimator 180 .
  • the output of the motion estimator 180 is connected in signal communication with a first input of a motion compensator 190 .
  • a second output of the reference picture stores 170 is connected in signal communication with a second input of the motion compensator 190 .
  • the output of the motion compensator 190 is connected in signal communication with an inverting input of the combiner 110 .
  • FIG. 2 an exemplary video decoder to which the present principles may be applied is indicated generally by the reference numeral 200 .
  • the video decoder 200 includes an entropy decoder 210 for receiving a video sequence.
  • a first output of the entropy decoder 210 is connected in signal communication with an input of an inverse quantizer/transformer 220 .
  • An output of the inverse quantizer/transformer 220 is connected in signal communication with a first non-inverting input of a combiner 240 .
  • the output of the combiner 240 is connected in signal communication with an input of a deblock filter 290 .
  • An output of the deblock filter 290 is connected in signal communication with an input of a reference picture stores 250 .
  • the output of the reference picture stores 250 is connected in signal communication with a first input of a motion compensator 260 .
  • An output of the motion compensator 260 is connected in signal communication with a second non-inverting input of the combiner 240 .
  • a second output of the entropy decoder 210 is connected in signal communication with a second input of the motion compensator 260 .
  • the output of the deblock filter 290 is available as an output of the video decoder 200 .
  • a high level syntax is proposed for efficient processing of a multi-view sequence.
  • VPS View Parameter Set
  • the NAL unit types including a view identifier (id) in the NAL header to identify to which view the slice belongs.
  • view identifier id
  • high level syntax refers to syntax present in the bitstream that resides hierarchically above the macroblock layer.
  • high level syntax may refer to, but is not limited to, syntax at the slice header level, Supplemental Enhancement Information (SEI) level, picture parameter set level, and sequence parameter set level.
  • SEI Supplemental Enhancement Information
  • a base view may or may not be compatible with the MPEG-4 AVC standard, but an MPEG-4 AVC compatible view is always a base view.
  • an inter-view-temporal prediction structure based on the MPEG-4 AVC standard, using hierarchical B pictures is indicated generally by the reference numeral 300 .
  • the variable I denotes an intra coded picture
  • the variable P denotes a predictively coded picture
  • the variable B denotes a bi-predictively coded picture
  • the variable T denotes a location of a particular picture
  • the variable S denotes a particular view to which corresponds a particular picture.
  • Anchor picture is defined as a picture the decoding of which does not involve any picture sampled at a different time instance.
  • An anchor picture is signaled by setting the nal_ref_idc to 3.
  • all pictures in locations T 0 , T 8 . . . , T 96 , and T 100 are examples of anchor pictures.
  • Non-anchor picture is defined as a picture which does not have the above constraint specified for an anchor picture.
  • pictures B 2 , B 3 , and B 4 are non-anchor pictures.
  • Base view is a view which does not depend on any other view and can be independently decoded.
  • view S 0 is an example of base view.
  • a new parameter set is proposed called the View Parameter Set with its own NAL unit type and two new NAL unit types to support Multi-view Video Coding slices.
  • the MPEG-4 AVC standard includes the following two parameter sets: (1) Sequence Parameter Set (SPS), which includes information that is not expected to change over an entire sequence; and (2) Picture Parameter Set (PPS), which includes information that is not expected to change for each picture.
  • SPS Sequence Parameter Set
  • PPS Picture Parameter Set
  • Multi-view Video Coding has additional information which is specific to each view, we have created a separate View Parameter Set (VPS) in order to transmit this information. All the information that is needed to determine the dependency between the different views is indicated in the View Parameter Set.
  • the syntax table for the proposed View Parameter Set is shown in TABLE 1 (View Parameter Set RBSP syntax). This View Parameter Set is included in a new NAL unit type, for example, type 14 as shown in TABLE 2 (NAL unit type codes).
  • view_parameter_set_id identifies the view parameter set that is referred to in the slice header.
  • the value of the view_parameter_set_id shall be in the range of 0 to 255.
  • number_of_views_minus — 1 plus 1 identifies the total number of views in the bitstream.
  • the value of the number_of_view_minus — 1 shall be in the range of 0 to 255.
  • avc_compatible_view_id indicates the view_id of the AVC compatible view.
  • the value of avc_compatible_view_id shall be in the range of 0 to 255.
  • is_base_view_flag[i] 1 indicates that the view i is a base view and is independently decodable is_base_view_flag[i] equal to 0 indicates that the view i is not a base view.
  • the value of is_base_view_flag[i] shall be equal to 1 for an AVC compatible view i.
  • dependency_update_flag 1 indicates that dependency information for this view is updated in the VPS.
  • dependency_update_flag 0 indicates that the dependency information for this view is not updated and should not be changed.
  • anchor_picture_dependency_maps[i][j] 1 indicates the anchor pictures with view_id equal to j will depend on the anchor pictures with view_id equal to i.
  • non_anchor_picture_dependency_maps[i][j] 1 indicates the non-anchor pictures with view_id equal to j will depend on the non-anchor pictures with view_id equal to i.
  • non_anchor_picture_dependency_maps[i][j] is present only when anchor_picture_dependency_maps[i][j] equals 1. If anchor_picture_dependency_maps[i][j] is present and is equal to zero non_anchor_picture_dependency_maps[i][j] shall be inferred as being equal to 0.
  • Optional parameters in the View Parameter Set include the following:
  • camera_parameters_present_flag 1 indicates that a projection matrix is signaled as follows.
  • camera_parameters presuming camera parameter is conveyed in the form of a 3 ⁇ 4 projection matrix P, which can be used to map a point in the 3D world to the 2D image coordinate:
  • Each element camera_parameters_*_* can be represented according to the IEEE single precision floating point (32 bits) standard.
  • the decoder can create a map using all the dependency information once it receives the View Parameter Set. This enables it to know before it receives any slice which views are needed for decoding a particular view. As a result of this, we only need to parse the slice header to obtain the view_id and determine if this view is needed to decode a target view as indicated by a user. Thus, we do not need to buffer any frames or wait until a certain point to determine which frames are needed for decoding a particular view.
  • the dependency information and whether it is a base view is indicated in the View Parameter Set. Even an MPEG-4 AVC compatible base view has associated with it information that is specific to that view (e.g., camera parameters). This information may be used by other views for several purposes including view interpolation/synthesis.
  • Non-Multi-view Video Coding decoder By restricting it to just one such view, it is guaranteed that a non-Multi-view Video Coding decoder will be able to correctly decode the view and a Multi-view Video Coding decoder can easily identify such a view from the View Parameter Set using the syntax avc_compatible_view_id. All other base views (non-MPEG-4 AVC compatible) can be identified using the is_base_view_flag.
  • a new slice header for Multi-view Video Coding slices is proposed.
  • the View Parameter Set is identified using the view_parameter_set_id.
  • the view_id information is needed for several Multi-view Video Coding requirements including view interpolation/synthesis, view random access, parallel processing, and so forth. This information can also be useful for special coding modes that only relate to cross-view prediction.
  • view_parameter_set_id specifies the view parameter set in use.
  • the value of the view_parameter_set_id shall be in the range 0 to 255.
  • view_id indicates the view id of the current view.
  • the value of the view_parameter_set_id shall be in the range 0 to 255.
  • View random access is a Multi-view Video Coding requirement. The goal is to get access to any view with minimum decoding effort. Let us consider a simple example of view random access for the prediction structure shown in FIG. 3 .
  • view_id for the views are numbered consecutively from 0 to 7 in the slice header syntax and there is only one View Parameter Set present with view_parameter_set equal to 0. number_of_views_minus — 1 is set to 7. avc_compatible_view_id could be set to 0.
  • is_base_view_flag is set to 1 and for other views it is set to 0.
  • the dependency map for S 0 , S 1 , S 2 , S 3 , and S 4 will look as shown in TABLE 4A (Dependency table for S 0 anchor_picture_dependency_map) and TABLE 4B (dependency table for S 0 non_anchor_picture_dependency_map).
  • the dependency map for the other views can be written in a similar way.
  • the decoder can easily determine if a slice it receives is needed to decode a particular view.
  • the decoder only needs to parse the slice header to determine the view_id of the current slice and for the target view S 3 it can look up the S 3 columns in the two tables (TABLE 4a and TABLE 4B) to determine whether or not it should keep the current slice.
  • the decoder needs to distinguish between anchor pictures and non-anchor pictures since they may have different dependencies as can be seen from TABLE 4a and TABLE 4b.
  • For the target view S 3 we need to decode the anchor pictures of views S 0 , S 2 , and S 4 but only need to decode the non-anchor pictures of views S 2 and S 4 .
  • FIG. 4 an exemplary method for encoding multiple views of multi-view video content is indicated generally by the reference numeral 400 .
  • the method 400 includes a start block 405 that passes control to a function block 410 .
  • the function block 410 reads a configuration file for the encoding parameters to be used to encode the multiple views, and passes control to a function block 415 .
  • the function block sets N to be equal to the number of views to be encoded, and passes control to a function block 420 .
  • the function block 420 sets number_of_views_minus — 1 equal to N ⁇ 1, sets avc_compatible_view_id equal to the view_id of the MPEG-4 AVC compatible view, and passes control to a function block 425 .
  • the function block 425 sets view_parameter_set_id equal to a valid integer, initializes a variable i to be equal to zero, and passes control to a decision block 430 .
  • the decision block 430 determines whether or not i is greater than N. If so, then control is passed to a decision block 435 . Otherwise, control is passed to a function block 470 .
  • the decision block 435 determines whether or not the current view is a base view. If so, then control is passed to a function block 440 . Otherwise, control is passed to a function block 480 .
  • the function block 440 sets is_base_view_flag[i] equal to one, and passes control to a decision block 445 .
  • the decision block 445 determines whether or not the dependency is being updated. If so, the control is passed to a function block 450 . Otherwise, control is passed to a function block 485 .
  • the function block 450 sets dependency_update_flag equal to one, and passes control to a function block 455 .
  • the function block 455 sets a variable j equal to 0, and passes control to a decision block 460 .
  • the decision block 460 determines whether or not j is less than N. If so, then control is passed to a function block 465 . Otherwise, control is passed to the function block 487 .
  • the function block 465 sets anchor_picture_dependency_maps[i][j] and non_anchor_picture_dependency_maps[i][j] to values indicated by configuration file, and passes control to a function block 467 .
  • the function block 467 increments the variable j by one, and returns control to the decision block 460 .
  • the function block 470 sets camera_parameters_present_flag equal to one when camera parameters are present, sets camera_parameters_present_flag equal to zero otherwise, and passes control to a decision block 472 .
  • the decision block 472 determines whether or not camera_parameters_present_flag is equal to one. If so, then control is passed to a function block 432 . Otherwise, control is passed to a function block 434 .
  • the function block 432 writes the camera parameters, and passes control to the function block 434 .
  • the function block 434 writes the View Parameter Set (VPS) or the Sequence Parameter Set (SPS), and passes control to an end block 499 .
  • VPS View Parameter Set
  • SPS Sequence Parameter Set
  • the function block 480 sets is_base_view_flag[i] equal to zero, and passes control to the decision block 445 .
  • the function block 485 sets dependency_update_flag equal to zero, and passes control to a function block 487 .
  • the function block 487 increments the variable i by 1, and returns control to the decision block 430 .
  • an exemplary method for decoding multiple views of multi-view video content is indicated generally by the reference numeral 500 .
  • the method 500 includes a start block 505 that passes control to a function block 510 .
  • the function block 510 parses a Sequence Parameter Set (SPS) or View Parameter Set (VPS), view_parameter_set_id, number_of_views_minus — 1, avc_compatible_view_id, sets variables I and j equal to zero, sets N equal to number_of_views_minus — 1, and passes control to a decision block 515 .
  • the decision block 515 determines whether or not i is less than or equal to N. If so, then control is passed to a function block 570 . Otherwise, control is passed to a function block 525 .
  • the function block 570 parses camera_parameters_present_flag, and passes control to a decision block 572 .
  • the decision block 572 determines whether or not camera_parameters_present_flag is equal to one. If so, then control is passed to a function block 574 . Otherwise, control is passed to a function block 576 .
  • the function block 574 parses the camera parameters, and passes control to the function block 576 .
  • the function block 576 continues decoding, and passes control to an end block 599 .
  • the function block 525 parses is_base_view_flag[i] and dependency_update_flag, and passes control to a decision block 530 .
  • the decision block 530 determines whether or not dependency_update_flag is equal to zero. If so, then control is passes to a function block 532 . Otherwise, control is passed to a decision block 535 .
  • the function block 532 increments i by one, and returns control to the decision block 515 .
  • the decision block 535 determines whether or not j is less than or equal to N. If so, then control is passed to a function block 540 . Otherwise, control is passes to a function block 537 .
  • the function block 540 parses anchor_picture_dependency_maps[i][j], and passes control to a decision block 545 .
  • the decision block 545 determines whether or not non_anchor_picture_dependency_maps[i][j] is equal to one. If so, then control is passed to a function block 550 . Otherwise, control is passes to a function block 547 .
  • the function block 550 parses the non_anchor_picture_dependency_maps[i][j], and passes control to the function block 547 .
  • the function block 547 increments j by one, and returns control to the decision block 535 .
  • the function block 537 increments i by one, and returns control to the function block 515 .
  • one advantage/feature is an apparatus that includes an encoder for encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include view specific information.
  • the view specific information indicates a decoding interdependency between at least some of the at least two views.
  • Another advantage/feature is the apparatus having the encoder as described above, wherein the decoding interdependency allows a corresponding decoding of at least one of the at least two views using only a subset of the at least two views for the corresponding decoding.
  • Yet another advantage/feature is the apparatus having the encoder as described above, wherein the decoding interdependency indicated in the view specific information is used for random access of at least one of the at least two views by dropping slices related to any other ones of the at least two views indicated as non-interdependent with respect to the at least one view. Still another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information is included in a high level syntax. A further advantage/feature is the apparatus having the encoder as described above, wherein the view specific information is included in a parameter set compliant with the International Organization for Standardization/international Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/international Telecommunication Union, Telecommunication Sector H.264 recommendation.
  • a yet further advantage/feature is the apparatus having the encoder as described above, wherein the view specific information is included in a View Parameter Set.
  • a still further advantage/feature is the apparatus having the encoder wherein the view specific information is included in a View Parameter Set as described above, wherein the View Parameter Set is assigned a NAL unit type specifically for use only with the View Parameter Set.
  • An additional advantage/feature is the apparatus having the encoder wherein a NAL unit type is assigned specifically for use only with the View Parameter Set as described above, wherein the NAL unit type is 14.
  • another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information includes at least one syntax element for indicating a View Parameter Set id.
  • another advantage/feature is the apparatus having the encoder wherein the view specific information includes at least one syntax element for indicating a View Parameter Set id as described above, wherein the at least one syntax element is denoted by a view_parameter_set_id syntax element. Also, another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information includes at least one syntax element for indicating a number of views. Additionally, another advantage/feature is the apparatus having the encoder wherein the view specific information includes at least one syntax element for indicating a number of views as described above, wherein the at least one syntax element is denoted by a number_of_views_minus — 1 syntax element.
  • another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information includes at least one syntax element for indicating a view id for a particular one of the at least two views, when the particular one of the at least two views is encoded in a resultant bitstream that is compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/international Telecommunication Union, Telecommunication Sector H.264 recommendation.
  • the view specific information includes at least one syntax element for indicating a view id for a particular one of the at least two views as described above, wherein the at least one syntax element is de noted by an avc_compatible_view_id syntax element.
  • another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information includes at least one syntax element or is implicitly derivable from a high level syntax, the at least one syntax element and the high level syntax for indicating that a particular one of the at least two views is compatible with the International Organization for Standardization/international Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation. Additionally, another advantage/feature is the apparatus having the encoder wherein the view specific information includes at least one syntax element or is implicitly derivable from a high level syntax as described above, wherein the at least one syntax element is denoted by an is_base_view_flag syntax element.
  • another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information includes at least one syntax element for indicating whether dependency information for at least one of the at least two views is present in the resultant bitstream. Further, another advantage/feature is the apparatus having the encoder wherein the view specific information includes at least one syntax element for indicating whether dependency information for at least one of the at least two views is present in the resultant bitstream as described above, wherein the at least one syntax element is denoted by a dependency_update_flag syntax element.
  • another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information includes at least one syntax element for indicating whether at least one anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views.
  • the view specific information includes at least one syntax element for indicating whether at least one anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views as described above, wherein the at least one syntax element is denoted by an anchor_picture_dependency_maps[i][j] syntax element.
  • another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information includes at least one syntax element for indicating whether at least one non-anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views. Additionally, another advantage/feature is the apparatus having the encoder wherein the view specific information includes at least one syntax element for indicating whether at least one non-anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views as described above, wherein the at least one syntax element is denoted by a non_anchor_picture_dependency_maps[i][j] syntax element.
  • another advantage/feature is the apparatus having the encoder as described above, wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views. Further, another advantage/feature is the apparatus having the encoder wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views as described above, wherein the at least one syntax is included in a parameter set corresponding to the resultant bitstream.
  • another advantage/feature is an apparatus that includes an encoder for encoding at least two views corresponding to multi-view video content by defining as a base view any of the at least two views that, for a decoding thereof, is independent of any other of the at least two views.
  • another advantage/feature is an apparatus that includes an encoder for encoding at least two views corresponding to multi-view video content by encoding at least one of the at least two views in a resultant bitstream that is syntax compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation, for backwards compatibility therewith.
  • another advantage/feature is the apparatus having the encoder as described above, wherein the at least one view is a base view that, for a decoding thereof, is independent of any other of the at least two views.
  • an avc_compatible_view_id syntax element identifies the at least one view as being encoded in the resultant bitstream that is syntax compliant with the International Organization for Standardization/international Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation for backwards compatibility.
  • another advantage/feature is an apparatus that includes an encoder for encoding at least one of at least two views corresponding to multi-view video content by selecting between one of two-pre-defined slice types. Further, another advantage/feature is the apparatus having the encoder as described above, wherein the two pre-defined slice types are an Instantaneous Decoding Refresh slice type and a non-Instantaneous Decoding Refresh slice type.
  • another advantage/feature is the apparatus having the encoder that selects between the Instantaneous Decoding Refresh slice type and the non-Instantaneous Decoding Refresh slice type as described above, wherein NAL unit type 22 is used for the Instantaneous Decoding Refresh slice type and NAL unit type 23 is used for the non-Instantaneous Decoding Refresh slices.
  • another advantage/feature is the apparatus having the encoder as described above, wherein slice headers for at least one of the at least two slices includes view specific syntax.
  • another advantage/feature is the apparatus having the encoder wherein slice headers for at least one of the at least two slices includes view specific syntax as described above, wherein the view specific syntax is conditioned on NAL unit type 23 and NAL unit type 24 .
  • another advantage/feature is the apparatus having the encoder wherein slice headers for at least one of the at least two slices includes view specific syntax as described above, wherein the view specific syntax includes a view parameter set identifier and a view identifier. Also, another advantage/feature is the apparatus having the encoder wherein the view specific syntax includes a view parameter set identifier and a view identifier as described above, wherein the view parameter set identifier is denoted by a view_parameter_set_id syntax element and the view identifier is denoted by a view_id syntax element.
  • another advantage/feature is an apparatus that includes an encoder for encoding at least two views corresponding to multi-view content into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one camera parameter corresponding to at least one of the at least two views.
  • another advantage/feature is the apparatus having the encoder as described above, wherein the resultant bitstream is encoded to include a View Parameter Set, and the at least one camera parameter is included in the View Parameter Set.
  • another advantage/feature is the apparatus having the encoder as described above, wherein a presence of the at least one camera parameter is indicated by a syntax element.
  • another advantage/feature is the apparatus having the encoder wherein a presence of the at least one camera parameter is indicated by a syntax element as described above, wherein the syntax element is a camera_parameters_present_flag syntax element. Additionally, another advantage/feature is the apparatus having the encoder as described above, wherein the at least one camera parameter is denoted by a camera_parameters syntax element.
  • another advantage/feature is an apparatus that includes an encoder for encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views.
  • another advantage/feature is the apparatus having the encoder as described above, wherein the at least one syntax is a high level syntax element.
  • another advantage/feature is the apparatus having the encoder as described above, wherein the at least one syntax is included in a parameter set corresponding to the resultant bitstream.
  • the teachings of the present principles are implemented as a combination of hardware and software.
  • the software may be implemented as an application program tangibly embodied on a program storage unit.
  • the application program may be uploaded to, and executed by, a machine comprising any suitable architecture.
  • the machine is implemented on a computer platform having hardware such as one or more central processing units (“CPU”), a random access memory (“RAM”), and input/output (“I/O”) interfaces.
  • CPU central processing units
  • RAM random access memory
  • I/O input/output
  • the computer platform may also include an operating system and microinstruction code.
  • the various processes and functions described herein may be either part of the microinstruction code or part of the application program, or any combination thereof, which may be executed by a CPU.
  • various other peripheral units may be connected to the computer platform such as an additional data storage unit and a printing unit.

Abstract

There are provided method and apparatus for use in a multi-view video coding system. The apparatus includes an encoder for encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include view specific information. The view specific information indicates a decoding interdependency between at least some of the at least two views.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application claims the benefit of U.S. Provisional Application Ser. No. 60/787,092, filed 29 Mar., 2006, which is incorporated by reference herein in its entirety. This application is also closely related to two other applications concurrently filed with the instant application and having common inventors and a common assignee (Attorney docket Nos. PU070031 and PU070032), each of which are incorporated by reference in their entireties.
  • TECHNICAL FIELD
  • The present principles relate generally to video encoding and decoding and, more particularly, to methods and apparatus for use in a multi-view video coding system.
  • BACKGROUND
  • In the current implementation of Multi-view Video Coding (MVC) compliant with the International Organization for Standardization/International Electrotechnical Commission (ISO/IEC) Moving Picture Experts Group-4 (MPEG-4) Part 10 Advanced Video Coding (AVC) standard/international Telecommunication Union, Telecommunication Sector (ITU-T) H.264 recommendation (hereinafter the “MPEG-4 AVC standard”), there is no provision to identify a specific view and to signal the camera parameters. This view information is needed for several reasons. View scalability, view random access, parallel processing, view generation, and view synthesis are all Multi-view Video Coding requirements which utilize the view information. Moreover, several of these requirements also utilize camera parameters which are currently not passed in a standardized way.
  • A method has been proposed to enable efficient random access in multi-view compressed bit streams. In the proposed method, a V-picture type and a View Dependency Supplemental Enhancement Information (SEI) message are defined. A feature required in the proposed V-picture type is that V-pictures shall have no temporal dependence on other pictures in the same camera and may only be predicted from pictures in other cameras at the same time. The proposed View Dependency Supplemental Enhancement Information message will describe exactly which views a V-picture, as well as the preceding and following sequences of pictures, may depend on. The following are the details of the proposed changes.
  • With respect to V-Picture syntax and semantics, a particular syntax table relating to the MPEG-4 AVC standard is extended to include a Network Abstraction Layer (NAL) unit type of 14 corresponding to a V-picture. Also, the V-picture type is defined to have the following semantics:
  • V-picture: A coded picture in which all slices reference only slices with the same temporal index (i.e., only slices in other views and not slices in the current view). When a V-picture would be output or displayed, it also causes the decoding process to mark all pictures from the same view which are not IDR-pictures or V-pictures and which precede the V-picture in output order to be marked as “unused for reference”. Each V-picture shall be associated with a View Dependency SEI message occurring in the same NAL.
  • With respect to the view dependency Supplemental Enhancement Information message syntax and semantics, a View Dependency Supplemental Enhancement Information message is defined with the following syntax:
  • view_dependency ( payloadSize ) {
      num_seq_reference_views ue(v)
      seq_reference_view_0 ue(v)
      seq_reference_view_1 ue(v)
      ...
      seq_reference_view_N ue(v)
      num_pic_reference_views ue(v)
      pic_reference_view_0 ue(v)
      pic_reference_view_1 ue(v)
      ...
      pic_reference_view_N ue(v)
    }

    where num_seq_reference_views/num_pic_reference_views denotes the number of potential views that can be used as a reference for the current sequence/picture, and seq_reference_view_i/pic_reference_view_i denotes the view number for the ith reference view.
  • The picture associated with a View Dependency Supplemental Enhancement Information message shall only reference the specified views described by pic_reference_view_i. Similarly, all subsequent pictures in output order of that view until the next View Dependency Supplemental Enhancement Information message in that view shall only reference the specified views described by seq_reference_view_i.
  • A View Dependency Supplemental Enhancement Information message shall be associated with each Instantaneous Decoding Refresh (IDR) picture and V-picture.
  • At least one drawback of this method is the complexity introduced in the decoder due to the dependency being recursively obtained. Additionally, this method requires that every V-picture carry an SEI message (which is a non-normative part of the MPEG-4 AVC standard), resulting in the dependency being unable to be used for normative behavior such as reference picture selection.
  • SUMMARY
  • These and other drawbacks and disadvantages of the prior art are addressed by the present principles, which are directed to methods and apparatus for use in a multi-view video coding system.
  • According to an aspect of the present principles, there is provided an apparatus. The apparatus includes an encoder for encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include view specific information. The view specific information indicates a decoding interdependency between at least some of the at least two views.
  • According to another aspect of the present principles, there is provided a method. The method includes encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include view specific information. The view specific information indicates a decoding interdependency between at least some of the at least two views.
  • According to yet another aspect of the present principles, there is provided an apparatus. The apparatus includes a decoder for decoding at least two views corresponding to multi-view video content from a bitstream, wherein the bitstream is decoded to determine view specific information included therein, the view specific information indicating a decoding interdependency between at least some of the at least two views.
  • According to still another aspect of the present principles, there is provided a method. The method includes decoding at least two views corresponding to multi-view video content from a bitstream, wherein the bitstream is decoded to determine view specific information included therein. The view specific information indicates a decoding interdependency between at least some of the at least two views.
  • According to a further aspect of the present principles, there is provided an apparatus. The apparatus includes an encoder for encoding at least two views corresponding to multi-view video content by defining as a base view any of the at least two views that, for a decoding thereof, is independent of any other of the at least two views.
  • According to a yet further aspect of the present principles, there is provided a method. The method includes encoding at least two views corresponding to multi-view video content by defining as a base view any of the at least two views that, for a decoding thereof, is independent of any other of the at least two views.
  • According to a still further aspect of the present principles, there is provided an apparatus. The apparatus includes a decoder for decoding at least two views corresponding to multi-view video content, wherein the decoder determines which, if any, of the at least two views is a base view that, for a decoding thereof, is independent of any other of the at least two views.
  • According to an additional aspect of the present principles, there is provided a method. The method includes decoding at least two views corresponding to multi-view video content, wherein the decoding step determines which, if any, of the at least two views is a base view that, for a decoding thereof, is independent of any other of the at least two views.
  • According to an aspect of the present principles, there is provided an apparatus. The apparatus includes an encoder for encoding at least two views corresponding to multi-view video content by encoding at least one of the at least two views in a resultant bitstream that is syntax compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/international Telecommunication Union, Telecommunication Sector H.264 recommendation, for backwards compatibility therewith.
  • According to another aspect of the present principles, there is provided a method. The method includes encoding at least two views corresponding to multi-view video content by encoding at least one of the at least two views in a resultant bitstream that is syntax compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation, for backwards compatibility therewith.
  • According to yet another aspect of the present principles, there is provided an apparatus. The apparatus includes a decoder for decoding at least two views corresponding to multi-view video content, wherein at least one of the at least two views is included in a bitstream that is syntax compliant with the International Organization for Standardization/international Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation, for backwards compatibility therewith.
  • According to still another aspect of the present principles, there is provided a method. The method includes decoding at least two views corresponding to multi-view video content, wherein at least one of the at least two views is included in a bitstream that is syntax compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation, for backwards compatibility therewith.
  • According to a further aspect of the present principles, there is provided an apparatus. The apparatus includes an encoder for encoding at least one of at least two views corresponding to multi-view video content by selecting between one of two-pre-defined slice types.
  • According to a yet further aspect of the present principles, there is provided a method. The method includes encoding at least one of at least two views corresponding to multi-view video content by selecting between one of two-pre-defined slice types.
  • According to a still further aspect of the present principles, there is provided an apparatus. The apparatus includes a decoder for decoding at least one of at least two views corresponding to multi-view video content by determining between one of two-pre-defined slice types.
  • According to an additional aspect of the present principles, there is provided a method. The method includes decoding at least one of at least two views corresponding to multi-view video content by determining between one of two-pre-defined slice types.
  • According to an aspect of the present principles, there is provided an apparatus. The apparatus includes an encoder for encoding at least two views corresponding to multi-view content into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one camera parameter corresponding to at least one of the at least two views.
  • According to another aspect of the present principles, there is provided a method. The method includes encoding at least two views corresponding to multi-view content into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one camera parameter corresponding to at least one of the at least two views.
  • According to yet another aspect of the present principles, there is provided an apparatus. The apparatus includes a decoder for decoding at least two views corresponding to multi-view content from a bitstream, wherein the bitstream is decoded to determine at least one camera parameter included therein. The at least one camera parameter corresponds to at least one of the at least two views.
  • According to still another aspect of the present principles, there is provided a method. The method includes decoding at least two views corresponding to multi-view content from a bitstream, wherein the bitstream is decoded to deter mine at least one camera parameter included therein. The at least one camera parameter corresponds to at least one of the at least two views.
  • According to a further aspect of the present principles, there is provided an apparatus. The apparatus includes an encoder for encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views.
  • According to a yet further aspect of the present principles, there is provided a method. The method includes encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views.
  • According to a still further aspect of the present principles, there is provided an apparatus. The apparatus includes a decoder for decoding at least two views corresponding to multi-view video content from a bitstream, wherein the bitstream is decoded to determine at least one camera parameter for at least one of the at least two views based on at least one syntax element included in the bitstream.
  • According to an additional aspect of the present principles, there is provided a method. The method includes decoding at least two views corresponding to multi-view video content from a bitstream, wherein the bitstream is decoded to determine at least one camera parameter for at least one of the at least two views based on at least one syntax element included in the bitstream.
  • These and other aspects, features and advantages of the present principles will become apparent from the following detailed description of exemplary embodiments, which is to be read in connection with the accompanying drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The present principles may be better understood in accordance with the following exemplary figures, in which:
  • FIG. 1 is a block diagram for an exemplary video encoder to which the present principles may be applied, in accordance with an embodiment of the present principles;
  • FIG. 2 is a block diagram for an exemplary video decoder to which the present principles may be applied, in accordance with an embodiment of the present principles;
  • FIG. 3 is a diagram for an inter-view-temporal prediction structure based on the MPEG-4 AVC standard, using hierarchical B pictures, in accordance with an embodiment of the present principles;
  • FIG. 4 is a flow diagram for an exemplary method for encoding multiple views of multi-view video content, in accordance with an embodiment of the present principles; and
  • FIG. 5 is a flow diagram for an exemplary method for decoding multiple views of multi-view video content, in accordance with an embodiment of the present principles.
  • DETAILED DESCRIPTION
  • The present principles are directed to methods and apparatus for use in a multi-view video coding system.
  • The present description illustrates the present principles. It will thus be appreciated that those skilled in the art will be able to devise various arrangements that, although not explicitly described or shown herein, embody the present principles and are included within its spirit and scope.
  • All examples and conditional language recited herein are intended for pedagogical purposes to aid the reader in understanding the present principles and the concepts contributed by the inventor(s) to furthering the art, and are to be construed as being without limitation to such specifically recited examples and conditions.
  • Moreover, all statements herein reciting principles, aspects, and embodiments of the present principles, as well as specific examples thereof, are intended to encompass both structural and functional equivalents thereof. Additionally, it is intended that such equivalents include both currently known equivalents as well as equivalents developed in the future, i.e., any elements developed that perform the same function, regardless of structure.
  • Thus, for example, it will be appreciated by those skilled in the art that the block diagrams presented herein represent conceptual views of illustrative circuitry embodying the present principles. Similarly, it will be appreciated that any flow charts, flow diagrams, state transition diagrams, pseudocode, and the like represent various processes which may be substantially represented in computer readable media and so executed by a computer or processor, whether or not such computer or processor is explicitly shown.
  • The functions of the various elements shown in the figures may be provided through the use of dedicated hardware as well as hardware capable of executing software in association with appropriate software. When provided by a processor, the functions may be provided by a single dedicated processor, by a single shared processor, or by a plurality of individual processors, some of which may be shared. Moreover, explicit use of the term “processor” or “controller” should not be construed to refer exclusively to hardware capable of executing software, and may implicitly include, without limitation, digital signal processor (“DSP”) hardware, read-only memory (“ROM”) for storing software, random access memory (“RAM”), and non-volatile storage.
  • Other hardware, conventional and/or custom, may also be included. Similarly, any switches shown in the figures are conceptual only. Their function may be carried out through the operation of program logic, through dedicated logic, through the interaction of program control and dedicated logic, or even manually, the particular technique being selectable by the implementer as more specifically understood from the context.
  • In the claims hereof, any element expressed as a means for performing a specified function is intended to encompass any way of performing that function including, for example, a) a combination of circuit elements that performs that function or b) software in any form, including, therefore, firmware, microcode or the like, combined with appropriate circuitry for executing that software to perform the function. The present principles as defined by such claims reside in the fact that the functionalities provided by the various recited means are combined and brought together in the manner which the claims call for. It is thus regarded that any means that can provide those functionalities are equivalent to those shown herein.
  • Reference in the specification to “one embodiment” or “an embodiment” of the present principles means that a particular feature, structure, characteristic, and so forth described in connection with the embodiment is included in at least one embodiment of the present principles. Thus, the appearances of the phrase “in one embodiment” or “in an embodiment” appearing in various places throughout the specification are not necessarily all referring to the same embodiment.
  • Turning to FIG. 1, an exemplary video encoder to which the present principles may be applied is indicated generally by the reference numeral 100.
  • An input to the video encoder 100 is connected in signal communication with a non-inverting input of a combiner 110. The output of the combiner 110 is connected in signal communication with a transformer/quantizer 120. The output of the transformer/quantizer 120 is connected in signal communication with an entropy coder 140. An output of the entropy coder 140 is available as an output of the encoder 100.
  • The output of the transformer/quantizer 120 is further connected in signal communication with an inverse transformer/quantizer 150. An output of the inverse transformer/quantizer 150 is connected in signal communication with an input of a deblock filter 160. An output of the deblock filter 160 is connected in signal communication with reference picture stores 170. A first output of the reference picture stores 170 is connected in signal communication with a first input of a motion estimator 180. The input to the encoder 100 is further connected in signal communication with a second input of the motion estimator 180. The output of the motion estimator 180 is connected in signal communication with a first input of a motion compensator 190. A second output of the reference picture stores 170 is connected in signal communication with a second input of the motion compensator 190. The output of the motion compensator 190 is connected in signal communication with an inverting input of the combiner 110.
  • Turning to FIG. 2, an exemplary video decoder to which the present principles may be applied is indicated generally by the reference numeral 200.
  • The video decoder 200 includes an entropy decoder 210 for receiving a video sequence. A first output of the entropy decoder 210 is connected in signal communication with an input of an inverse quantizer/transformer 220. An output of the inverse quantizer/transformer 220 is connected in signal communication with a first non-inverting input of a combiner 240.
  • The output of the combiner 240 is connected in signal communication with an input of a deblock filter 290. An output of the deblock filter 290 is connected in signal communication with an input of a reference picture stores 250. The output of the reference picture stores 250 is connected in signal communication with a first input of a motion compensator 260. An output of the motion compensator 260 is connected in signal communication with a second non-inverting input of the combiner 240. A second output of the entropy decoder 210 is connected in signal communication with a second input of the motion compensator 260. The output of the deblock filter 290 is available as an output of the video decoder 200.
  • In an embodiment of the present principles, a high level syntax is proposed for efficient processing of a multi-view sequence. In particular, we propose creating a new parameter set called View Parameter Set (VPS) with its own NAL unit type and two more new NAL unit types to support multi-view slices, with the NAL unit types including a view identifier (id) in the NAL header to identify to which view the slice belongs. For view scalability and backward compatibility with decoders compliant with the MPEG-4 AVC standard, we propose to maintain one MPEG-4 AVC compliant view which we call an “MPEG-4 AVC compliant Base View”.
  • As used herein, “high level syntax” refers to syntax present in the bitstream that resides hierarchically above the macroblock layer. For example, high level syntax, as used herein, may refer to, but is not limited to, syntax at the slice header level, Supplemental Enhancement Information (SEI) level, picture parameter set level, and sequence parameter set level.
  • In the current implementation of the Multi-view Video Coding system described above as having no provision to identify a specific view and to signal camera parameters, different views are interleaved to form a single sequence instead of treating the different views as separate views. Since the syntax is compatible with the MPEG-4 AVC standard, as noted above, it is presently not possible to identify which view a given slice belongs to. This view information is needed for several reasons. View scalability, view random access, parallel processing, view generation, and view synthesis are all Multi-view Video Coding requirements which need to identify a view. For efficient support of view random access and view scalability, it is important for the decoder to know how different pictures depend on each other, so only pictures that are necessary are decoded. Camera parameters are needed for view synthesis. If view synthesis is eventually used in the decoding loop, a standardized way of signaling camera parameters needs to be specified. In accordance with an embodiment, a view parameter set is used.
  • In an embodiment, it is presumed that one view is needed that is fully backward compatible with the MPEG-4 AVC standard for the purpose of supporting non-MVC compatible but MPEG-4 AVC compatible decoders. In an embodiment, it is presumed that there will be views that are independently decodable to facilitate fast view random access. We refer to these views as “base views”. A base view may or may not be compatible with the MPEG-4 AVC standard, but an MPEG-4 AVC compatible view is always a base view.
  • Turning to FIG. 3, an inter-view-temporal prediction structure based on the MPEG-4 AVC standard, using hierarchical B pictures, is indicated generally by the reference numeral 300. In FIG. 3, the variable I denotes an intra coded picture, the variable P denotes a predictively coded picture, the variable B denotes a bi-predictively coded picture, the variable T denotes a location of a particular picture, and the variable S denotes a particular view to which corresponds a particular picture.
  • In accordance with an embodiment, the following terms are defined:
  • Anchor picture is defined as a picture the decoding of which does not involve any picture sampled at a different time instance. An anchor picture is signaled by setting the nal_ref_idc to 3. In FIG. 3, all pictures in locations T0, T8 . . . , T96, and T100 are examples of anchor pictures.
  • Non-anchor picture is defined as a picture which does not have the above constraint specified for an anchor picture. In FIG. 3, pictures B2, B3, and B4 are non-anchor pictures.
  • Base view is a view which does not depend on any other view and can be independently decoded. In FIG. 3, view S0 is an example of base view.
  • Also, in an embodiment, a new parameter set is proposed called the View Parameter Set with its own NAL unit type and two new NAL unit types to support Multi-view Video Coding slices. We also modify the slice header syntax to indicate the view_id and the view parameter set to be used.
  • The MPEG-4 AVC standard includes the following two parameter sets: (1) Sequence Parameter Set (SPS), which includes information that is not expected to change over an entire sequence; and (2) Picture Parameter Set (PPS), which includes information that is not expected to change for each picture.
  • Since Multi-view Video Coding has additional information which is specific to each view, we have created a separate View Parameter Set (VPS) in order to transmit this information. All the information that is needed to determine the dependency between the different views is indicated in the View Parameter Set. The syntax table for the proposed View Parameter Set is shown in TABLE 1 (View Parameter Set RBSP syntax). This View Parameter Set is included in a new NAL unit type, for example, type 14 as shown in TABLE 2 (NAL unit type codes).
  • In accordance with the description of the present invention, the following terms are defined:
  • view_parameter_set_id identifies the view parameter set that is referred to in the slice header. The value of the view_parameter_set_id shall be in the range of 0 to 255.
  • number_of_views_minus 1 plus 1 identifies the total number of views in the bitstream. The value of the number_of_view_minus 1 shall be in the range of 0 to 255.
  • avc_compatible_view_id indicates the view_id of the AVC compatible view. The value of avc_compatible_view_id shall be in the range of 0 to 255.
  • is_base_view_flag[i] equal to 1 indicates that the view i is a base view and is independently decodable is_base_view_flag[i] equal to 0 indicates that the view i is not a base view. The value of is_base_view_flag[i] shall be equal to 1 for an AVC compatible view i.
  • dependency_update_flag equal to 1 indicates that dependency information for this view is updated in the VPS. dependency_update_flag equal to 0 indicates that the dependency information for this view is not updated and should not be changed.
  • anchor_picture_dependency_maps[i][j] equal to 1 indicates the anchor pictures with view_id equal to j will depend on the anchor pictures with view_id equal to i.
  • non_anchor_picture_dependency_maps[i][j] equal to 1 indicates the non-anchor pictures with view_id equal to j will depend on the non-anchor pictures with view_id equal to i. non_anchor_picture_dependency_maps[i][j] is present only when anchor_picture_dependency_maps[i][j] equals 1. If anchor_picture_dependency_maps[i][j] is present and is equal to zero non_anchor_picture_dependency_maps[i][j] shall be inferred as being equal to 0.
  • TABLE 1
    view_parameter_set_rbsp( ) { C Descriptor
    view_parameter_set_id 0 ue(v)
    number_of_views_minus_1 0 ue(v)
    avc_compatible_view_id 0 ue(v)
    for( i = 0; i <= number_of_views_minus_1; i++ ) {
    is_base_view_flag[i] 0 u(1)
    dependency_update_flag 0 u(1)
    if (dependency_update_flag == 1) {
    for(j = 0; j <= number_of__views_minus_1; j++) {
    anchor_picture_dependency_maps[i][j] 0 f(1)
    if (anchor_picture_dependency_maps[i][j] == 1)
    non_anchor_picture_dependency_maps[i][j] 0 f(1)
    }
    }
    }
    for( i = 0; i <= number_of_views_minus_1; i++ ) {
    camera_parameters_present_flag 0 u(1)
    if (camera_parameters_present_flag == 1) {
    camera_parameters_1_1[i] 0 f(32)
    ***
    camera_parameters_3_4[i] 0 f(32)
    }
    }
    }
  • TABLE 2
    nal_unit_type Content of NAL unit and RBSP syntax structure C
    0 Unspecified
    1 Coded slice of a non-IDR picture 2, 3, 4
    slice_layer_without_partitioning_rbsp( )
    2 Coded slice data partition A 2
    slice_data_partition_a_layer_rbsp( )
    3 Coded slice data partition B 3
    slice_data_partition_b_layer_rbsp( )
    4 Coded slice data partition C 4
    slice_data_partition_c_layer_rbsp( )
    5 Coded slice of an IDR picture 2, 3
    slice_layer_without_partitioning_rbsp( )
    6 Supplemental enhancement information (SEI) 5
    sei_rbsp( )
    7 Sequence parameter set 0
    seq_parameter_set_rbsp( )
    8 Picture parameter set 1
    pic_parameter_set_rbsp( )
    9 Access unit delimiter 6
    access_unit_delimiter_rbsp( )
    10 End of sequence 7
    end_of_seq_rbsp( )
    11 End of stream 8
    end_of_stream_rbsp( )
    12 Filler data 9
    filler_data_rbsp( )
    13 Sequence parameter set extension 10 
    seq_parameter_set_extension_rbsp( )
    14 View parameter set 11 
    view_parameter_set_rbps( )
    15 . . . 18 Reserved
    19 Coded slice of an auxiliary coded picture without partitioning 2, 3, 4
    slice_layer_without_partitioning_rbsp( )
    20 Coded slice of a non-IDR picture in scalable extension 2, 3, 4
    slice_layer_in_scalable_extension_rbsp( )
    21 Coded slice of an IDR picture in scalable extension 2, 3
    slice_layer_in_scalable_extension_rbsp( )
    22 Coded slice of a non-IDR picture in multi-view extension 2, 3, 4
    slice_layer_in_mvc_extension_rbsp( )
    23 Coded slice of an IDR picture in multi-view extension 2, 3
    slice_layer_in_mvc_extension_rbsp( )
    24 . . . 31 Unspecified
  • Optional parameters in the View Parameter Set include the following:
  • camera_parameters_present_flag equal to 1 indicates that a projection matrix is signaled as follows.
  • camera_parameters, presuming camera parameter is conveyed in the form of a 3×4 projection matrix P, which can be used to map a point in the 3D world to the 2D image coordinate:

  • I=P*[X w :Y w :Z w:1]
  • where I is in homogeneous coordinates I=[λ·Ix:λ·Iy:λ].
  • Each element camera_parameters_*_* can be represented according to the IEEE single precision floating point (32 bits) standard.
  • The advantage of putting this information in a separate parameter set is that we still maintain Sequence Parameter Sets (SPS) and Picture Parameter Sets (PPS) that are compatible with the MPEG-4 AVC standard. If we put this information in a Sequence Parameter Set or a Picture Parameter Set then, for each view, we need to send a separate Sequence Parameter Set and Picture Parameter Set. This is too restrictive. Also, this information does not fit well in either a Sequence Parameter Set or a Picture Parameter Set. Another reason is that since we propose to have an MPEG-4 AVC standard compatible base view we would have to use separate (MPEG-4 AVC compatible) Sequence Parameter Sets and Picture Parameter Sets for such a view and a separate Sequence Parameter Sets/Picture Parameter Sets (with view specific information) for all other views.
  • Placing all the dependency information in a single View Parameter Set at the very beginning of the sequence is very beneficial. The decoder can create a map using all the dependency information once it receives the View Parameter Set. This enables it to know before it receives any slice which views are needed for decoding a particular view. As a result of this, we only need to parse the slice header to obtain the view_id and determine if this view is needed to decode a target view as indicated by a user. Thus, we do not need to buffer any frames or wait until a certain point to determine which frames are needed for decoding a particular view.
  • The dependency information and whether it is a base view is indicated in the View Parameter Set. Even an MPEG-4 AVC compatible base view has associated with it information that is specific to that view (e.g., camera parameters). This information may be used by other views for several purposes including view interpolation/synthesis. We propose to support only one MPEG-4 AVC compatible view since if there are multiple MPEG-4 AVC compatible views; this makes it difficult to identify for each such slice which view it belongs to and a non-Multi-view Video Coding decoder can easily get confused.
  • By restricting it to just one such view, it is guaranteed that a non-Multi-view Video Coding decoder will be able to correctly decode the view and a Multi-view Video Coding decoder can easily identify such a view from the View Parameter Set using the syntax avc_compatible_view_id. All other base views (non-MPEG-4 AVC compatible) can be identified using the is_base_view_flag.
  • A new slice header for Multi-view Video Coding slices is proposed. In order to support view scalability, view random access, and so forth, we need to know which views the current slice depends upon. For view synthesis and view interpolation we may potentially also need camera parameters. This information is present in the View Parameter Set as shown above in TABLE 1. The View Parameter Set is identified using the view_parameter_set_id. We propose to add the view_parameter_set_id in the slice header of all the non-MPEG-4 AVC compatible slices as shown in TABLE 3 (Slice Header Syntax). The view_id information is needed for several Multi-view Video Coding requirements including view interpolation/synthesis, view random access, parallel processing, and so forth. This information can also be useful for special coding modes that only relate to cross-view prediction. In order to find the corresponding parameters from the View Parameter Set for this view, we need to send the view_id in the slice header.
  • TABLE 3
    slice_header( ) { C Descriptor
    first_mb_in_slice
    2 ue(v)
    slice_type 2 ue(v)
    pic_parameter_set_id 2 ue(v)
    if (nal_unit_type == 22 ∥ nal_unit_type == 23) {
    view_parameter_set_id 2 ue(v)
    view_id 2 ue(v)
    }
    frame_num 2 u(v)
    if( !frame_mbs_only_flag ) {
    field_pic_flag 2 u(1)
    if( field_pic_flag )
    bottom_field_flag 2 u(1)
    }
    ........
    }
  • For the new Multi-view Video Coding slices we propose to create new NAL unit types for each slice type (Instantaneous Decoding Refresh (IDR) and non-IDR). We propose to use type 22 for IDR slices and type 23 for non-IDR slices as shown in TABLE 2.
  • view_parameter_set_id specifies the view parameter set in use. The value of the view_parameter_set_id shall be in the range 0 to 255.
  • view_id indicates the view id of the current view. The value of the view_parameter_set_id shall be in the range 0 to 255.
  • An example of view random access will now be described in accordance with an embodiment of the present principles.
  • View random access is a Multi-view Video Coding requirement. The goal is to get access to any view with minimum decoding effort. Let us consider a simple example of view random access for the prediction structure shown in FIG. 3.
  • Suppose a user requests to decode view S3. From FIG. 3, we see that this view depends on view S0, view S2, and view S4. An example View Parameter Set is illustrated below.
  • Let us presume that the view_id for the views are numbered consecutively from 0 to 7 in the slice header syntax and there is only one View Parameter Set present with view_parameter_set equal to 0. number_of_views_minus 1 is set to 7. avc_compatible_view_id could be set to 0.
  • For views S0, is_base_view_flag is set to 1 and for other views it is set to 0. The dependency map for S0, S1, S2, S3, and S4 will look as shown in TABLE 4A (Dependency table for S0 anchor_picture_dependency_map) and TABLE 4B (dependency table for S0 non_anchor_picture_dependency_map). The dependency map for the other views can be written in a similar way.
  • Once this table is available at the decoder, the decoder can easily determine if a slice it receives is needed to decode a particular view. The decoder only needs to parse the slice header to determine the view_id of the current slice and for the target view S3 it can look up the S3 columns in the two tables (TABLE 4a and TABLE 4B) to determine whether or not it should keep the current slice. The decoder needs to distinguish between anchor pictures and non-anchor pictures since they may have different dependencies as can be seen from TABLE 4a and TABLE 4b. For the target view S3, we need to decode the anchor pictures of views S0, S2, and S4 but only need to decode the non-anchor pictures of views S2 and S4.
  • TABLE 4A
    j
    i S0 S1 S2 S3 S4 S5 S6 S7
    S0
    0 1 1 1 1 1 1 1
    S1 0 0 0 0 0 0 0 0
    S2 0 1 0 1 1 1 1 1
    S3 0 0 0 0 0 0 0 0
    S4 0 0 0 1 0 1 1 1
  • TABLE 4B
    j
    i S0 S1 S2 S3 S4 S5 S6 S7
    S0
    0 1 0 0 0 0 0 0
    S1 0 0 0 0 0 0 0 0
    S2 0 1 0 1 0 0 0 0
    S3 0 0 0 0 0 0 0 0
    S4 0 0 0 1 0 1 0 0
  • Turning to FIG. 4, an exemplary method for encoding multiple views of multi-view video content is indicated generally by the reference numeral 400.
  • The method 400 includes a start block 405 that passes control to a function block 410. The function block 410 reads a configuration file for the encoding parameters to be used to encode the multiple views, and passes control to a function block 415. The function block sets N to be equal to the number of views to be encoded, and passes control to a function block 420. The function block 420 sets number_of_views_minus1 equal to N−1, sets avc_compatible_view_id equal to the view_id of the MPEG-4 AVC compatible view, and passes control to a function block 425. The function block 425 sets view_parameter_set_id equal to a valid integer, initializes a variable i to be equal to zero, and passes control to a decision block 430. The decision block 430 determines whether or not i is greater than N. If so, then control is passed to a decision block 435. Otherwise, control is passed to a function block 470.
  • The decision block 435 determines whether or not the current view is a base view. If so, then control is passed to a function block 440. Otherwise, control is passed to a function block 480.
  • The function block 440 sets is_base_view_flag[i] equal to one, and passes control to a decision block 445. The decision block 445 determines whether or not the dependency is being updated. If so, the control is passed to a function block 450. Otherwise, control is passed to a function block 485.
  • The function block 450 sets dependency_update_flag equal to one, and passes control to a function block 455. The function block 455 sets a variable j equal to 0, and passes control to a decision block 460. The decision block 460 determines whether or not j is less than N. If so, then control is passed to a function block 465. Otherwise, control is passed to the function block 487.
  • The function block 465 sets anchor_picture_dependency_maps[i][j] and non_anchor_picture_dependency_maps[i][j] to values indicated by configuration file, and passes control to a function block 467. The function block 467 increments the variable j by one, and returns control to the decision block 460.
  • The function block 470 sets camera_parameters_present_flag equal to one when camera parameters are present, sets camera_parameters_present_flag equal to zero otherwise, and passes control to a decision block 472. The decision block 472 determines whether or not camera_parameters_present_flag is equal to one. If so, then control is passed to a function block 432. Otherwise, control is passed to a function block 434.
  • The function block 432 writes the camera parameters, and passes control to the function block 434.
  • The function block 434 writes the View Parameter Set (VPS) or the Sequence Parameter Set (SPS), and passes control to an end block 499.
  • The function block 480 sets is_base_view_flag[i] equal to zero, and passes control to the decision block 445.
  • The function block 485 sets dependency_update_flag equal to zero, and passes control to a function block 487. The function block 487 increments the variable i by 1, and returns control to the decision block 430.
  • Turning to FIG. 5, an exemplary method for decoding multiple views of multi-view video content is indicated generally by the reference numeral 500.
  • The method 500 includes a start block 505 that passes control to a function block 510. The function block 510 parses a Sequence Parameter Set (SPS) or View Parameter Set (VPS), view_parameter_set_id, number_of_views_minus 1, avc_compatible_view_id, sets variables I and j equal to zero, sets N equal to number_of_views_minus 1, and passes control to a decision block 515. The decision block 515 determines whether or not i is less than or equal to N. If so, then control is passed to a function block 570. Otherwise, control is passed to a function block 525.
  • The function block 570 parses camera_parameters_present_flag, and passes control to a decision block 572. The decision block 572 determines whether or not camera_parameters_present_flag is equal to one. If so, then control is passed to a function block 574. Otherwise, control is passed to a function block 576.
  • The function block 574 parses the camera parameters, and passes control to the function block 576.
  • The function block 576 continues decoding, and passes control to an end block 599.
  • The function block 525 parses is_base_view_flag[i] and dependency_update_flag, and passes control to a decision block 530. The decision block 530 determines whether or not dependency_update_flag is equal to zero. If so, then control is passes to a function block 532. Otherwise, control is passed to a decision block 535.
  • The function block 532 increments i by one, and returns control to the decision block 515.
  • The decision block 535 determines whether or not j is less than or equal to N. If so, then control is passed to a function block 540. Otherwise, control is passes to a function block 537.
  • The function block 540 parses anchor_picture_dependency_maps[i][j], and passes control to a decision block 545. The decision block 545 determines whether or not non_anchor_picture_dependency_maps[i][j] is equal to one. If so, then control is passed to a function block 550. Otherwise, control is passes to a function block 547.
  • The function block 550 parses the non_anchor_picture_dependency_maps[i][j], and passes control to the function block 547.
  • The function block 547 increments j by one, and returns control to the decision block 535.
  • The function block 537 increments i by one, and returns control to the function block 515.
  • A description will now be given of some of the many attendant advantages/features of the present invention, some of which have been mentioned above. For example, one advantage/feature is an apparatus that includes an encoder for encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include view specific information. The view specific information indicates a decoding interdependency between at least some of the at least two views. Another advantage/feature is the apparatus having the encoder as described above, wherein the decoding interdependency allows a corresponding decoding of at least one of the at least two views using only a subset of the at least two views for the corresponding decoding. Yet another advantage/feature is the apparatus having the encoder as described above, wherein the decoding interdependency indicated in the view specific information is used for random access of at least one of the at least two views by dropping slices related to any other ones of the at least two views indicated as non-interdependent with respect to the at least one view. Still another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information is included in a high level syntax. A further advantage/feature is the apparatus having the encoder as described above, wherein the view specific information is included in a parameter set compliant with the International Organization for Standardization/international Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/international Telecommunication Union, Telecommunication Sector H.264 recommendation. A yet further advantage/feature is the apparatus having the encoder as described above, wherein the view specific information is included in a View Parameter Set. A still further advantage/feature is the apparatus having the encoder wherein the view specific information is included in a View Parameter Set as described above, wherein the View Parameter Set is assigned a NAL unit type specifically for use only with the View Parameter Set. An additional advantage/feature is the apparatus having the encoder wherein a NAL unit type is assigned specifically for use only with the View Parameter Set as described above, wherein the NAL unit type is 14. Moreover, another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information includes at least one syntax element for indicating a View Parameter Set id. Further, another advantage/feature is the apparatus having the encoder wherein the view specific information includes at least one syntax element for indicating a View Parameter Set id as described above, wherein the at least one syntax element is denoted by a view_parameter_set_id syntax element. Also, another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information includes at least one syntax element for indicating a number of views. Additionally, another advantage/feature is the apparatus having the encoder wherein the view specific information includes at least one syntax element for indicating a number of views as described above, wherein the at least one syntax element is denoted by a number_of_views_minus 1 syntax element. Moreover, another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information includes at least one syntax element for indicating a view id for a particular one of the at least two views, when the particular one of the at least two views is encoded in a resultant bitstream that is compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/international Telecommunication Union, Telecommunication Sector H.264 recommendation. Further, another advantage/feature is the apparatus having the encoder wherein the view specific information includes at least one syntax element for indicating a view id for a particular one of the at least two views as described above, wherein the at least one syntax element is de noted by an avc_compatible_view_id syntax element. Also, another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information includes at least one syntax element or is implicitly derivable from a high level syntax, the at least one syntax element and the high level syntax for indicating that a particular one of the at least two views is compatible with the International Organization for Standardization/international Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation. Additionally, another advantage/feature is the apparatus having the encoder wherein the view specific information includes at least one syntax element or is implicitly derivable from a high level syntax as described above, wherein the at least one syntax element is denoted by an is_base_view_flag syntax element. Moreover, another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information includes at least one syntax element for indicating whether dependency information for at least one of the at least two views is present in the resultant bitstream. Further, another advantage/feature is the apparatus having the encoder wherein the view specific information includes at least one syntax element for indicating whether dependency information for at least one of the at least two views is present in the resultant bitstream as described above, wherein the at least one syntax element is denoted by a dependency_update_flag syntax element. Also, another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information includes at least one syntax element for indicating whether at least one anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views. Moreover, another advantage/feature is the apparatus having the encoder wherein the view specific information includes at least one syntax element for indicating whether at least one anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views as described above, wherein the at least one syntax element is denoted by an anchor_picture_dependency_maps[i][j] syntax element. Also, another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information includes at least one syntax element for indicating whether at least one non-anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views. Additionally, another advantage/feature is the apparatus having the encoder wherein the view specific information includes at least one syntax element for indicating whether at least one non-anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views as described above, wherein the at least one syntax element is denoted by a non_anchor_picture_dependency_maps[i][j] syntax element. Moreover, another advantage/feature is the apparatus having the encoder as described above, wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views. Further, another advantage/feature is the apparatus having the encoder wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views as described above, wherein the at least one syntax is included in a parameter set corresponding to the resultant bitstream.
  • Moreover, another advantage/feature is an apparatus that includes an encoder for encoding at least two views corresponding to multi-view video content by defining as a base view any of the at least two views that, for a decoding thereof, is independent of any other of the at least two views.
  • Further, another advantage/feature is an apparatus that includes an encoder for encoding at least two views corresponding to multi-view video content by encoding at least one of the at least two views in a resultant bitstream that is syntax compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation, for backwards compatibility therewith. Also, another advantage/feature is the apparatus having the encoder as described above, wherein the at least one view is a base view that, for a decoding thereof, is independent of any other of the at least two views. Additionally, another advantage/feature is the apparatus having the encoder as described above, wherein an avc_compatible_view_id syntax element identifies the at least one view as being encoded in the resultant bitstream that is syntax compliant with the International Organization for Standardization/international Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation for backwards compatibility.
  • Moreover, another advantage/feature is an apparatus that includes an encoder for encoding at least one of at least two views corresponding to multi-view video content by selecting between one of two-pre-defined slice types. Further, another advantage/feature is the apparatus having the encoder as described above, wherein the two pre-defined slice types are an Instantaneous Decoding Refresh slice type and a non-Instantaneous Decoding Refresh slice type. Also, another advantage/feature is the apparatus having the encoder that selects between the Instantaneous Decoding Refresh slice type and the non-Instantaneous Decoding Refresh slice type as described above, wherein NAL unit type 22 is used for the Instantaneous Decoding Refresh slice type and NAL unit type 23 is used for the non-Instantaneous Decoding Refresh slices. Additionally, another advantage/feature is the apparatus having the encoder as described above, wherein slice headers for at least one of the at least two slices includes view specific syntax. Moreover, another advantage/feature is the apparatus having the encoder wherein slice headers for at least one of the at least two slices includes view specific syntax as described above, wherein the view specific syntax is conditioned on NAL unit type 23 and NAL unit type 24. Further, another advantage/feature is the apparatus having the encoder wherein slice headers for at least one of the at least two slices includes view specific syntax as described above, wherein the view specific syntax includes a view parameter set identifier and a view identifier. Also, another advantage/feature is the apparatus having the encoder wherein the view specific syntax includes a view parameter set identifier and a view identifier as described above, wherein the view parameter set identifier is denoted by a view_parameter_set_id syntax element and the view identifier is denoted by a view_id syntax element.
  • Additionally, another advantage/feature is an apparatus that includes an encoder for encoding at least two views corresponding to multi-view content into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one camera parameter corresponding to at least one of the at least two views. Moreover, another advantage/feature is the apparatus having the encoder as described above, wherein the resultant bitstream is encoded to include a View Parameter Set, and the at least one camera parameter is included in the View Parameter Set. Further, another advantage/feature is the apparatus having the encoder as described above, wherein a presence of the at least one camera parameter is indicated by a syntax element. Also, another advantage/feature is the apparatus having the encoder wherein a presence of the at least one camera parameter is indicated by a syntax element as described above, wherein the syntax element is a camera_parameters_present_flag syntax element. Additionally, another advantage/feature is the apparatus having the encoder as described above, wherein the at least one camera parameter is denoted by a camera_parameters syntax element.
  • Moreover, another advantage/feature is an apparatus that includes an encoder for encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views. Further, another advantage/feature is the apparatus having the encoder as described above, wherein the at least one syntax is a high level syntax element. Also, another advantage/feature is the apparatus having the encoder as described above, wherein the at least one syntax is included in a parameter set corresponding to the resultant bitstream.
  • These and other features and advantages of the present principles may be readily ascertained by one of ordinary skill in the pertinent art based on the teachings herein. It is to be understood that the teachings of the present principles may be implemented in various forms of hardware, software, firmware, special purpose processors, or combinations thereof.
  • Most preferably, the teachings of the present principles are implemented as a combination of hardware and software. Moreover, the software may be implemented as an application program tangibly embodied on a program storage unit. The application program may be uploaded to, and executed by, a machine comprising any suitable architecture. Preferably, the machine is implemented on a computer platform having hardware such as one or more central processing units (“CPU”), a random access memory (“RAM”), and input/output (“I/O”) interfaces. The computer platform may also include an operating system and microinstruction code. The various processes and functions described herein may be either part of the microinstruction code or part of the application program, or any combination thereof, which may be executed by a CPU. In addition, various other peripheral units may be connected to the computer platform such as an additional data storage unit and a printing unit.
  • It is to be further understood that, because some of the constituent system components and methods depicted in the accompanying drawings are preferably implemented in software, the actual connections between the system components or the process function blocks may differ depending upon the manner in which the present principles are programmed. Given the teachings herein, one of ordinary skill in the pertinent art will be able to contemplate these and similar implementations or configurations of the present principles.
  • Although the illustrative embodiments have been described herein with reference to the accompanying drawings, it is to be understood that the present principles is not limited to those precise embodiments, and that various changes and modifications may be effected therein by one of ordinary skill in the pertinent art without departing from the scope or spirit of the present principles. All such changes and modifications are intended to be included within the scope of the present principles as set forth in the appended claims.

Claims (112)

1. An apparatus, comprising:
an encoder for encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include view specific information, the view specific information indicating a decoding interdependency between at least some of the at least two views.
2. The apparatus of claim 1, wherein the decoding interdependency allows a corresponding decoding of at least one of the at least two views using only a subset of the at least two views for the corresponding decoding.
3. The apparatus of claim 1, wherein the decoding interdependency indicated in the view specific information is used for random access of at least one of the at least two views by dropping slices related to any other ones of the at least two views indicated as non-interdependent with respect to the at least one view.
4. The apparatus of claim 1, wherein the view specific information is included in a high level syntax.
5. The apparatus of claim 1, wherein the view specific information is included in a parameter set compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation.
6. The apparatus of claim 1, wherein the view specific information is included in a View Parameter Set.
7. The apparatus of claim 6, wherein the View Parameter Set is assigned a NAL unit type specifically for use only with the View Parameter Set.
8. The apparatus of claim 7, wherein the NAL unit type is 14.
9. The apparatus of claim 1, wherein the view specific information includes at least one syntax element for indicating a View Parameter Set id.
10. The apparatus of claim 9, wherein the at least one syntax element is denoted by a view_parameter_set_id syntax element.
11. The apparatus of claim 1, wherein the view specific information includes at least one syntax element for indicating a number of views.
12. The apparatus of claim 11, wherein the at least one syntax element is denoted by a number_of_views_minus1 syntax element.
13. The apparatus of claim 1, wherein the view specific information includes at least one syntax element for indicating a view id for a particular one of the at least two views, when the particular one of the at least two views is encoded in a resultant bitstream that is compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/international Telecommunication Union, Telecommunication Sector H.264 recommendation.
14. The apparatus of claim 13, wherein the at least one syntax element is denoted by an avc_compatible_view_id syntax element.
15. The apparatus of claim 1, wherein the view specific information includes at least one syntax element or is implicitly derivable from a high level syntax, the at least one syntax element and the high level syntax for indicating that a particular one of the at least two views is compatible with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation.
16. The apparatus of claim 15, wherein the at least one syntax element is denoted by an is_base_view_flag syntax element.
17. The apparatus of claim 1, wherein the view specific information includes at least one syntax element for indicating whether dependency information for at least one of the at least two views is present in the resultant bitstream.
18. The apparatus of claim 17, wherein the at least one syntax element is denoted by a dependency_update_flag syntax element.
19. The apparatus of claim 1, wherein the view specific information includes at least one syntax element for indicating whether at least one anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views.
20. The apparatus of claim 19, wherein the at least one syntax element is denoted by an anchor_picture_dependency_maps[i][j] syntax element.
21. The apparatus of claim 1, wherein the view specific information includes at least one syntax element for indicating whether at least one non-anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views.
22. The apparatus of claim 21, wherein the at least one syntax element is denoted by a non_anchor_picture_dependency_maps[i][j] syntax element.
23. The apparatus of claim 1, wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views.
24. The apparatus of claim 23, wherein the at least one syntax is included in a parameter set corresponding to the resultant bitstream.
25. A method, comprising:
encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include view specific information, the view specific information indicating a decoding interdependency between at least some of the at least two views.
26. The method of claim 25, wherein the decoding interdependency allows a corresponding decoding of at least one of the at least two views using only a subset of the at least two views for the corresponding decoding.
27. The method of claim 25, wherein the decoding interdependency indicated in the view specific information is used for random access of at least one of the at least two views by dropping slices related to any other ones of the at least two views indicated as non-interdependent with respect to the at least one view.
28. The method of claim 25, wherein the view specific information is included in a high level syntax.
29. The method of claim 25, wherein the view specific information is included in a parameter set compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation.
30. The method of claim 25, wherein the view specific information is included in a View Parameter Set.
31. The method of claim 30, wherein the View Parameter Set is assigned a NAL unit type specifically for use only with the View Parameter Set.
32. The method of claim 31, wherein the NAL unit type is 14.
33. The method of claim 25, wherein the view specific information includes at least one syntax element for indicating a View Parameter Set id.
34. The method of claim 33, wherein the at least one syntax element is denoted by a view_parameter_set_id syntax element.
35. The method of claim 25, wherein the view specific information includes at least one syntax element for indicating a number of views.
36. The method of claim 35, wherein the at least one syntax element is denoted by a number_of_views_minus1 syntax element.
37. The method of claim 25, wherein the view specific information includes at least one syntax element for indicating a view id for a particular one of the at least two views, when the particular one of the at least two views is encoded in a resultant bitstream that is compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation.
38. The method of claim 37, wherein the at least one syntax element is denoted by an avc_compatible_view_id syntax element.
39. The method of claim 25, wherein the view specific information includes at least one syntax element or is implicitly derivable from a high level syntax, the at least one syntax element and the high level syntax for indicating that a particular one of the at least two views is compatible with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation.
40. The method of claim 39, wherein the at least one syntax element is denoted by an is_base_view_flag syntax element.
41. The method of claim 25, wherein the view specific information includes at least one syntax element for indicating whether dependency information for at least one of the at least two views is present in the resultant bitstream.
42. The method of claim 41, wherein the at least one syntax element is denoted by a dependency_update_flag syntax element.
43. The method of claim 25, wherein the view specific information includes at least one syntax element for indicating whether at least one anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views.
44. The method of claim 43, wherein the at least one syntax element is denoted by an anchor_picture_dependency_maps[i][j] syntax element.
45. The method of claim 25, wherein the view specific information includes at least one syntax element for indicating whether at least one non-anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views.
46. The method of claim 45, wherein the at least one syntax element is denoted by a non_anchor_picture_dependency_maps[i][j] syntax element.
47. The method of claim 25, wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views.
48. The method of claim 47, wherein the at least one syntax is included in a parameter set corresponding to the resultant bitstream.
49. An apparatus, comprising:
a decoder for decoding at least two views corresponding to multi-view video content from a bitstream, wherein the bitstream is decoded to determine view specific information included therein, the view specific information indicating a decoding interdependency between at least some of the at least two views.
50. The apparatus of claim 49, wherein the decoding interdependency allows said decoder to decode at least one of the at least two views using only a subset of the at least two views.
51. The apparatus of claim 49, wherein the decoding interdependency indicated in the view specific information is used for random access of at least one of the at least two views by dropping slices related to any other ones of the at least two views indicated as non-interdependent with respect to the at least one view.
52. The apparatus of claim 49, wherein the view specific information is included in a high level syntax.
53. The apparatus of claim 49, wherein the view specific information is included in a parameter set compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation.
54. The apparatus of claim 49, wherein the view specific information is included in a View Parameter Set.
55. The apparatus of claim 54, wherein the View Parameter Set is assigned a NAL unit type specifically for use only with the View Parameter Set.
56. The apparatus of claim 55, wherein the NAL unit type is 14.
57. The apparatus of claim 49, wherein the view specific information includes at least one syntax element for indicating a View Parameter Set id.
58. The apparatus of claim 57, wherein the at least one syntax element is denoted by a view_parameter_set_id syntax element.
59. The apparatus of claim 49, wherein the view specific information includes at least one syntax element for indicating a number of views.
60. The apparatus of claim 59, wherein the at least one syntax element is denoted by a number_of_views_minus1 syntax element.
61. The apparatus of claim 49, wherein the view specific information includes at least one syntax element for indicating a view id for a particular one of the at least two views, when the particular one of the at least two views is encoded in a resultant bitstream that is compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation.
62. The apparatus of claim 61, wherein the at least one syntax element is denoted by an avc_compatible_view_id syntax element.
63. The apparatus of claim 49, wherein the view specific information includes at least one syntax element or is implicitly derived from a high level syntax, the at least one syntax element and the high level syntax for indicating that a particular one of the at least two views is compatible with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation.
64. The apparatus of claim 63, wherein the at least one syntax element is denoted by an is_base_view_flag syntax element.
65. The apparatus of claim 49, wherein the view specific information includes at least one syntax element for indicating whether dependency information for at least one of the at least two views is present in the bitstream.
66. The apparatus of claim 65, wherein the at least one syntax element is denoted by a dependency_update_flag syntax element.
67. The apparatus of claim 49, wherein the view specific information includes at least one syntax element for indicating whether at least one anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views.
68. The apparatus of claim 67, wherein the at least one syntax element is denoted by an anchor_picture_dependency_maps[i][j] syntax element.
69. The apparatus of claim 49, wherein the view specific information includes at least one syntax element for indicating whether at least one non-anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views.
70. The apparatus of claim 69, wherein the at least one syntax element is denoted by a non_anchor_picture_dependency_maps[i][j] syntax element.
71. The apparatus of claim 49, wherein said decoder determines, from the bitstream, at least one syntax element related to at least one camera parameter for at least one of the at least two views.
72. The apparatus of claim 71, wherein the at least one syntax is included in a parameter set corresponding to the bitstream
73. A method, comprising:
decoding at least two views corresponding to multi-view video content from a bitstream, wherein the bitstream is decoded to determine view specific information included therein, the view specific information indicating a decoding interdependency between at least some of the at least two views.
74. The method of claim 73, wherein the decoding interdependency allows said decoding step to decode at least one of the at least two views using only a subset of the at least two views.
75. The method of claim 73, wherein the decoding interdependency indicated in the view specific information is used for random access of at least one of the at least two views by dropping slices related to any other ones of the at least two views indicated as non-interdependent with respect to the at least one view.
76. The method of claim 73, wherein the view specific information is included in a high level syntax.
77. The method of claim 73, wherein the view specific information is included in a parameter set compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation.
78. The method of claim 73, wherein the view specific information is included in a View Parameter Set.
79. The method of claim 78, wherein the View Parameter Set is assigned a NAL unit type specifically for use only with the View Parameter Set.
80. The method of claim 79, wherein the NAL unit type is 14.
81. The method of claim 73, wherein the view specific information includes at least one syntax element for indicating a View Parameter Set id.
82. The method of claim 81, wherein the at least one syntax element is denoted by a view_parameter_set_id syntax element.
83. The method of claim 73, wherein the view specific information includes at least one syntax element for indicating a number of views.
84. The method of claim 83, wherein the at least one syntax element is denoted by a number_of_views_minus1 syntax element.
85. The method of claim 73, wherein the view specific information includes at least one syntax element for indicating a view id for a particular one of the at least two views, when the particular one of the at least two views is encoded in a resultant bitstream that is compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation.
86. The method of claim 85, wherein the at least one syntax element is denoted by an avc_compatible_view_id syntax element.
87. The method of claim 73, wherein the view specific information includes at least one syntax element or is implicitly derived from a high level syntax, the at least one syntax element and the high level syntax for indicating that a particular one of the at least two views is compatible with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation.
88. The method of claim 87, wherein the at least one syntax element is denoted by an is_base_view_flag syntax element.
89. The method of claim 73, wherein the view specific information includes at least one syntax element for indicating whether dependency information for at least one of the at least two views is present in the bitstream.
90. The method of claim 89, wherein the at least one syntax element is denoted by a dependency_update_flag syntax element.
91. The method of claim 73, wherein the view specific information includes at least one syntax element for indicating whether at least one anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views.
92. The method of claim 91, wherein the at least one syntax element is denoted by an anchor_picture_dependency_maps[i][j] syntax element.
93. The method of claim 73, wherein the view specific information includes at least one syntax element for indicating whether at least one non-anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views.
94. The method of claim 93, wherein the at least one syntax element is denoted by a non_anchor_picture_dependency_maps[i][j] syntax element.
95. The method of claim 73, wherein said decoding step determines, from the bitstream, at least one syntax element related to at least one camera parameter for at least one of the at least two views.
96. The method of claim 95, wherein the at least one syntax is included in a parameter set corresponding to the bitstream.
97. A video signal structure for video encoding, comprising:
at least two views corresponding to multi-view video content encoded into a resultant bitstream, wherein the resultant bitstream is encoded to include view specific information, wherein the view specific information indicates a decoding interdependency between at least some of the at least two views.
98. A storage media having video signal data encoded thereupon, comprising:
at least two views corresponding to multi-view video content encoded into a resultant bitstream, wherein the resultant bitstream is encoded to include view specific information, wherein the view specific information indicates a decoding interdependency between at least some of the at least two views.
99. An apparatus, comprising:
an encoder for encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views.
100. The apparatus of claim 99, wherein the at least one syntax is a high level syntax element.
101. The apparatus of claim 99, wherein the at least one syntax is included in a parameter set corresponding to the resultant bitstream.
102. A method, comprising:
encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views.
103. The apparatus of claim 102, wherein the at least one syntax is a high level syntax element.
104. The apparatus of claim 102, wherein the at least one syntax is included in a parameter set corresponding to the resultant bitstream.
105. An apparatus, comprising:
a decoder for decoding at least two views corresponding to multi-view video content from a bitstream, wherein the bitstream is decoded to determine at least one camera parameter for at least one of the at least two views based on at least one syntax element included in the bitstream.
106. The apparatus of claim 105, wherein the at least one syntax is a high level syntax element.
107. The apparatus of claim 105, wherein the at least one syntax is included in a parameter set corresponding to the resultant bitstream.
108. A method, comprising:
decoding at least two views corresponding to multi-view video content from a bitstream, wherein the bitstream is decoded to determine at least one camera parameter for at least one of the at least two views based on at least one syntax element included in the bitstream.
109. The method of claim 108, wherein the at least one syntax is a high level syntax element.
110. The method of claim 108, wherein the at least one syntax is included in a parameter set corresponding to the resultant bitstream.
111. A video signal structure for video encoding, comprising:
at least two views corresponding to multi-view video content encoded into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views.
112. A storage media having video signal data encoded thereupon, comprising:
at least two views corresponding to multi-view video content encoded into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views.
US12/224,816 2006-03-29 2007-02-27 Multi-View Video Coding Method and Device Abandoned US20090225826A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/224,816 US20090225826A1 (en) 2006-03-29 2007-02-27 Multi-View Video Coding Method and Device

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US78709206P 2006-03-29 2006-03-29
US12/224,816 US20090225826A1 (en) 2006-03-29 2007-02-27 Multi-View Video Coding Method and Device
PCT/US2007/004972 WO2007126509A2 (en) 2006-03-29 2007-02-27 Multi-view video coding method and device

Publications (1)

Publication Number Publication Date
US20090225826A1 true US20090225826A1 (en) 2009-09-10

Family

ID=38515387

Family Applications (3)

Application Number Title Priority Date Filing Date
US12/224,817 Abandoned US20090207904A1 (en) 2006-03-29 2007-02-27 Multi-View Video Coding Method and Device
US12/224,816 Abandoned US20090225826A1 (en) 2006-03-29 2007-02-27 Multi-View Video Coding Method and Device
US12/224,814 Expired - Fee Related US9100659B2 (en) 2006-03-29 2007-02-27 Multi-view video coding method and device using a base view

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US12/224,817 Abandoned US20090207904A1 (en) 2006-03-29 2007-02-27 Multi-View Video Coding Method and Device

Family Applications After (1)

Application Number Title Priority Date Filing Date
US12/224,814 Expired - Fee Related US9100659B2 (en) 2006-03-29 2007-02-27 Multi-view video coding method and device using a base view

Country Status (11)

Country Link
US (3) US20090207904A1 (en)
EP (3) EP1999968A2 (en)
JP (8) JP5213064B2 (en)
KR (3) KR101383735B1 (en)
CN (3) CN101416519B (en)
AU (2) AU2007243935A1 (en)
BR (3) BRPI0708305A2 (en)
MX (2) MX2008012382A (en)
RU (2) RU2529881C2 (en)
WO (3) WO2007126508A2 (en)
ZA (2) ZA200807142B (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080198924A1 (en) * 2007-02-06 2008-08-21 Gwangju Institute Of Science And Technology Method of computing disparity, method of synthesizing interpolation view, method of encoding and decoding multi-view video using the same, and encoder and decoder using the same
US20090279608A1 (en) * 2006-03-30 2009-11-12 Lg Electronics Inc. Method and Apparatus for Decoding/Encoding a Video Signal
US20100046619A1 (en) * 2006-08-25 2010-02-25 Han Suh Koo Method and apparatus for decoding/encoding a video signal
US20100091845A1 (en) * 2006-03-30 2010-04-15 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100177824A1 (en) * 2006-06-19 2010-07-15 Han Suh Koo Method and apparatus for processing a video signal
US20100266010A1 (en) * 2009-01-19 2010-10-21 Chong Soon Lim Coding method, decoding method, coding apparatus, decoding apparatus, program, and integrated circuit
US20110019746A1 (en) * 2009-04-28 2011-01-27 Chong Soon Lim Image decoding method, image coding method, image decoding apparatus, and image coding apparatus
US20110032999A1 (en) * 2009-08-07 2011-02-10 Ying Chen Signaling characteristics of an mvc operation point
US20110216827A1 (en) * 2010-02-23 2011-09-08 Jiancong Luo Method and apparatus for efficient encoding of multi-view coded video data

Families Citing this family (58)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101199498B1 (en) 2005-03-31 2012-11-09 삼성전자주식회사 Apparatus for encoding or generation of multi-view video by using a camera parameter, and a method thereof, and a recording medium having a program to implement thereof
US7714838B2 (en) * 2006-04-27 2010-05-11 Research In Motion Limited Handheld electronic device having hidden sound openings offset from an audio source
KR101450921B1 (en) * 2006-07-05 2014-10-15 톰슨 라이센싱 Methods and apparatus for multi-view video encoding and decoding
WO2008030067A1 (en) * 2006-09-07 2008-03-13 Lg Electronics Inc. Method and apparatus for decoding/encoding of a video signal
US8875199B2 (en) 2006-11-13 2014-10-28 Cisco Technology, Inc. Indicating picture usefulness for playback optimization
US8873932B2 (en) 2007-12-11 2014-10-28 Cisco Technology, Inc. Inferential processing to ascertain plural levels of picture interdependencies
US8155207B2 (en) 2008-01-09 2012-04-10 Cisco Technology, Inc. Processing and managing pictures at the concatenation of two video streams
US8416859B2 (en) * 2006-11-13 2013-04-09 Cisco Technology, Inc. Signalling and extraction in compressed video of pictures belonging to interdependency tiers
KR100896290B1 (en) * 2006-11-17 2009-05-07 엘지전자 주식회사 Method and apparatus for decoding/encoding a video signal
JP5108896B2 (en) * 2006-12-21 2012-12-26 トムソン ライセンシング Method and apparatus for improved signaling using high level syntax for multi-view video encoding and decoding
BR122018004904B1 (en) 2007-04-12 2019-11-05 Dolby Int Ab video coding and decoding tiling
BRPI0810584A2 (en) * 2007-04-25 2014-10-29 Thomson Licensing INTERVIEW PREDICTION
US8958486B2 (en) 2007-07-31 2015-02-17 Cisco Technology, Inc. Simultaneous processing of media and redundancy streams for mitigating impairments
US8804845B2 (en) 2007-07-31 2014-08-12 Cisco Technology, Inc. Non-enhancing media redundancy coding for mitigating transmission impairments
EP2198620A2 (en) * 2007-10-05 2010-06-23 Thomson Licensing Methods and apparatus for incorporating video usability information (vui) within a multi-view video (mvc) coding system
JP2011509631A (en) * 2008-01-11 2011-03-24 トムソン ライセンシング Video and depth encoding
US8416858B2 (en) 2008-02-29 2013-04-09 Cisco Technology, Inc. Signalling picture encoding schemes and associated picture properties
CN101562745B (en) * 2008-04-18 2012-07-04 华为技术有限公司 Method and device for encoding and decoding multi-viewpoint video image
KR101727311B1 (en) 2008-04-25 2017-04-14 톰슨 라이센싱 Multi-view video coding with disparity estimation based on depth information
US8886022B2 (en) 2008-06-12 2014-11-11 Cisco Technology, Inc. Picture interdependencies signals in context of MMCO to assist stream manipulation
US8705631B2 (en) 2008-06-17 2014-04-22 Cisco Technology, Inc. Time-shifted transport of multi-latticed video for resiliency from burst-error effects
US8699578B2 (en) 2008-06-17 2014-04-15 Cisco Technology, Inc. Methods and systems for processing multi-latticed video streams
US8971402B2 (en) 2008-06-17 2015-03-03 Cisco Technology, Inc. Processing of impaired and incomplete multi-latticed video streams
EP2356812B1 (en) 2008-11-12 2015-06-10 Cisco Technology, Inc. Processing of a video program having plural processed representations of a single video signal for reconstruction and output
AU2011250757B2 (en) * 2009-01-19 2012-09-06 Panasonic Intellectual Property Corporation Of America Coding method, decoding method, coding apparatus, decoding apparatus, program, and integrated circuit
KR101676059B1 (en) * 2009-01-26 2016-11-14 톰슨 라이센싱 Frame packing for video coding
CN102685514B (en) * 2009-02-19 2014-02-19 松下电器产业株式会社 Reproduction device, recording method and recording medium reproduction system
US8326131B2 (en) 2009-02-20 2012-12-04 Cisco Technology, Inc. Signalling of decodable sub-sequences
US8782261B1 (en) 2009-04-03 2014-07-15 Cisco Technology, Inc. System and method for authorization of segment boundary notifications
JP4962525B2 (en) * 2009-04-08 2012-06-27 ソニー株式会社 REPRODUCTION DEVICE, REPRODUCTION METHOD, AND PROGRAM
WO2010126613A2 (en) 2009-05-01 2010-11-04 Thomson Licensing Inter-layer dependency information for 3dv
US8949883B2 (en) 2009-05-12 2015-02-03 Cisco Technology, Inc. Signalling buffer characteristics for splicing operations of video streams
US8411746B2 (en) * 2009-06-12 2013-04-02 Qualcomm Incorporated Multiview video coding over MPEG-2 systems
US8780999B2 (en) 2009-06-12 2014-07-15 Qualcomm Incorporated Assembling multiview video coding sub-BITSTREAMS in MPEG-2 systems
US8279926B2 (en) 2009-06-18 2012-10-02 Cisco Technology, Inc. Dynamic streaming with latticed representations of video
JP5722349B2 (en) 2010-01-29 2015-05-20 トムソン ライセンシングThomson Licensing Block-based interleaving
US20110222837A1 (en) * 2010-03-11 2011-09-15 Cisco Technology, Inc. Management of picture referencing in video streams for plural playback modes
US9716920B2 (en) * 2010-08-05 2017-07-25 Qualcomm Incorporated Signaling attributes for network-streamed video data
WO2012036903A1 (en) 2010-09-14 2012-03-22 Thomson Licensing Compression methods and apparatus for occlusion data
US20130182074A1 (en) * 2010-10-13 2013-07-18 University-Industry Cooperation Group Of Kyung Hee University Method and apparatus for transmitting stereoscopic video information
MX2013005402A (en) * 2010-11-15 2013-07-03 Lg Electronics Inc Method for transforming frame format and apparatus using same method.
US9712804B2 (en) * 2010-12-03 2017-07-18 Lg Electronics Inc. Receiving device and method for receiving multiview three-dimensional broadcast signal
EP2654305A2 (en) * 2010-12-13 2013-10-23 Electronics And Telecommunications Research Institute Signaling method for a stereoscopic video service and apparatus using the method
US11496760B2 (en) 2011-07-22 2022-11-08 Qualcomm Incorporated Slice header prediction for depth maps in three-dimensional video codecs
US9635355B2 (en) 2011-07-28 2017-04-25 Qualcomm Incorporated Multiview video coding
US9674525B2 (en) 2011-07-28 2017-06-06 Qualcomm Incorporated Multiview video coding
EP2744201A4 (en) * 2011-08-09 2016-03-23 Samsung Electronics Co Ltd Method and device for encoding a depth map of multi viewpoint video data, and method and device for decoding the encoded depth map
EP2752011B1 (en) * 2011-08-31 2020-05-20 Nokia Technologies Oy Multiview video coding and decoding
US9258559B2 (en) 2011-12-20 2016-02-09 Qualcomm Incorporated Reference picture list construction for multi-view and three-dimensional video coding
US9451252B2 (en) 2012-01-14 2016-09-20 Qualcomm Incorporated Coding parameter sets and NAL unit headers for video coding
WO2013115562A1 (en) * 2012-01-30 2013-08-08 삼성전자 주식회사 Method and apparatus for multiview video encoding based on prediction structures for viewpoint switching, and method and apparatus for multiview video decoding based on prediction structures for viewpoint switching
TW201342884A (en) 2012-01-31 2013-10-16 Sony Corp Encoding device and encoding method, and decoding device and decoding method
KR20130116782A (en) * 2012-04-16 2013-10-24 한국전자통신연구원 Scalable layer description for scalable coded video bitstream
US10205961B2 (en) 2012-04-23 2019-02-12 Qualcomm Incorporated View dependency in multi-view coding and 3D coding
CN103379333B (en) * 2012-04-25 2018-12-04 浙江大学 The decoding method and its corresponding device of decoding method, video sequence code stream
US9565437B2 (en) 2013-04-08 2017-02-07 Qualcomm Incorporated Parameter set designs for video coding extensions
KR20160003070A (en) * 2013-07-19 2016-01-08 미디어텍 인크. Method and apparatus of camera parameter signaling in 3d video coding
CN104980763B (en) * 2014-04-05 2020-01-17 浙江大学 Video code stream, video coding and decoding method and device

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6055012A (en) * 1995-12-29 2000-04-25 Lucent Technologies Inc. Digital multi-view video compression with complexity and compatibility constraints
US6151362A (en) * 1998-10-30 2000-11-21 Motorola, Inc. Joint rate control for stereoscopic video coding
US6192186B1 (en) * 1997-11-06 2001-02-20 Sanyo Electric Co. Ltd Method and apparatus for providing/reproducing MPEG data
US20020012315A1 (en) * 2000-02-25 2002-01-31 Sony Corporation Recording medium, recording apparatus, and reading apparatus
US20060026646A1 (en) * 2004-07-27 2006-02-02 Microsoft Corporation Multi-view video format
US7903737B2 (en) * 2005-11-30 2011-03-08 Mitsubishi Electric Research Laboratories, Inc. Method and system for randomly accessing multiview videos with known prediction dependency
US7961786B2 (en) * 2003-09-07 2011-06-14 Microsoft Corporation Signaling field type information

Family Cites Families (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6161382A (en) * 1992-07-30 2000-12-19 Brotz; Gregory R. Thermoelectric actuator
US5557331A (en) * 1993-03-11 1996-09-17 Matsushita Electric Industrial Co., Ltd. Image encoding method, an image encoding circuit, an image encoding apparatus, and an optical disk
DE4331376C1 (en) * 1993-09-15 1994-11-10 Fraunhofer Ges Forschung Method for determining the type of encoding to selected for the encoding of at least two signals
US5771081A (en) * 1994-02-28 1998-06-23 Korea Telecommunication Authority Bit system for transmitting digital video data
US5619256A (en) * 1995-05-26 1997-04-08 Lucent Technologies Inc. Digital 3D/stereoscopic video compression technique utilizing disparity and motion compensated predictions
US5763943A (en) * 1996-01-29 1998-06-09 International Business Machines Corporation Electronic modules with integral sensor arrays
JPH09261653A (en) 1996-03-18 1997-10-03 Sharp Corp Multi-view-point picture encoder
KR980007751A (en) * 1996-06-26 1998-03-30 구자홍 Apparatus and method for parallel processing of MPEG-2 variable-length decoder
US6055274A (en) * 1997-12-30 2000-04-25 Intel Corporation Method and apparatus for compressing multi-view video
EP1034656A2 (en) 1998-06-11 2000-09-13 Koninklijke Philips Electronics N.V. Trick play signal generation for a digital video recorder
US6056012A (en) * 1999-02-25 2000-05-02 Ecolab Inc. Inline check valve
KR100433516B1 (en) * 2000-12-08 2004-05-31 삼성전자주식회사 Transcoding method
KR100433625B1 (en) 2001-11-17 2004-06-02 학교법인 포항공과대학교 Apparatus for reconstructing multiview image using stereo image and depth map
KR100446635B1 (en) * 2001-11-27 2004-09-04 삼성전자주식회사 Apparatus and method for depth image-based representation of 3-dimensional object
RU2237283C2 (en) 2001-11-27 2004-09-27 Самсунг Электроникс Ко., Лтд. Device and method for presenting three-dimensional object on basis of images having depth
US7292691B2 (en) * 2002-01-02 2007-11-06 Sony Corporation Progressive video refresh slice detection
KR100481732B1 (en) * 2002-04-20 2005-04-11 전자부품연구원 Apparatus for encoding of multi view moving picture
KR100475060B1 (en) * 2002-08-07 2005-03-10 한국전자통신연구원 The multiplexing method and its device according to user's request for multi-view 3D video
JP4045913B2 (en) * 2002-09-27 2008-02-13 三菱電機株式会社 Image coding apparatus, image coding method, and image processing apparatus
MY134659A (en) * 2002-11-06 2007-12-31 Nokia Corp Picture buffering for prediction references and display
ES2353872T3 (en) * 2003-01-28 2011-03-07 Thomson Licensing DIFFUSION SPACED IN ROBUST MODE.
US7778328B2 (en) * 2003-08-07 2010-08-17 Sony Corporation Semantics-based motion estimation for multi-view video coding
KR100965881B1 (en) * 2003-10-10 2010-06-24 삼성전자주식회사 System for encoding video data and system for decoding video data
KR100987775B1 (en) * 2004-01-20 2010-10-13 삼성전자주식회사 3 Dimensional coding method of video
JP2007525906A (en) 2004-02-27 2007-09-06 ティディヴィジョン コーポレイション エス.エー. デ シー.ヴィ. Stereo 3D video image digital coding system and method
KR100679740B1 (en) * 2004-06-25 2007-02-07 학교법인연세대학교 Method for Coding/Decoding for Multiview Sequence where View Selection is Possible
US7515759B2 (en) * 2004-07-14 2009-04-07 Sharp Laboratories Of America, Inc. 3D video coding using sub-sequences
US20060028846A1 (en) * 2004-08-06 2006-02-09 Hsiao-Chung Yang Connection device for solar panels in a solar powered lantern to enable thesolar panels to extend horizontally to the solar powered lantern
JP4638874B2 (en) 2004-08-11 2011-02-23 株式会社日立製作所 Coded stream recording medium and image coding apparatus
US7672378B2 (en) * 2005-01-21 2010-03-02 Stmicroelectronics, Inc. Spatio-temporal graph-segmentation encoding for multiple video streams
US8369406B2 (en) * 2005-07-18 2013-02-05 Electronics And Telecommunications Research Institute Apparatus of predictive coding/decoding using view-temporal reference picture buffers and method using the same
BRPI0617728A8 (en) * 2005-10-11 2016-05-10 Nokia Corp METHOD FOR ENABLING LOW TO HIGH RESIZABLE LAYER SWITCHING IN A CONTINUOUS STREAM OF VIDEO BITS; PRODUCT COMPUTER PROGRAM TO ENABLE RESIZABLE LOW TO HIGH LAYER SWITCHING IN A CONTINUOUS STREAM OF VIDEO BITS; ELECTRONIC DEVICE; AND ELEMENT OF A SYSTEM FOR SENDING INFORMATION IN PACKETS IN CONTINUOUS FLOW
KR101385015B1 (en) * 2006-03-02 2014-04-14 가부시키가이샤 다이센 고교 Foaming resin molder, and its running method

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6055012A (en) * 1995-12-29 2000-04-25 Lucent Technologies Inc. Digital multi-view video compression with complexity and compatibility constraints
US6192186B1 (en) * 1997-11-06 2001-02-20 Sanyo Electric Co. Ltd Method and apparatus for providing/reproducing MPEG data
US6151362A (en) * 1998-10-30 2000-11-21 Motorola, Inc. Joint rate control for stereoscopic video coding
US20020012315A1 (en) * 2000-02-25 2002-01-31 Sony Corporation Recording medium, recording apparatus, and reading apparatus
US7961786B2 (en) * 2003-09-07 2011-06-14 Microsoft Corporation Signaling field type information
US20060026646A1 (en) * 2004-07-27 2006-02-02 Microsoft Corporation Multi-view video format
US7903737B2 (en) * 2005-11-30 2011-03-08 Mitsubishi Electric Research Laboratories, Inc. Method and system for randomly accessing multiview videos with known prediction dependency

Cited By (104)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8565319B2 (en) 2006-03-30 2013-10-22 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US20100316136A1 (en) * 2006-03-30 2010-12-16 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20090296811A1 (en) * 2006-03-30 2009-12-03 Byeong Moon Jeon Method and Apparatus for Decoding/Encoding a Video Signal
US20100020870A1 (en) * 2006-03-30 2010-01-28 Byeong Moon Jeon Method and Apparatus for Decoding/Encoding a Video Signal
US20100026883A1 (en) * 2006-03-30 2010-02-04 Byeong Moon Jeon Method and Apparatus for Decoding/Encoding a Video Signal
US20100027682A1 (en) * 2006-03-30 2010-02-04 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100026882A1 (en) * 2006-03-30 2010-02-04 Byeong Moon Jeon Method and Apparatus for Decoding/Encoding a Video Signal
US20100027653A1 (en) * 2006-03-30 2010-02-04 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100027654A1 (en) * 2006-03-30 2010-02-04 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100027659A1 (en) * 2006-03-30 2010-02-04 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100026884A1 (en) * 2006-03-30 2010-02-04 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100027660A1 (en) * 2006-03-30 2010-02-04 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US8634475B2 (en) 2006-03-30 2014-01-21 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal using a reference picture list for inter-view prediction
US20100074334A1 (en) * 2006-03-30 2010-03-25 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100080293A1 (en) * 2006-03-30 2010-04-01 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100086036A1 (en) * 2006-03-30 2010-04-08 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100091884A1 (en) * 2006-03-30 2010-04-15 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100091885A1 (en) * 2006-03-30 2010-04-15 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100091886A1 (en) * 2006-03-30 2010-04-15 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100091845A1 (en) * 2006-03-30 2010-04-15 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100091843A1 (en) * 2006-03-30 2010-04-15 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100091883A1 (en) * 2006-03-30 2010-04-15 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100091844A1 (en) * 2006-03-30 2010-04-15 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US8611427B2 (en) 2006-03-30 2013-12-17 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US8611419B2 (en) 2006-03-30 2013-12-17 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US8576920B2 (en) 2006-03-30 2013-11-05 Lg Electronics, Inc. Method and apparatus for decoding/encoding a video signal
US20100316135A1 (en) * 2006-03-30 2010-12-16 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20090279608A1 (en) * 2006-03-30 2009-11-12 Lg Electronics Inc. Method and Apparatus for Decoding/Encoding a Video Signal
US8472519B2 (en) 2006-03-30 2013-06-25 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US8565303B2 (en) * 2006-03-30 2013-10-22 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US8559505B2 (en) 2006-03-30 2013-10-15 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US8559523B2 (en) 2006-03-30 2013-10-15 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US8526504B2 (en) 2006-03-30 2013-09-03 Lg Electronics Inc. Method and apparatus for decoding/encoding a multi-view video signal with inter-view reference picture list management
US8483273B2 (en) 2006-03-30 2013-07-09 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US8170108B2 (en) 2006-03-30 2012-05-01 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US8457207B2 (en) 2006-03-30 2013-06-04 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US8432972B2 (en) 2006-03-30 2013-04-30 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US8428130B2 (en) 2006-03-30 2013-04-23 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US20100316360A1 (en) * 2006-03-30 2010-12-16 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US8411744B2 (en) 2006-03-30 2013-04-02 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US8363732B2 (en) 2006-03-30 2013-01-29 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US7782944B2 (en) * 2006-03-30 2010-08-24 Lg Electronics, Inc. Method and apparatus for decoding/encoding a video signal
US7782946B2 (en) * 2006-03-30 2010-08-24 Lg Electronics, Inc. Method and apparatus for decoding/encoding a video signal
US7782950B2 (en) * 2006-03-30 2010-08-24 Lg Electronics, Inc. Method and apparatus for decoding/encoding a video signal
US7782947B2 (en) * 2006-03-30 2010-08-24 Lg Electronics, Inc. Method and apparatus for decoding/encoding a video signal
US7782943B2 (en) * 2006-03-30 2010-08-24 Lg Electronics, Inc. Method and apparatus for decoding/encoding a video signal
US7782949B2 (en) * 2006-03-30 2010-08-24 Lg Electronics, Inc. Method and apparatus for decoding/encoding a video signal
US7782948B2 (en) * 2006-03-30 2010-08-24 Lg Electronics, Inc. Method and apparatus for decoding/encoding a video signal
US7782945B2 (en) * 2006-03-30 2010-08-24 Lg Electronics, Inc. Method and apparatus for decoding/encoding a video signal
US20100215100A1 (en) * 2006-03-30 2010-08-26 Byeong Moon Jeon Method and Apparatus for Decoding/Encoding a Video Signal
US20100316361A1 (en) * 2006-03-30 2010-12-16 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100111169A1 (en) * 2006-03-30 2010-05-06 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US20100316362A1 (en) * 2006-03-30 2010-12-16 Byeong Moon Jeon Method and apparatus for decoding/encoding a video signal
US8571113B2 (en) 2006-03-30 2013-10-29 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US8325814B2 (en) 2006-06-19 2012-12-04 Lg Electronics Inc. Method and apparatus for processing a video signal
US20100177824A1 (en) * 2006-06-19 2010-07-15 Han Suh Koo Method and apparatus for processing a video signal
US20100111170A1 (en) * 2006-08-25 2010-05-06 Han Suh Koo Method and apparatus for decoding/encoding a video siganl
US8532183B2 (en) 2006-08-25 2013-09-10 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US8855200B2 (en) 2006-08-25 2014-10-07 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US8767827B2 (en) 2006-08-25 2014-07-01 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US20100046619A1 (en) * 2006-08-25 2010-02-25 Han Suh Koo Method and apparatus for decoding/encoding a video signal
US8761255B2 (en) 2006-08-25 2014-06-24 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US8724700B2 (en) 2006-08-25 2014-05-13 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US20100202519A1 (en) * 2006-08-25 2010-08-12 Han Suh Koo Method and apparatus for decoding/encoding a video signal
US8718136B2 (en) 2006-08-25 2014-05-06 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US20100202521A1 (en) * 2006-08-25 2010-08-12 Han Suh Koo Method and apparatus for decoding/encoding a video signal
US20100158117A1 (en) * 2006-08-25 2010-06-24 Han Suh Koo Method and apparatus for decoding/encoding a video siganl
US20100158112A1 (en) * 2006-08-25 2010-06-24 Han Suh Koo Method and apparatus for decoding/encoding a video siganl
US8711932B2 (en) 2006-08-25 2014-04-29 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US20100158114A1 (en) * 2006-08-25 2010-06-24 Han Suh Koo Method and apparatus for decoding/encoding a video signal
US20100158113A1 (en) * 2006-08-25 2010-06-24 Han Suh Koo Method and apparatus for decoding/encoding a video signal
US20100158118A1 (en) * 2006-08-25 2010-06-24 Han Suh Koo Method and apparatus for decoding/encoding a video siganl
US20100150236A1 (en) * 2006-08-25 2010-06-17 Han Suh Koo Method and apparatus for decoding/encoding a video signal
US8532181B2 (en) 2006-08-25 2013-09-10 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US8532184B2 (en) 2006-08-25 2013-09-10 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US8649433B2 (en) 2006-08-25 2014-02-11 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US8532180B2 (en) 2006-08-25 2013-09-10 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US8532182B2 (en) 2006-08-25 2013-09-10 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US8532178B2 (en) 2006-08-25 2013-09-10 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US8681863B2 (en) 2006-08-25 2014-03-25 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US8660179B2 (en) 2006-08-25 2014-02-25 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US8559508B2 (en) 2006-08-25 2013-10-15 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US20100150234A1 (en) * 2006-08-25 2010-06-17 Han Suh Koo Method and apparatus for decoding/encoding a video signal
US20100150235A1 (en) * 2006-08-25 2010-06-17 Han Suh Koo Method and apparatus for decoding/encoding a video signal
US8559507B2 (en) 2006-08-25 2013-10-15 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US20100111173A1 (en) * 2006-08-25 2010-05-06 Han Suh Koo Method and apparatus for decoding/encoding a video signal
US20100111171A1 (en) * 2006-08-25 2010-05-06 Han Suh Koo Method and apparatus for decoding/encoding a video signal
US20100111172A1 (en) * 2006-08-25 2010-05-06 Han Suh Koo Method and apparatus for decoding/encoding a video signal
US8630344B2 (en) 2006-08-25 2014-01-14 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US20100104014A1 (en) * 2006-08-25 2010-04-29 Han Suh Koo Method and apparatus for decoding/encoding a video signal
US20100104012A1 (en) * 2006-08-25 2010-04-29 Han Suh Koo Method and apparatus for decoding/encoding a video signal
US20080198924A1 (en) * 2007-02-06 2008-08-21 Gwangju Institute Of Science And Technology Method of computing disparity, method of synthesizing interpolation view, method of encoding and decoding multi-view video using the same, and encoder and decoder using the same
US8165201B2 (en) * 2007-02-06 2012-04-24 Gwangju Institute Of Science And Technology Method of computing disparity, method of synthesizing interpolation view, method of encoding and decoding multi-view video using the same, and encoder and decoder using the same
US8451890B2 (en) 2009-01-19 2013-05-28 Panasonic Corporation Coding method, decoding method, coding apparatus, decoding apparatus, program, and integrated circuit
US8553761B2 (en) 2009-01-19 2013-10-08 Panasonic Corporation Coding method, decoding method, coding apparatus, decoding apparatus, program, and integrated circuit
US8548040B2 (en) 2009-01-19 2013-10-01 Panasonic Corporation Coding method, decoding method, coding apparatus, decoding apparatus, program, and integrated circuit
US20100266010A1 (en) * 2009-01-19 2010-10-21 Chong Soon Lim Coding method, decoding method, coding apparatus, decoding apparatus, program, and integrated circuit
US8369414B2 (en) 2009-04-28 2013-02-05 Panasonic Corporation Image decoding method, image coding method, image decoding apparatus, and image coding apparatus
US20110019746A1 (en) * 2009-04-28 2011-01-27 Chong Soon Lim Image decoding method, image coding method, image decoding apparatus, and image coding apparatus
US8149923B2 (en) 2009-04-28 2012-04-03 Panasonic Corporation Image decoding method, image coding method, image decoding apparatus, and image coding apparatus
US8908771B2 (en) 2009-04-28 2014-12-09 Panasonic Corporation Image decoding method, image coding method, image decoding apparatus, and image coding apparatus
US20110032999A1 (en) * 2009-08-07 2011-02-10 Ying Chen Signaling characteristics of an mvc operation point
US8948241B2 (en) 2009-08-07 2015-02-03 Qualcomm Incorporated Signaling characteristics of an MVC operation point
US20110216827A1 (en) * 2010-02-23 2011-09-08 Jiancong Luo Method and apparatus for efficient encoding of multi-view coded video data

Also Published As

Publication number Publication date
AU2007243935A1 (en) 2007-11-08
CN101416519B (en) 2012-01-11
EP1999968A2 (en) 2008-12-10
JP2014131348A (en) 2014-07-10
BRPI0709167A2 (en) 2011-06-28
KR101353193B1 (en) 2014-01-21
ZA200807023B (en) 2009-11-25
KR20080108448A (en) 2008-12-15
KR101361896B1 (en) 2014-02-12
KR20080108449A (en) 2008-12-15
CN101416517A (en) 2009-04-22
CN101416519A (en) 2009-04-22
JP2009531967A (en) 2009-09-03
EP1999966A2 (en) 2008-12-10
WO2007126509A3 (en) 2008-06-19
WO2007126508A2 (en) 2007-11-08
JP2009531968A (en) 2009-09-03
KR101383735B1 (en) 2014-04-08
KR20090007293A (en) 2009-01-16
MX2008012382A (en) 2008-11-18
WO2007126511A2 (en) 2007-11-08
RU2008142774A (en) 2010-05-10
JP2009531966A (en) 2009-09-03
WO2007126511A3 (en) 2008-01-03
CN101416518B (en) 2013-07-10
WO2007126509A2 (en) 2007-11-08
RU2488973C2 (en) 2013-07-27
JP5845299B2 (en) 2016-01-20
US9100659B2 (en) 2015-08-04
JP2013017215A (en) 2013-01-24
US20090207904A1 (en) 2009-08-20
AU2007243933A1 (en) 2007-11-08
CN101416518A (en) 2009-04-22
MX2008011652A (en) 2008-09-22
RU2529881C2 (en) 2014-10-10
BRPI0709194A2 (en) 2011-06-28
US20090185616A1 (en) 2009-07-23
EP1999967A2 (en) 2008-12-10
RU2008142771A (en) 2010-05-10
BRPI0708305A2 (en) 2011-05-24
JP2016054526A (en) 2016-04-14
JP2012235478A (en) 2012-11-29
ZA200807142B (en) 2010-02-24
JP5255558B2 (en) 2013-08-07
JP2013118671A (en) 2013-06-13
AU2007243933B2 (en) 2012-09-13
JP5213064B2 (en) 2013-06-19
WO2007126508A3 (en) 2008-03-27
JP5213088B2 (en) 2013-06-19
JP5669273B2 (en) 2015-02-12

Similar Documents

Publication Publication Date Title
US9100659B2 (en) Multi-view video coding method and device using a base view
US20090323824A1 (en) Methods and Apparatus for Use in Multi-View Video Coding
JP6395667B2 (en) Method and apparatus for improved signaling using high level syntax for multi-view video encoding and decoding
KR101450921B1 (en) Methods and apparatus for multi-view video encoding and decoding
KR101558627B1 (en) Methods and Apparatus for Incorporating Video Usability Information within a Multi-view Video Coding System
US20090147860A1 (en) Method and apparatus for signaling view scalability in multi-view video coding
AU2012203039B2 (en) Methods and apparatus for use in a multi-view video coding system
AU2012261656A1 (en) Methods and apparatus for use in a multi-view video coding system

Legal Events

Date Code Title Description
AS Assignment

Owner name: THOMSON LICENSING, FRANCE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PANDIT, PURVIN BIBHAS;SU, YEPING;YIN, PENG;AND OTHERS;REEL/FRAME:021518/0956;SIGNING DATES FROM 20060330 TO 20080330

STCB Information on status: application discontinuation

Free format text: ABANDONED -- AFTER EXAMINER'S ANSWER OR BOARD OF APPEALS DECISION