AU2012203039B2 - Methods and apparatus for use in a multi-view video coding system - Google Patents

Methods and apparatus for use in a multi-view video coding system Download PDF

Info

Publication number
AU2012203039B2
AU2012203039B2 AU2012203039A AU2012203039A AU2012203039B2 AU 2012203039 B2 AU2012203039 B2 AU 2012203039B2 AU 2012203039 A AU2012203039 A AU 2012203039A AU 2012203039 A AU2012203039 A AU 2012203039A AU 2012203039 B2 AU2012203039 B2 AU 2012203039B2
Authority
AU
Australia
Prior art keywords
view
views
syntax
decoding
video coding
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
AU2012203039A
Other versions
AU2012203039A1 (en
Inventor
Jill Macdonald Boyce
Cristina Gomila
Purvin Bibhas Pandit
Yeping Su
Peng Yin
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
InterDigital VC Holdings Inc
Original Assignee
InterDigital VC Holdings Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from AU2007243935A external-priority patent/AU2007243935A1/en
Application filed by InterDigital VC Holdings Inc filed Critical InterDigital VC Holdings Inc
Priority to AU2012203039A priority Critical patent/AU2012203039B2/en
Publication of AU2012203039A1 publication Critical patent/AU2012203039A1/en
Application granted granted Critical
Publication of AU2012203039B2 publication Critical patent/AU2012203039B2/en
Assigned to INTERDIGITAL VC HOLDINGS, INC. reassignment INTERDIGITAL VC HOLDINGS, INC. Request for Assignment Assignors: THOMSON LICENSING
Ceased legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

There are provided method and apparatus for use in a multi-view video coding system. The apparatus includes an encoder (100) for encoding at least two views 5 corresponding to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include view specific information. The view specific information indicates a decoding interdependency between at least some of the at least two views.

Description

AUSTRALIA Patents Act COMPLETE SPECIFICATION (ORIGINAL) Class Int. Class Application Number: Lodged: Complete Specification Lodged: Accepted: Published: Priority Related Art: Name of Applicant: Thomson Licensing Actual Inventor(s): Cristina Gomila, Jill MacDonald Boyce, Yeping Su, Purvin Bibhas Pandit, Peng Yin Address for Service and Correspondence: PHILLIPS ORMONDE FITZPATRICK Patent and Trade Mark Attorneys 367 Collins Street Melbourne 3000 AUSTRALIA Invention Title: METHODS AND APPARATUS FOR USE IN A MULTI-VIEW VIDEO CODING SYSTEM Our Ref: 942537 POF Code: 84175/475312 The following statement is a full description of this invention, including the best method of performing it known to applicant(s): -1 6oosq la METHODS AND APPARATUS FOR USE IN A MULTI-VIEW VIDEO CODING SYSTEM The present application is a divisional application from Australian Patent Application No. 2007243935, the entire disclosure of which is incorporated herein by 5 reference. CROSS-REFERENCE TO RELATED APPLICATIONS This application claims the benefit of U.S. Provisional Application Serial No. 60/787,092, filed 29 March, 2006, which is incorporated by reference herein in its entirety. This application is also closely related to two other applications 10 concurrently filed with the instant application and having common inventors and a common assignee (Attorney docket Nos. PU060064 and PU070032), each of which are incorporated by reference in their entireties. TECHNICAL FIELD The present principles relate generally to video encoding and decoding and, 15 more particularly, to methods and apparatus for use in a multi-view video coding system. BACKGROUND In the current implementation of Multi-view Video Coding (MVC) compliant with the International Organization for Standardization/International Electrotechnical 20 Commission (ISO/IEC) Moving Picture Experts Group-4 (MPEG-4) Part 10 Advanced Video Coding (AVC) standard/International Telecommunication Union, Telecommunication Sector (ITU-T) H.264 recommendation (hereinafter the "MPEG 4 AVC standard"), there is no provision to identify a specific view and to signal the camera parameters. This view information is needed for several reasons. View 25 scalability, view random access, parallel processing, view generation, and view synthesis are all Multi-view Video Coding requirements which utilize the view information. Moreover, several of these requirements also utilize camera parameters which are currently not passed in a standardized way. A method has been proposed to enable efficient random access in multi-view 30 compressed bit streams. In the proposed method, a V-picture type and a View Dependency Supplemental Enhancement Information (SEI) message are defined. A feature required in the proposed V-picture type is that V-pictures shall have no temporal dependence on other pictures in the same camera and may only be 942537 WO 2007/126511 PCT/US2007/004997 2 predicted from pictures in other cameras at the same time. The proposed View Dependency Supplemental Enhancement Information message will describe exactly which views a V-picture, as well as the preceding and following sequences of pictures, may depend on. The following are the details of the proposed changes. 5 With respect to V-Picture syntax and semantics, a particular syntax table relating to the MPEG-4 AVC standard is extended to include a Network Abstraction Layer (NAL) unit type of 14 corresponding to a V-picture. Also, the V-picture type is defined to have the following semantics: 10 V-picture: A coded picture in which all slices reference only slices with the same temporal -index (i.e., only slices in other views and not sllce In the current view). When a V-picture would be output or displayed, it also causes the decoding process to mark all pictures from the same view which are not.lDR-pictures or V pictures and which precede the V-picture in output order to be marked as "unused 15 for reference". Each V-picture shall be associated with a View Dependency SE1 message occurring in the same NAL. With respect to the view dependency Supplemental Enhancement Information message syntax and semartics, a View Dependency Supplemental Enhancement 20 information message is defined with the following syntax: view_dependency ( payloadSize) { num-seqjreference_vews ue(v) 25 seq-reference-view_0 ue(v) seqjreference-view_1 ue(v) seq reference-viewN ue(v) 30 numpc_reference_views ue(v) plc-reference-ylewO ue(v) pic_referenceview-1 ue(v) 35 plc-eeference_viewN ue(v) 3 where num_seqreferenceviews/numpicreferenceviews denotes the number of potential views that can be used as a reference for the current sequence/picture, and seq reference_view i/pic referenceviewi denotes the view number for the ith reference view. 5 The picture associated with a View Dependency Supplemental Enhancement Information message shall only reference the specified views described by pic referenceviewi. Similarly, all subsequent pictures in output order of that view until the next View Dependency Supplemental Enhancement Information message in that view shall only reference the specified views 10 described by seqreferenceviewi. A View Dependency Supplemental Enhancement Information message shall be associated with each Instantaneous Decoding Refresh (IDR) picture and V-picture. At least one drawback of this method is the complexity introduced in the 15 decoder due to the dependency being recursively obtained. Additionally, this method requires that every V-picture carry an SEI message (which is a non normative part of the MPEG-4 AVC standard), resulting in the dependency being unable to be used for normative behavior such as reference picture selection. A reference herein to a patent document or other matter which is given as 20 prior art is not to be taken as an admission or a suggestion that that document or matter was known or that the information it contains was part of the common general knowledge as at the priority date of any of the claims. SUMMARY 25 One aspect of the invention provides an apparatus, comprising: an encoder for encoding at least two views corresponding to multi-view video content by defining as a base view any of the at least two views that, for a decoding thereof, is independent of any other of the at least two views, wherein an avc compatibleview id syntax element identifies which one of the base views 30 encoded in the resultant bitstream is syntax compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation for backwards compatibility and wherein a new syntax 942537 3a structure is included that indicates dependency information between views, and a Multi-View Video Coding slice header includes a parameter indicative of which view a particular slice belongs. Another aspect of the invention provides a method, comprising: encoding at 5 least two views corresponding to multi-view video content by defining as a base view any of the at least two views that, for a decoding thereof, is independent of any other of the at least two views, wherein an avc compatibleview id syntax element identifies which one of the base views encoded in the resultant bitstream is syntax compliant with the International Organization for 10 Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation for backwards compatibility and wherein a new syntax structure is included that indicates dependency information between views, and a Multi-View Video Coding 15 slice header includes a parameter indicative of which view a particular slice belongs. Yet another aspect of the invention provides an apparatus, comprising: a decoder for decoding at least two views corresponding to multi-view video content, wherein said decoder determines which, if any, of the at least two views is a base 20 view that, for a decoding thereof, is independent of any other of the at least two views, wherein an avc compatibleview id syntax element identifies which one of the base views encoded in the resultant bitstream is syntax compliant with the International Organization for Standardization/ International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding 25 standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation for backwards compatibility and wherein a new syntax structure is included that indicates dependency information between views, and a Multi-View Video Coding slice header includes a parameter indicative of which view a particular slice belongs. 30 A further aspect of the invention provides a method, comprising: decoding at least two views corresponding to multi-view video content, wherein said decoding step determines which, if any, of the at least two views is a base view that, for a decoding thereof, is independent of any other of the at least two views, 942537 3b wherein an avc compatibleview id syntax element identifies which one of the base views encoded in the resultant bitstream is syntax compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/ International 5 Telecommunication Union, Telecommunication Sector H.264 recommendation for backwards compatibility and wherein a new syntax structure is included that indicates dependency information between views, and a Multi-View Video Coding slice header includes a parameter indicative of which view a particular slice belongs. Another aspect of the invention provides a storage media having video signal 10 data encoded thereupon executed by a computer platform, comprising: at least two views corresponding to multi-view video content encoded by defining as a base view any of the at least two views that, for a decoding thereof, is independent of any other of the at least two views, wherein an avc compatibleview id syntax element identifies which one of the base views encoded in the resultant bitstream is syntax 15 compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation for backwards compatibility and wherein a new syntax structure is included that indicates dependency information between views, and a 20 Multi-View Video Coding slice header includes a parameter indicative of which view a particular slice belongs. According to an aspect of the present principles, there is provided an apparatus. The apparatus includes an encoder for encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the 25 resultant bitstream is encoded to include view specific information. The view specific information indicates a decoding interdependency between at least some of the at least two views. According to another aspect of the present principles, there is provided a method. The method includes encoding at least two views corresponding to multi 30 view video content into a resultant bitstream, wherein the resultant bitstream is 942537 WO 2007/126511 PCT/US2007/004997 4 encoded to include view specific information. The view specific information indicates a decoding Interdependency between at least some of the at least two views. According to yet another aspect of the present principles, there is provided an* apparatus. The apparatus includes a decoder for decoding at least two views 5 corresponding to multi-view video content from a bltstream, wherein the bitstream Is decoded to determine view specific information included therein, the view specific information indicating a decodirig interdependency between at least some of the at least two views. According to still another aspect of the present principles, there Is-provided a 10 method. The method includes decoding at least two views corresponding to multi view video content from a bitstream, wherein the bitstream Is decoded to determine view specific Information included therein. The view specific information indicates a decoding interdependency between at least some of the at leAst two views. According to a further aspect of the present principles, there is provided an 15 apparatus. The apparatus includes an encoder for'encoding at least two views corresponding to multi-view video content by defining as a base view any of the at least two views that, for a decoding thereof, is independent of any other of the at least two views. According to a yet further aspect of the present principles, there is provided a 20 method. The method includes encoding at least two views corresponding to multi view video content-by defining as a base view any of the at least two views that, for a decoding thereof, is independent of ary other of the at least two views. According to a still further aspect of the present principles, there is provided an apparatus.. The apparatus Includes a decoder for decoding at least two views 25 corresponding to multi-view video content, wherein the decoder determines which, if any, of the at least two views is a base view that, for a decoding thereof, is independent of any other of the at least two views. According to an additional aspect of the present principles, there is provided a method. The method includes decoding at least two views corresponding to multi 30 view video content, wherein the decoding step determines.which, if any, of the at least two views is a base view that, for a decoding thereof, Is independent of any other of the at least two views.
WO 2007/126511 PCTIUS2007/004997 5 According to an aspect of the present principles, there is provided an apparatus. The apparatus Includes an encoder for encoding at least two views corresponding to multi-view video content by encoding at least one of the at least two views in a resultant bitstream that is syntax compliant with the International 5 Organization for Standardization/Intemational Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation, for backwards compatibility therewith. According to another aspect of the present principles, there is provided a 10 method. The method includes encoding at least two views corresponding to multi view video content by encoding at least one of the at least two views In a resultant bitstream that is syntax compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication 15 Union, Telecommunication Sector H.264 recommendation, for backwards compatibility therewith. . According to yet another aspect of the present principles, there is provided an apparatus. The apparatus includes a decoder for decoding at least two views corresponding to multi-view video content, wherein at least one of the at least two 20 views is included in a bitstream that is syntax compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding'standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation, for ' backwards compatibility therewith. 25 According to still another aspect of the present principles, there is provided. a method. The method includes decoding at least two views corresponding to multi. view video content, wherein at least one of the at least two views Is Included in a bitstream that is syntax compliant with the International Organization for Standardization/International @lectrotechnIcal Commission Moving Picture Experts 3.0 Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation, for backwards compatibility therewith.
WO 2007/126511 PCT/US2007/004997 6 According to a further aspect of the present principles, there is provided an apparatus. The apparatus includes an encoder for encoding at least one of at least two views corresponding to multi-view video content by selecting between one of two-pre-defined slice types. 5 According to a yet further aspect of the present principles, there is provided a method. The method includes encoding at least one of at least two views corresponding to multi-view video content by selecting between one of two-pre defined slice types. According to a still further aspect of the present principles, there is provided 10 an apparatus. The apparatus includes a decoder for decoding at least one of at least two views corresponding to multi-view video content by determining between one of two-pre-defined slice types. According to an additional aspect of the present principles, there is provided a method. The method includes decoding at least one of at least two views 15 corresponding to multi-view video content by determining between one of two-pre defined slice types. According to an aspect of the present principles, there is provided an apparatus. The apparatus Includes an encoder for encoding at least two views corresponding to multi-view content into a resultant bitstream, wherein the resultant 20 bitstream is encoded to include at least one camera parameter corresponding to at least one of the at least two views. According to another aspect of the present principles, there is provided a method. The method includes encoding at least two views corresponding'to multi 'view content into a resultant bitstream, wherein the resultant bitstream is encoded to 25 include at least one camera parameter corresponding to at least one of the at least two views. According to yet another aspect of the present principles, there is provided an apparatus.. The apparatus includes a decoder for decoding at least two views corresponding to multi-view content from a bitstream, wherein the bitstream is 30 decoded to determine at least one ibamera parameter included therein. The at least one camera parameter corresponds.to at least one of the at least two views. According to still another aspect of the present principles, there is provided a method. The method includes decoding at least two views corresponding to multi- WO 2007/126511 PCTIUS2007/004997 7 view content from a bitstream, wherein the bitstream is decoded to determine at. least one camera parameter included therein. The at least one camera parameter corresponds to at least one of the at least two views. According to a further aspect of the present principles, there.is provided an 5 apparatus. The apparatus includes an encoder for encoding at least two views corresponding.to multi-view video content into a resultant bitstream, wherein the resultant bitstream is encoded to include at least one syntax.element related to at least one camera parameter for at least one of the at least two views. According to a yet further aspect of the present principles,. there is provided a 10 method. The method includes encoding at least two views corresponding to multi view video content into a resultant bitstream, wherein the resultant-bitstream is encoded to Include at least one syntax element related to at least one camera parameter for at least one of the at least two views. According to a still further aspect of the present principles, there is provided 15 an apparatus. The apparatus Includes a decoder for decoding at least two views corresponding to multi-vieW video content from a bitstream, wherein the bitstream is decoded to determine at least one camera parameter for at least one of the at least two views based on at least orie syntax element included in the bitstream. According to an additional aspect of the pi-esent principles, there is provided a 20 method. .The method includes decoding at least two views corresponding to multi view video content from a bitstream, wherein th.e bitstream is decoded to determine at least one camera parameter for at least one of the at least two views based on at least one syntax element Included in the bitstream. These and other aspects, features and advantages of the present principles will 25 become apparent from the following detailed description of exemplary embodiments, which is to be read in connection with the accompanying drawings. BRIEF DESCRIPTION OF THE DRAWINGS The present principles nay be better understood in accordance with the 30 following exemplary figures, In which: FIG. .is a block diagram for an exemplary video encoder to which the. present principles may be applied, in accordance.with an embodiment of the present principles; WO 2007/126511 PCTIUS2007/004997 8 FIG. 2 Is a block diagram for an exemplary video decoder to which the present principles may be applied, In accordance with an embodiment of the present principles; FIG. 3 is a diagram for an inter-view-temporal prediction structure based on 5 the MPEG-4 AVC standard, using'hierarchical B pictures, In accordance with an embodiment of the present principles; FIG. 4 is a flow diagram for an exemplary method for encoding multiple views of multi-view video content, in accordance with an embodiment of the present principles; and 10 FIG. 5 Is a flow diagram for an exemplary method for decoding multiple views of multi-view video content, In accordance with an embodiment of the present principles. DETAILED DESCRIPTION. 15 The present principles are directed to methods and apparatus for use in a multi-view video coding system. The present description illustrates the present principles. It will thus be appreciated.that those skilled in the art will be able to devise various arrangements that, although not explicitly described or shown herein, embody the present 20 principles and are included within its spirit and scope. All examples and conditional language recited herein are intended for pedagogical purposes to aid the reader in understanding the present principles and the concepts contributed by the Inventor(s) to furthering the art, and are to be -construed as being without limitation to such specifically recited examples and 25 conditions. Moreover, all statements herein reciting principles, aspects, and embodiments of the present principles, as well as specific examples thereof, are. Intended to encompass both structural and functional equivalents thereof. Additionally, It is intended that such equivalents include both currently known 30 equivalents as well as equivalents developed in the future, i.e., any elements developed that perform the same function, regardless of structure. Thus, for example, it will be appreciated by those skilled in the drt that the block diagrams presented herein represent conceptual views of illustrative circuitry WO 2007/126511 PCT/US2007/004997 9 embodying the present principles. Similarly, it will be appreciate that any flow charts, flow diagrams, state transition diagrams, pseudocode, and the like represent various processes which may be substantially represented in computer readable media and so executed by a computer or processor, whether or not such computer 5 or processor is explicitly shown. The functions of the various elements shown In the figures may be provided through the use of dedicated hardware as well as hardware capable of executing software in association with appropriate software. When provided by a processor, the functions may be provided by a single dedicated processor, by a single shared 10 processor, or by a plurality of Individual processors, some of which may be shared. Moreover, explicit use of the term "processor" or "controller" should not be construed to refer exclusively to hardware capable of executing software, and may Implicitly include, without limitation, digital signal processor ("DSP") hardware, read-only memory ("ROM") for storing software, random access memory ("RAM"), and 15 non-volatile storage. Other hardware, conventional and/or custom, may also be included. Similarly, any switches shown in the figures are conceptual only. Their function may be carried out through the operation of program logic, through dedicated logic, through the interaction of program control and dedicated logic, or even -manually, the 20 particular technique being selectable by the implementer as more specifically understood from the context. In the claims hereof, any element expressed as a means for performing a specified function is intended to encompass any way of performing that function including, for example, a) a combination of circuit elements that performs that 25 function or b) software in any form, including, therefore, firmware, microcode or the like, combined with appropriate circuitry for executing that softWare to perform the function. The present principles as defined by such claims reside in the fact that the functionalitles provided by the various-recited means are combined and brought. together in the manner which the claims call for. It Is thus regarded that any means 30 that can provide those functionalities are equivalent to those shown herein. Reference in the specification to "one embodiment" or "an embodiment" of the present principles means that a particular feature, structure, characteristic, and so forth described in connection with the embodiment is included in at least one- WO 2007/126511 PCT/US2007/004997 10 embodiment of the present principles. Thus, the appearances of the phrase "in one embodiment" or "In an embodiment" appearing In various places throughout the specification are not necessarily all referring to the same embodiment. Turning to FIG. 1, an exemplary video encoder to which the present principles 5 may be applied Is Indicated generally by the reference numeral 100. An Input to the video encoder 100 Is connected-in signal communication with a non-Inverting Input of a combiner 110. The output of the combiner 11011s connected In signal- communication with a transformer/qudntzer 120.. The output of the transformer/quantizer 120 is connected in signal communication with an entropy 10 coder 140. An output of the entropy coder 140 is available as an output of the . encoder 100. The output of the transformer/quantizer 120 Is further connected in signal communication with an Inverse transformer/quantizer 150. An output of the inverse transformer/quantizer 150 Is connected in signal communication with an. input of a 15 deblock filter 160. An output of the deblock filter 1.60 Is connected in signal communication with reference picture stores 170. A first output of the reference picture stores 170 is connected In signal communication with first input of a motion estimator 180. The Input to the encoder 100 Is further connected in signal communication with a second input of the motion estimator 180. The output of the 20 motion.estimator 180 is connected in signal communication with a first Input of a motion compensator 190. A second output of the reference picture stores 170 is connected in signal communication with a second input of the motion compensator 190. The output of the motion compensator 190 is connected In signal communication with an inverting Input of the combiner- 110. 25 Turning to FIG. 2, an exemplary video decoder to which the present principles may be applied Is Indicated generally by the reference numeral 200. The video decoder 200 Includes an entropy decoder 210 for receiving a video sequence. A first output of the entropy decoder 210 is connected in signal communication with an input of an Inverse quantizer/transformer 220. An output of 30 the inverse quantizer/transformer 220 Is connected In signal communication with a first non-Inverting Input of a combiner 240. The output of the combiner 240 Is connected In signal communIation with an input of a deblock filter 290. An output of the deblock filter 290 is connected In WO 2007/126511 PCT/US2007/004997 11 signal communication with an input of a reference picture stores 250. The'output of the reference picture stores 250 Is connected in signal communication with a first input of a motion compensator 260. An output of the motion -compensator 260 Is connected in signal communication with a second non-Inverting inpyt of the 5 combiner 240. A second output of the entropy decoder 210 Is connected in signal communication with a second input of the motion compensator 260. The output of the deblock filter 290 Is available as an output of the video decoder 200. In an embodiment of the present principles, a high. level syntax is proposed for efficient processing of a multi-view sequence. In particular, we propose creating 10 a new parameter set called View Parameter Set (VPS) with its own NAL unit type and two more new NAL unit types to support mlti-view slices, with the NAL unit types including a view identifier (Id) in the NAL header to identify to'which vlewn the slice belongs. For view scalability and backward compatibility with decoders compliant with the MPEG-4 AVC standard, we propose to maintain one MPEG-4 15 AVC compliant view which we call an "MPEG-4 AVC compliant Base View". As used herein, "high level syntax" refers to.syntax present in the bitstream that resides hierarchically above the macroblock layer. For example, high level syntax, as used herein, may refer to, but is not limited to, syntax at the slice header level, Supplemental Enhancement Information (SEI) level-, picture parameter set 20 level; and sequence parameter set level.
In the current 'nplementation of the Multi-view Video Coding system described above as having no provision to Identify a specific view and to signal camera parameters, different views are Interleaved to form a single sequence instead.of treating the different views as separate views. Since the syntax Is 25 compatible with the MPEG-4 AVC standard, as noted above, it Is presently not possible to Identify which view a given slice belongs to. This view- information is needed for -several reasons. View scalability, view random access, parallel processing, view generation, and view- synthesis are all Multi-view Video Coding. requirements which need to identify a view. For efficient support of view random 30 access and view scalability, it is important for the decoder to know how different pictures depend on each other, so only pictures that are necessary are decoded. Camera parameters are needed for view synthesis. 'If view synthesis is eventually used In the decoding loop, a standardized way of signaling camera parameters WO 2007/126511 PCT/US2007/004997 12 needs to be specified. In accordance with an embodiment, a view parameter set is used. In an embodiment, it is presumed that one view is needed that is fully backward compatible with the MPEG-4AVC standard for the purpose of supporting 5 non-MVC compatible but MPEG-4-AVC compatible decoders. In an embodiment, it. Is presumed that there will be views that are independently decodable to facilitate fast view random access. We refer to these views as "base views". A base view may or may not be compatible with the MPEG-4 AVC standard, but an MPEG-4 AVC compatible view is always a base view. 10 Turning to FIG. 3, an inter-view-temporal prediction structure based on the MPEG-4 AVC standard, using hierarchical.B pictures, is indicated generally by the reference numeral 300. In FIG. 3, the variable I denotes an intra coded picture, the variable P denotes a predictively coded picture; the variable B denotes a bi predictively coded picture, the variable T denotes a location of a particular picture; 15 and the variable S. denotes a particular view to Which cortesponds.a particular picture. In accordance with an embodiment, the following terms are defined: Anchor picture 19 defined as a picture the decoding of which does not involve 20 any picture sampled at.a different time instance. An anchor picture Is signaled by setting the nal_refjdc to 3. In FIG. 3, all pictures in locations TO, T8..., T96, and TI 00 are examples of anchor pictures. Non-anchor picture is defined as a picture which does not have the above constraint specified for an anchor picture. In'FIG. 3, pictures B2, B3, and B4 are 25 non-anchor pictures. Base view is a view which does not depend on any other view and can be independently decoded. In FIG. 3, view SO is an example of base view. Also, in an embodiment, a new parameter set is proposed called the View 30. Parameter Set with -its own NAL.unit type and two new NAL unit types to support Multi-view Video Coding slices. We also modify the slice header syntax to indicate the view-id and the view parameter set to be used.
WO 2007/126511 PCT/US2007/004997 13 The MPEG-4 AVC standard includes the following two parameter sets: (1) Sequence Parameter Set (SPS), which includes information that is not expected to change over an entire sequence; and (2) Picture Parameter Set (PPS), which includes information that is not expected to change for each picture. 5 Since Multi-view Video Coding has additional Information which is specific to each view, we have created a separate View Parameter Set (VPS) in order to transmit this information. All.the- information that is needed to determine the dependency between the different views is indicated in the View Parameter Set. The syntax table for the proposed View Parameter Set is shown in TABLE 1 (View 10 Parameter Set RBSP syntax).. This View Parameter Set is included in a new NAL unit type, for example, type 14 as shown in TABLE 2 (NAL unit type codes). In accordance with -the description of the present Invention, the following terms are defined: 15 viewparametersetjid identifies the view parameter set that is referred to In the slice header. The value of the view_parametersetid shall be in the range of 0 to 255. numberofviews_minus -1 plus 1 Identifies the total number of views in the 20 bItstream. The value of the numberof_view_minus_1 shall be In the range of 0. to 255. avc-compatibleviewId indicates the viewId of the AVC compatible view. The value of avccompatible_viewId shall be in the range of 0 to 255. 25 Is...baseview-flago] equal to 1 Indicates that the view I is a base view and is independently decodable. is_basevlewflag[l] equal to 0 indicates that the view I is not a base view. The value of is_baseview_flag[] shall be equal to 1 for an AVC compatible view I. 30 dependency.updatejflag equal to 1 Indicates that dependency information for this view is updated in the VPS. dependency-update-flag equal to 0 Indicates that the dependency Information for this view is not updated and should not be changed.
WO 2007/126511 PCT/US2007/004997 14 anchor-picture dependency_.maps[i]Uj equal to 1 Indicates the anchor pictures with viewId equal to j will depend on the anchor pictures with viewed equal . to I. 5 non-anchorpicture-dependency-maps[i]o] equal to indicates the non anchor pictures with view-id equal to j will depend on the non-anchor pictures with viewed equal to I. nonanchor-pcturedependency-mapIsi]bI is present only when anchor-picturedependency-maps[][iI equals 1. If 10 anchor-picturedependency-maps[i]] Is present and is equal to zero non_anchor-picture-dependency-maps[i][I shall be inferred as being equal to 0. TABLE 1 view-parameterset-rbsp() { C Descripto r. viewjarameter-setsd 0 ue(v) number-ofviews-minue.1 0 ue(v) avccompatibleviewId 0 ue(v) for( i = 0; 1 <= number_of_views-minus_1; 1++) { Inbase _vewjlag[] o U(1) dependency-updatejflag 0 u(1) 1 (dependancy-update-flag == 1) { foray = 0; j <= number_ofviews_minut.1; J++) { anchor-picture_dpendencymaps[i[j] 0 f(1) If (anchor-ploture-dependency-maps[]m == 1) . non-anchor_plcturedependencymapsEi][] 0 f(1) . ) fo I= 0;, <= number-of-ewsmnus_ 1; ++ ) { camera-parameters_presentflag -. u(1) - If (camera-parameters-presentflag == 1) 1 camera-parametera .0- '_O f(32) camera-parameters_3.4[a . 0 1(32) 15 WO 2007/126511 PCT/US2007/004997 15 TABLE nalunlt_type Content of NAL unit and RBSP syntax structure C D Unspecified 1 Coded slice of a non-IDR picture 2, 3, 4 slicelayer-without-partitioning-rbsp() 2 Coded slice data partition A 2 -__ _ slice.data-partition-a_layer rbsp() 3 Coded slice data partition B 3 slice-data.partiton-b.bjlayer.rbsp() 4 Coded slice data partition C 4 slicedataj-artition-cjayer_rbsp() 5 Coded slice of an IDR picture 2,3 sicejlayer.' wfthouLpartitioning-rbsp( ' . 6 Supplemental enhancement information (SEI) 5 seLrbsp() 7 $equence parameter set seqparameter-seLrbsp() . B Picture parameter set. 1 pIc._parametersetrbsp() 9 Access. unit delimiter 6. accessuniLdelImiterrbsp() 10 End of sequence 7 end-of-seqjrbsp() 11 End of stream 8 end-ofLstrearnjbsp() 12 Filler data -9 fillerOlaAJbsp() ' 13 Sequence parameter set extension 10 . seqparameter-setextension_rbsp( ) 14 View parameter set 11 . vIewparametersetrbps() 15..18 Reserved 19 Coded slice of an auxiliary coded picture without partitioning 2, 3, 4 slicejayerwithouLpartitioningrbsp() 20 Coded slice of a non-IDA picture in scalable extension 2, 3, 4 slice-layerin-scalable-extensionrbsp() .. 21 Coded slice of an IDR picture In scalable extension 2, 3. slice_layer -inscalable-extension.rbsp() 22 Coded slice of a non-IDR pIcture In multi-view extension 2, 3, 4 sllce_layern-mvcextension rbsp() 93 Coded slice of .an IDR picture In multi-view extension 2, 3. 'sllce-layerIn.jnvc.extensionrbsp() 24..31 Unspecified WO 2007/126511 PCT/US2007/004997 16 Optional parameters In the View Paramater Set include the following: 5 camera-parameterspresent-flag equal to 1 Indicates that a projection matrix is signaled as follows. camera-parameters, presuming camera parameter. Is conveyed In the form of a 3x4 projection matrix P, which can be used to map a point in the 3D world to the 10 2D image coordinate: I= P * [XW :Yw :Zw: 1) where I Is in homogeneous coordinates I =[A-1x: A-y : A]. Each element cameraparameters_* can be represented according to the 15 IEEE single precision floating point (32 bits) standard. The advantage of putting this information in a separate parameter set Is that we still maintain Sequence Parameter.Sets (SPS) and Picture Parameter Sets (PPS) that are compatible with the MPEG-4 AVC standard. If we put this information in a Sequence Parameter Set or a Picture Parameter Set then, for each view, we 20 need to send a separate Sequence Parameter Set and Picture Parameter Set. This is too restrictive. Also, this Information does not fit well in either a Sequence Parameter Set or a Picture Parameter Set. Another reason is that since we propose to have an MPEG-4 AVC standard compatible base view we would have to use .separate (MPEG-4 AVC compatible) Sequence Parameter Sets and Picture 25 Parameter Sets for such a view and a separate Sequence Parameter Sets/Plcture Parameter Sets (with view specific Inforiation) for all other views. Placing all the dependency information In a single View Parameter Set at the very beginning of the sequence is very beneficial. The decoder can create a map using all the dependency information once it receives the View Parameter Set. This 30 enables it to know before it receives any slice which views are needed for decoding a particular view. As a result of this, we only need to parse the slice header to obtain-the view_id and.determine if this view is needed to decode a target view as WO 2007/126511 PCT/US2007/004997 17 indicated by a user. Thus, we do not need to buffer any frames or wait until a certain point to determine which frames are needed for decoding a particular view. The dependency information and whether it is a base view is indicated in the View Parameter Set. Even an MPEG-4 AVC compatible base view has associated 5 with it information that is specific to that view (e.g., camera parameters). This Information may be used by other views for several purposes including view interpolation/synthesis. We propose to support only one MPEG4 AVC compatible view since if there are multiple MPEG-4 AVC compatible views, this makes it difficult to Identify for each such slice which view it'belongs to and a non-Multi-vlew Video 10 Coding decoder can easily get confused. By restricting it to just one such view, it Is guaranteed that a non-Muti-view Video Coding decoder will be able to correctly decode the view and a Multi-view Video Coding decoder can easily identify such a view from the View Parameter Set using the syntax avc-compatible_viewid. All other base views (non-MPEG-4 AVC 15 compatible) can be Identified using the Is_base_view_flag. A new slice header for Multi-view Video Coding slices is proposed. In order to support view scalability, view random access, and so forth, we need to know which views the current slice depends upon. For view synthesis and view interpolation we may potentially also need came-a parameters. This information is 20 present In the View Parameter Set as shown above in TABLE 1. The View . . Parameter Set is identified using the view-parametersetid. We propose to add the view_parameter_set_d In the slice header of all the non-MPEG-4 AVC compatible slices as shown in TABLE 3 (Slice Header Syntax). The viewId Information is needed for several Multi-view Video Coding requirements including 25 view interpolation/synthesis, view random access, parallel processing, and so forth. This information can also be useful for special coding modes that only relate to cross-view-prediction. In order-to find the corresponding parameters from the View Parameter Set for this view, we need to send the view._Id in the slice header.
WO 2007/126511 PCT/US2007/004997 18 TABLE 3 slice-header() { C Descript or first-mb-in._slice 2 ue(v) . - - --- 2 ue(v) picparameter.etjid .2 ue(v) if (nal_uniLtype = 22 || nalunitype = 23) { ...... view_parameterset_1d .2 e(v) viewed .2 ue(v) frame-num rn-. .2 u(v) - if( lirame-mbs-only-flag) { . ield_pc...ag .2 u(1) lt( fleldplcjlag) bottom-field-flag 2 u(1)
)
For the new Multi-view Video Coding slices we propose to create new NAL unit types for each slice type (Instantaneous Decoding Refresh (IDR) and non-IDR). 5 We propose to use type 22 for IDR slices and type 23 for non-IDR slices as shown in TABLE 2. view-parameter-set-1d specifies the view parameter set in use. The value of the view-parameter-setid shall be in the range 0 to 255. 10 viewId Indicates the view Id of the current view. The value of the view_parametersetld shall be in the range 0 to 255.. An example.of view random access will now be described in accordance with 15 an embodiment of the present principles. View random access is a Multi-view Video Coding requirement. The goal Is t get access to any view With minimum decoding effort. Let us consider a simple. example of view random access for the prediction structure shown in FIG. 3. Suppose a user requests to decode view S3. From FIG. 3, we see that this 20 view depends on view So, view S2, and view S4. An example View Parameter Set Is iflustrated below.
WO 2007/126511 PCT/US20071004997 19 Let us presume that the view_id for the views are numbered consecutively from 0 to 7 In the slice header syntax and there is only one View Parameter Set present with viewparameterset equal to 0. number_of_views_minus_1 is set to 7. avccompatible_view_id could be set to 0.. 5 For views SO, isjbaseview_flag Is set to 1 and for other views It is set to 0. The dependency map for SO, S1, S2, S3, and S4 will look as shown In TABLE-4A (Dependency table for SO anchorpicture dependencymap) and TABLE 4B (dependency table for SO non-anchor..picture.dependency-map). The dependency map for the other views can be written In a similar way. 10 Once this table Is available at the decoder, the decoder can easily determlle if a slice it receives is needed to decode a particular view. The decoder only needs to parse the slice header to determine the viewed of the current slice and for the target view S3 it can look up the S3 columns in the two tables (TABLE 4a and TABLE 4B) to determine whether or not it should keep the current slice. The 15 decoder needs to distinguish between anchor pictures and non-anchor pictures since they may have different dependencies as can be seen from TABLE 4a and TABLE 4b. For the target view S3, we need to decode the anchor pictures of views SO, S2, and S4 but only need to decode the non-anchor pictures of views S2 and S4. 20 -- TABLE 4A I\ I s0 S1 S2 S3 S4 S5 S6 S7 SO 0 1 - 1 1 1 -1 , 1. 1 S1 0 0 0 0 0. 0 0 0 S2 0 1 0 1 1 1 1 1 S3 0 0 0 0 0 0 0 :::o0 S4 0 0 0 1 - 0 1 1 1 25 TABLE 4B i s S1 S2 S SS S6 S7 SO 0 1 0 0 .0 0 0 0 S1 0 0 0 0 0 0 0 0 'S2 0 1 0 1 0 0 0 0 S3 0 0 0 0_ 10 - 0 .0 10 S4 0 0 10 1 .O . 1 0 10 WO 2007/126511 PCT/US2007/004997 20 Turning to FIG. 4, an exemplary method for encoding multiple views of multi view video content is Indicated generally by the reference numeral 400. The method 400 includes a start block 405 that passes control to a function 5 block 410. The function block 410 reads a configuration file for the encoding - parameters to be used to encode the multiple views, and passes control to a function block 415. The function block sets N to be equal to the number of views to be encoded, and passes control to a function block 420. The function block 420 sets number_of_views_minus_1 equal to N - 1, sets avc..compatible_view_id equal to 10 the view_id of the MPEG-4 AVC compatible view, and passes control to a function block 425. The function block 425 sets view4parametersetsid equal to a valid integer, Initializes a variable I to be equal to zero, and passes control to a decision. block 430. The decision block 430 determines whether or not I Is greater than N. If so, then control is passed to a decision block 435. Otherwise, control Is.passed to a 15 function block 470. The decision block 435 determines whether or not the current view is a base view. If so, then control Is passed to a function block 440. Otherwise, control is passed to a function block 480. The function block 440 sets is_base_vlew_fag(] equal to one, and passes 20 control to a decision block 445. The decision block 445 determines whether or not the dependency Is being updated. If so, the control is passed to a function block 450. Otherwise, control is passed to a function block 485. The function block 450 sets dependency.updatejlag equal to one, and passes control to a function block 455. The function block 455 sets a variable j 25 equal to 0, and passes control to a decision block 460. The decision block 460 determines whether or not j is less than N. If so, then control is passed to a function block 465. Otherwise, control is passed to the function block 487. The function block 465 sets anchor-picture-dependency-maps[i][] and non-anchor-plcture-dependencymaps[1][J to.values Indicated by configuration file, 30 and passes control to a function, block 467. The function block 467 increments the variable j by one, and returns control to the decision block 460. The function block 470 sets cameraparameters-presenLflag equal. to one when camera parameters are present, sets cameraparameters_presenLflag equal WO 2007/126511 PCT/US2007/004997 21 to zero otherwise, and passes control to a decision block 472. The decision block 472 determines whether or not camera-parameterspresentflag is equal to one. If go, then control Is passed to a function block 432. Otherwise, control is passed to a function block 434. 5 The function block 432 writes the camera parameters, and passes control to the function block 434. The function block 434 writes the View Parameter Set (VPS) or the Sequence Parameter Set (SPS), and passes control to an end block 499. -The function block 480 sets is_base_vlew_flag[i] equal to zero, and passes 10 control to the decision block 445. The function block 485 sets-dependency.update-flag equal to zero, and passes control to a function block 487. The function block 487 increments the variable I by. 1, and returns control to the decision block 430. Turning to FIG. 5, an exemplary method for decoding multiple views of multi 15 view video content Is indicated generally by the reference numeral 500. The method 500 includes a start block 505'that passes control to a function block 510. The function block 510 parses a Sequence Parameter Set.(SPS) or View Parameter Set (VPS), view....parameterset-id, number-of_vlewsminus_1, avccompatible_viewId, sets variables I and j equal to zero, sets N equal to 20 number-of_vewsminus-1, and passes control to a decision block 515. The decision block 515 determines whether or not i is less than or equal to N. If so, then control is passed to a function block 570. Otherwise, control is passed to a function block 525. - The function block 570 parses camera.parameters-presenLflag, and passes 25 control to a decision block 572. The decision block 572 determines whether or not . camera.,.parameters-presentflag is equal to one. If so, then control Is passed to a function block 574.. Otherwise.-control is passed to a function block 576. The function block 574 parses the camera parameters, and passes control to the function block 576. 30 The function block 576 continues decoding, and passes control to an end block 599. The function block 525 parses is.base_view_fag[J and dependency.update-flag, and passes control to a decision block 530. The decision WO 2007/126511 PCT/US2007/004997 22 block 530 determines whether or not dependencyupdateflag is equal to zero. If so, then control is passes to a function block 532.- Otherwise, control Is passed to*a decision block 535. The function block 532 Increments I by one, and returns control to the 5 decision block 515. The decision block 535 determines whether or not j is less than or equal to N. If so, then control is passed to a function block 540. Otherwise, control Is passes to a function block 537. The function block 540 parses anchor_picture-dependencymaps[1I, and 10 passes control to a decision block 545. The decision block 545.determines whether or not non-anchor-picture-dependency..maps[i]j is equal to one. If so, then control is passed to a function block 550. Otherwise, control Is passes to a function block 547. The function block 550 parses tie 15 nonanchor-picture.dependencymaps[i][J, and passes control to the function block 547. The function block 547 increments j by one, and returns control to the decision block 535. The function block 537 increments I by one, and returns control to the function 20 block 515. A description will now be given of some of the many attendant advantages/features of the present Invention, some of which have been mentioned above. For example, one advantage/feature Is an apparatus that includes an encoder for encoding at least two views corresponding to multi-view video content 25 into a resultant bitstream, wherein the resultant bitstream is encoded to include view specific information. -The view specific Information indicates a decoding Interdependency between at least some of the at least two views. -Another advantage/feature Is the apparatus having the encoder as described above, wherein the decoding Interdependency allows a corresponding decoding of at least one of 30 the at least two views using only a subset of the at least two views for the corresponding decoding. Yet another advantage/feature Is the apparatus having the encoder as described above, wherein the decoding interdependency Indicated In the view specific Information Is used for random access of at least one of the at least two WO 2007/126511 PCT/US2007/004997 23 views by dropping slices related to any other ones of the at least two views indicated as non-interdependent with respect to the at least one view. Still another advantage/feature Is the apparatus having the encoder as described above, wherein the view specific information is included in a high level syntax. A further 5 advantage/feature is the apparatus having the encoder as described above, wherein the view specific Information Is Included In a parameter set compliant with the International Organization for Standardization/international Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/international Telecommunication Union, Telecommunication Sector H.264 10 recommendation. A yet further advantage/feature Is the apparatus having the. encoder as described above, wherein the view specific Information is included in a View Parameter Set. A still further advantage/feature is the apparatus having the encoder wherein the view specific information is Included in a View Parameter Set as described above, wherein the View Parameter Set is assigned a NAL unit type 15 specifically for use only with the View Parameter Set. An additional advantage/feature Is the apparatus having the endoder wherein'a NAL unit type Is assigned specifically for use only with the View Parameter Set as described above, wherein the NAL unit type is 14. Mo-eover, another advantage/feature Is the apparatus having the encoder as described above, wherein the view specific 20 Information includes at least one syntax element for indicating a View Parameter Set Id. Further, another advantage/feature is the apparatus having the encoder wherein the view specific Information includes at least one syntax element for indicating a View Parameter Set id as described above, wherein the at least one syntax element - Is denoted by a view._parameter_setL d syntax element: Also, another 25 advantage/feature Is the apparatus having the encoder as described above, wherein . the view specific Information includes at least one syntax elemerit for indicating a number of views. Additionally, another advantage/feature Is the apparatus having the encoder wherein the view specific -information includes at least one syntax. element for indicating.a number of views as described above, wherein the at least 30 one syntax element is denoted by a number_of_views_minus_1 syntax element. Moreover, anotheradvantage/feature is the apparatus having the encoder as. described above, wherein the view specific information includes at least one syntax element for indicating a view Id for a particular one of the at feast two views, when WO 2007/126511 PCT/US2007/004997 24 the particular one of the at least two views is encoded in a resultant bitstream that is compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication 5 Sector H.264 recommendation. Further, another advantage/feature is the apparatus having the encoder wherein the view specific information includes at least one syntax element for indicating a view Id for a particular one of the at least two views as described above, wherein the at least one syntax element Is denoted by an avc_compatible._vlew jd syntax element. Also, another advantage/feature Is the 10 apparatus having the encoder as described above, wherein the.view specific information includes at least one syntax element or is implicitly derivable from a high level syntax, the at least one syntax element and the high level syntax for-indicating that a particular one of the.at least two views is compatible -with the International Organization for Standardization/International Electrotechnical Commission Moving 15 Picture Experts Group-4 Part 10 Advanced Video Coding standard/International .;iqcommiiunication Union, Telecommunication Sector -H.264 recommendation. Additionally, another advantage/eature is the apparatus having the encoder wherein the view specific information includes at least one syntax element or is implicitly derivable from a high level.syntax as described above, wherein the at least one 20 syntax element is denoted by an Is_base_viewjlagsyntax element. Moreover, another advantage/feature is the apparatus having the encoder. as described above, wherein the view specific Information includes at least onesy.ntax-element for - idicating whether dependency Information for at least one of the at least two views is present in the resultant bitstream. Further, an ther-advantage/feature is the 25 app.ratus having the encoder hereIn-tne view specific information Includes at least one yntax elemda efrIndicatira whether dependency information for at least one of the at least two views is present in the resultant bitstream as described above,. wherein the at least one.syntax element is denoted by a dependency-updatejlag syntax element. Also, another advantage/feature is. the apparatus having the 30 encoder as described above, wherein the view specific information includes at least one syntax element for. indicating whether at least one anchor pictore:Irza:current one of the at least two views is used for decoding any other ones of the at least two views. Moreover, another advantage/feature is the apparatus having the encoder WO 2007/126511 PCT/US2007/004997 25 wherein the view specific information includes at least one syntax elsnient for indicating whether at least one anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views as described above, wherein the at least one syntax element is denoted by an 5 anchor.picture-dependency-maps[]j] syntax element Also, another advantage/feature is the apparatus having the encoder as described above, wherein the view specific information includes at least one syntax element for indicating whether at least one non-anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two Views. AdditiOnally, another 10 advantage/feature Is the apparatus having the encoder wherein the view specific Information includes at least one syntax element for indicating whether at least one non-anchor picture in a current one of the at least two views is used for decoding any other ones of the at least two views as described above, wherein the at least one syntax element is denoted by a non_anchor_pcturedpendencymaps[1][] 15 syntax element. Moreover, another advantage/feature is the apparatus having the encoder as described above, wherein the resultant bitstream is encoded to include at least one syntax element related to at least one camera parameter for at least one of the at least two views. Further, another advantage/feature is- the apparatus having the encoder wherein the resultant bitstream is encoded to include at least 20 one syntax element related to at least one camera parameter for at least one of. the at least two views as described above, wherein.the at least one -syntax is included in a parameter set corresponding to the resultant bitstream. Moreover, another advantage/feature is an apparatus that includes an encoder for encoding at least two views corresponding to multi-view video content 25 by defining as a base view any of the at least two views that, for a decoding thereof, Is independent of any other of the at least two views. Further, another advantage/feature is an apparatus that includes an encoder for encoding at least two views corresponding to multi-view video content by encoding'at least one of the at least two views in a resultant bitstrearTi that is syntax 30 compliant with the International Organization for Standardization/international Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation, for backwards compatibility therewith. Also, another WO 2007/126511 PCTIUS2007/004997 26 advantage/feature is the apparatus having the encoder as described above, wherein the at least one view is a base view that, for a decoding thereof, Is independent of' any other of the at least two views. Additionally, another advantage/feature is the apparatus having the encoder as described above, wherein an 5 avc-compatibleview-Id syntax element identifies the at least one view as being encoded in the resultant bitstream that Is syntax compliant with the Intemational Organization for Standardization/Intemational.Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation for 10 backwards compatibility. Moreover, another advantage/feature is an apparatus that'includes an encoder for encoding at least one of at least two views corresponding to multi-view video content by selecting between one of two-pre-defined slice types. Further, another advantage/feature is the apparatus having the encoder as described above, 15 wherein the two pre-defined slice types are an Instantaneous Decoding Refresh slice type and a non-instantaneous Decoding Refresh slice type. Also, another advantage/feature is the apparatus having the encoder that selects between the Instantaneous Decod.ing Refresh slice type and the non-instantaneous Decoding Refresh slice type as described above, wherein NAL unit type 22 Is used for the 20 lnstantpneous Decoding Refresh slice type and NAL unit type 23 Is used for the non Instantaneous Decoding Refresh slices. Additionally, another advantage/feature is the.apparatus having the encoder as described above, wherein slice headers for at least one of the at least two slices includes view specific syntax. Moreover, another advantage/feature is the apparatus having the encoder wherein slice headers for at 25 least one of the at least two slices includes view specific syntax as described.above, wherein the view specific syntax is conditioned on NAL unit type 23 and NAL unit type 24. Further, another advantage/feature is the apparatus having the encoder wherein slice headers for at least one of the at least two slices includes view specific syntax as described above, wherein the view specific syntax includes a view 30 parameter set identifier and a view identifier. Also, another advantage/feature is the apparatus having the encoder wherein the view specific syntax includes a view parameter set identifier and a view identifier as described above, wherein the view WO 2007/126511 PCT/US2007/004997 27 parameter set Identifier is denoted by a view-parameter-setid syntax element and the view identifier is denoted by a viewed syntax element. Additionally, another advantage/feature is an apparatus that includes an encoder for encoding at least two views corresponding to multi-view content Into a 5 resultant bitstream, wherein the resultant bitstream is encoded to include at least one camera parameter corresponding to at least one of the at least two view. MAoreover, another advantage/feature is the apparatus having the encoder as described above, wherein the resultant bitstream is encoded to include a View Parameter Set, and the at least one camera parameter is included in the View 10 Parameter Set. Further, another advantage/feature is the apparatus having the encoder as described above, wherein a presence of the at least one camera parameter is Indicated by a syntax element. Also, another advantage/feature Is the apparatus having the encoder wherein a presence of the at least one camera parameter is indicated by a syntax element as described above, wherein the syntax 15 element Is a cameraJparameters4presentflag syntax element. Additionally, another advantage/feature is the apparatus having the encoder as described above, wherein the at least one camera parameter Is denoted by a camera-parameters syntax element. Moreover, another advantage/feature is an apparatus that includes an 20 encoder for encoding at least two views corresponding to multi-view video content into a resultant bitstream, wherein the resultant.bitstream is encoded to Include at least one syntax element related to at least one camera parameter for at least one of the at least two views. Further, another advantage/feature is the apparatus having the encoder as described above, wherein the at least one syntax is a high level 25 syntax element. Also, another advantage/feature is the apparatus having the encoder as described above, wherein the at least one syntax is included in a parameter set corresponding to the resultant bitstream. These and other features and advantages of the present principles may be readily ascertained by one of ordinary skill in the pertinent art based on the 30 teachings herein. It is to be understood that the teachings of the present principles may be implemented in various forms of hardware, software, firmware, special purpose processors, or combinations thereof.
28 Most preferably, the teachings of the present principles are implemented as a combination of hardware and software. Moreover, the software may be implemented as an application program tangibly embodied on a program storage unit. The 5 application program may be uploaded to, and executed by, a machine comprising any suitable architecture. Preferably, the machine is implemented on a computer platform having hardware such as one or more central processing units ("CPU"), a random access memory ("RAM"), and input/output ("I/O") interfaces. The computer platform may also include an operating system and microinstruction code. The 10 various processes and functions described herein may be either part of the microinstruction code or part of the application program, or any combination thereof, which may be executed by a CPU. In addition, various other peripheral units may be connected to the computer platform such as an additional data storage unit and a printing unit. 15 It is to be further understood that, because some of the constituent system components and methods depicted in the accompanying drawings are preferably implemented in software, the actual connections between the system components or the process function blocks may differ depending upon the manner in which the present principles are programmed. Given the teachings herein, one of ordinary skill 20 in the pertinent art will be able to contemplate these and similar implementations or configurations of the present principles. Although the illustrative embodiments have been described herein with reference to the accompanying drawings, it is to be understood that the present principles is not limited to those precise embodiments, and that various changes 25 and modifications may be effected therein by one of ordinary skill in the pertinent art without departing from the scope or spirit of the present principles. All such changes and modifications are intended to be included within the scope of the present principles as set forth in the appended claims. Where the terms "comprise", "comprises", "comprised" or "comprising" are 30 used in this specification (including the claims) they are to be interpreted as specifying the presence of the stated features, integers, steps or components, but not precluding the presence of one or more other feature, integer, step, component or group thereof. 942537

Claims (5)

1. An apparatus, comprising: an encoder for encoding at least two views corresponding to multi-view video 5 content by defining as a base view any of the at least two views that, for a decoding thereof, is independent of any other of the at least two views, wherein an avc compatibleviewid syntax element identifies which one of the base views encoded in the resultant bitstream is syntax compliant with the International Organization for Standardization/International Electrotechnical Commission Moving 10 Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation for backwards compatibility and wherein a new syntax structure is included that indicates dependency information between views, and a Multi-View Video Coding slice header includes a parameter indicative of which view a particular slice belongs. 15
2. A method, comprising: encoding at least two views corresponding to multi-view video content by defining as a base view any of the at least two views that, for a decoding thereof, is independent of any other of the at least two views, wherein an 20 avc compatibleviewid syntax element identifies which one of the base views encoded in the resultant bitstream is syntax compliant with the International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation for 25 backwards compatibility and wherein a new syntax structure is included that indicates dependency information between views, and a Multi-View Video Coding slice header includes a parameter indicative of which view a particular slice belongs.
3. An apparatus, comprising: 30 a decoder for decoding at least two views corresponding to multi-view video content, wherein said decoder determines which, if any, of the at least two views is a base view that, for a decoding thereof, is independent of any other of the at least two views, wherein an avc compatibleview id syntax element identifies which one of the base views encoded in the resultant bitstream is syntax compliant with the 042537 30 International Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation for backwards compatibility and wherein a new syntax structure is 5 included that indicates dependency information between views, and a Multi-View Video Coding slice header includes a parameter indicative of which view a particular slice belongs.
4. A method, comprising: 10 decoding at least two views corresponding to multi-view video content, wherein said decoding step determines which, if any, of the at least two views is a base view that, for a decoding thereof, is independent of any other of the at least two views , wherein an avc compatibleview id syntax element identifies which one of the base views encoded in the resultant bitstream is syntax compliant with the International 15 Organization for Standardization/International Electrotechnical Commission Moving Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation for backwards compatibility and wherein a new syntax structure is included that indicates dependency information between views, and a Multi-View Video Coding slice header 20 includes a parameter indicative of which view a particular slice belongs.
5. A storage media having video signal data encoded thereupon executed by a computer platform, comprising: at least two views corresponding to multi-view video content encoded by 25 defining as a base view any of the at least two views that, for a decoding thereof, is independent of any other of the at least two views, wherein an avc compatibleviewid syntax element identifies which one of the base views encoded in the resultant bitstream is syntax compliant with the International Organization for Standardization/International Electrotechnical Commission Moving 30 Picture Experts Group-4 Part 10 Advanced Video Coding standard/International Telecommunication Union, Telecommunication Sector H.264 recommendation for backwards compatibility and wherein a new syntax structure is included that indicates dependency information between views, and a Multi-View Video Coding slice header includes a parameter indicative of which view a particular slice belongs. 942537
AU2012203039A 2006-03-29 2012-05-23 Methods and apparatus for use in a multi-view video coding system Ceased AU2012203039B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU2012203039A AU2012203039B2 (en) 2006-03-29 2012-05-23 Methods and apparatus for use in a multi-view video coding system

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US60/787,092 2006-03-29
AU2007243935A AU2007243935A1 (en) 2006-03-29 2007-02-27 Multi view video coding method and device
AU2012203039A AU2012203039B2 (en) 2006-03-29 2012-05-23 Methods and apparatus for use in a multi-view video coding system

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
AU2007243935A Division AU2007243935A1 (en) 2006-03-29 2007-02-27 Multi view video coding method and device

Publications (2)

Publication Number Publication Date
AU2012203039A1 AU2012203039A1 (en) 2012-06-14
AU2012203039B2 true AU2012203039B2 (en) 2013-09-26

Family

ID=46642825

Family Applications (1)

Application Number Title Priority Date Filing Date
AU2012203039A Ceased AU2012203039B2 (en) 2006-03-29 2012-05-23 Methods and apparatus for use in a multi-view video coding system

Country Status (1)

Country Link
AU (1) AU2012203039B2 (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5907178A (en) * 1996-01-29 1999-05-25 International Business Machines Corporation Multi-view imaging apparatus
US6055012A (en) * 1995-12-29 2000-04-25 Lucent Technologies Inc. Digital multi-view video compression with complexity and compatibility constraints
US6055274A (en) * 1997-12-30 2000-04-25 Intel Corporation Method and apparatus for compressing multi-view video
US6151362A (en) * 1998-10-30 2000-11-21 Motorola, Inc. Joint rate control for stereoscopic video coding
US20050031035A1 (en) * 2003-08-07 2005-02-10 Sundar Vedula Semantics-based motion estimation for multi-view video coding
US20060026646A1 (en) * 2004-07-27 2006-02-02 Microsoft Corporation Multi-view video format
US6999513B2 (en) * 2002-04-20 2006-02-14 Korea Electronics Technology Institute Apparatus for encoding a multi-view moving picture

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6055012A (en) * 1995-12-29 2000-04-25 Lucent Technologies Inc. Digital multi-view video compression with complexity and compatibility constraints
US5907178A (en) * 1996-01-29 1999-05-25 International Business Machines Corporation Multi-view imaging apparatus
US6055274A (en) * 1997-12-30 2000-04-25 Intel Corporation Method and apparatus for compressing multi-view video
US6151362A (en) * 1998-10-30 2000-11-21 Motorola, Inc. Joint rate control for stereoscopic video coding
US6999513B2 (en) * 2002-04-20 2006-02-14 Korea Electronics Technology Institute Apparatus for encoding a multi-view moving picture
US20050031035A1 (en) * 2003-08-07 2005-02-10 Sundar Vedula Semantics-based motion estimation for multi-view video coding
US20060026646A1 (en) * 2004-07-27 2006-02-02 Microsoft Corporation Multi-view video format

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Broadcast Solution Guide, Issue 1, September 2005 [retrieved on 19 August 2010]. Retrieved from Internet *
COLLINS, R. et al., 'An Active Camera System for Acquiring Multi-view Video', Proceedings of the 2002 International Conference on Image Processing (ICIP '02), September, 2002, pp. 517 - 520 *
WANG, YE-KUI, 'Signalling of scalability information', JVT of ISO/IEC MPEG and ITU-T VECG, 16th Meeting: Poznan Poland, 24-29 July 2005 *

Also Published As

Publication number Publication date
AU2012203039A1 (en) 2012-06-14

Similar Documents

Publication Publication Date Title
AU2007243933B2 (en) Multi-view video coding method and device
US20090323824A1 (en) Methods and Apparatus for Use in Multi-View Video Coding
KR101558627B1 (en) Methods and Apparatus for Incorporating Video Usability Information within a Multi-view Video Coding System
US20090279612A1 (en) Methods and apparatus for multi-view video encoding and decoding
AU2012203039B2 (en) Methods and apparatus for use in a multi-view video coding system
AU2012261656A1 (en) Methods and apparatus for use in a multi-view video coding system

Legal Events

Date Code Title Description
FGA Letters patent sealed or granted (standard patent)
PC Assignment registered

Owner name: INTERDIGITAL VC HOLDINGS, INC.

Free format text: FORMER OWNER(S): THOMSON LICENSING

MK14 Patent ceased section 143(a) (annual fees not paid) or expired