CN102257818B - Sharing of motion vector in 3d video coding - Google Patents
Sharing of motion vector in 3d video coding Download PDFInfo
- Publication number
- CN102257818B CN102257818B CN200980151318.7A CN200980151318A CN102257818B CN 102257818 B CN102257818 B CN 102257818B CN 200980151318 A CN200980151318 A CN 200980151318A CN 102257818 B CN102257818 B CN 102257818B
- Authority
- CN
- China
- Prior art keywords
- depth map
- view
- picture
- motion vector
- texture
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/128—Adjusting depth or disparity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/161—Encoding, multiplexing or demultiplexing different image signal components
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/189—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
- H04N19/196—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/30—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/513—Processing of motion vectors
- H04N19/517—Processing of motion vectors by encoding
- H04N19/52—Processing of motion vectors by encoding by predictive encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/60—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
- H04N19/61—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/70—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N2013/0074—Stereoscopic image analysis
- H04N2013/0081—Depth or disparity estimation from stereoscopic image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N2013/0074—Stereoscopic image analysis
- H04N2013/0085—Motion estimation from stereoscopic image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N2213/00—Details of stereoscopic systems
- H04N2213/003—Aspects relating to the "2D+depth" image format
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computing Systems (AREA)
- Theoretical Computer Science (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Joint coding of depth map video and texture video is provided, where a motion vector for a texture video is predicted from a respective motion vector of a depth map video or vice versa. For scalable video coding, depth map video is coded as a base layer and texture video is coded as an enhancement layer(s). Inter-layer motion prediction predicts motion in texture video from motion in depth map video. With more than one view in a bit stream (for multi view coding), depth map videos are considered monochromatic camera views and are predicted from each other. If joint multi-view video model coding tools are allowed, inter-view motion skip issued to predict motion vectors of texture images from depth map images. Furthermore, scalable multi-view coding is utilized, where interview prediction is applied between views in the same dependency layer, and inter-layer (motion) prediction is applied between layers in the same view.
Description
Technical field
Three-dimensional (3D) video content that different execution modes relate generally to presenting depth map image carries out Video coding.More particularly, different execution mode relates to multi-view video and the degree of depth is carried out to combined optimization to support high coding efficiency.
Background technology
This part is intended to provide background of the present invention or the context about describing in detail in claims.Here, description can comprise and can be probed into but must not be the concept that formerly envisions or probe into.Therefore, unless separately had contrary instruction at this, otherwise what in this part, describe is not the prior art about the specification in the application and claims, and owing to being included in, in this part, to be recognized as not be prior art.
Video encoding standard comprise ITU-T H.261, visual, the ITU-T of ISO/IEC motion picture expert group (MPEG)-1 H.262 or ISO/IEC MPEG-2 video, ITU-TH.263, ISO/IEC MPEG-4 are visual and H.264 ITU-T (is also referred to as ISO/IECMPEG-4 advanced video coding (AVC)).In addition, there is the effort about exploitation new video coding standard.Such standard is scalable video coding (SVC) standard, and it is the scalable expansion to H.264/AVC.Another the such standard just having completed is multi-view video coding (MVC) standard, and it becomes another expansion to H.264/AVC.
In multi-view video coding, each of the video sequence of exporting from different cameras, corresponding to different views, is a bit stream by described video sequence coding.After decoding, in order to show certain view, rebuild and show belonging to the decoding picture of that view.Also can rebuild and show for more than one view.
Multi-view video coding has diversified application, comprises free viewpoint video/TV, 3D TV and monitors application.Current, the joint video team (JVT) of ITU-T Video coding expert group and ISO/IEC motion picture expert group (MPEG) is just making great efforts to develop MVC standard, and it is just becoming expansion H.264/AVC.These standards are hereinafter referred to as MVC and AVC at this.At JVT-AB204 (" Joint Draft Multi-view Video Coding ", the 28th JVT meeting, Hanover, Germany, in July, 2008) in the up-to-date working draft of MVC has been described, it can obtain at ftp3.itu.ch/av-arch/jvt-site/2008_07_Hannover/JVT-AB204. zip.
Except the feature defining in the working draft of MVC, other potential features, particularly pay close attention to the feature of coding tools, also in associating multi-view video model (JMVM), are described.At JVT-AA207 (" Joint Multiview Video Model (JMVM) 8.0 ", the 24th JVT meeting, Geneva, Switzerland, in April, 2008) in the latest edition of JMVM has been described, it can obtain at ftp3.itu.ch/av-arch/jvt-site/2008_04_Geneva/JVT-AA207.zi p.
Fig. 1 shows the expression of traditional MVC decoding order (, bit stream order).Decoding order is arranged and is called as time priority encoding.Define each access unit to comprise for example, for example, encoded picture for whole views (, S0, S1, S2...) of an output time point (time instance) (, T0, T1, T2...).The decoding order that it should be noted in the discussion above that access unit can be not identical with output or DISPLAY ORDER.Figure 2 illustrates traditional MVC prediction for multi-view video coding (comprise inter-picture prediction in each view and inter-view prediction the two) structure.In Fig. 2, by arrow indication predicting, wherein each sensing (pointed-to) object uses accordingly and refers to from (point-from) object as prediction reference.
Anchor point picture is such encoded picture, and therein, whole only references of cutting into slices have the section of same time index, that is, and the only not section of reference in picture morning of working as front view with reference to the section in other views.Being set to 1 pair of anchor point picture by anchor_pic_flag indicates.After anchor point picture is decoded, can, in the case of the infra-frame prediction (inter-prediction) of any picture of decoding before not carrying out comfortable anchor point picture, the whole encoded picture subsequently in DISPLAY ORDER be decoded.If the picture in a view is anchor point picture, whole pictures in other views so with same time index can be also anchor point pictures.Therefore, the decoding of any view can be from the time index corresponding to anchor point picture.Having anchor_pic_flag equals 0 picture and is called as non-anchor some picture.
In the associating draft of MVC, in sequence parameter set (SPS) MVC expansion, specify drawings relativity.Specify independently the correlation for anchor point picture and non-anchor some picture.Therefore, anchor point picture and non-anchor some picture can have different views correlation.But for the picture set with reference to identical SPS, all anchor point pictures must have identical view correlation, and all non-anchor some pictures must have identical view correlation.In SPS MVC expansion, can be for indicate respectively subordinate view as the view using with reference to picture in RefPicList0 and RefPicList1.Among an access unit, in the time that view component A directly depends on view component B, this means view component A by view component B for inter-view prediction.If view component B directly depends on view component C, and if view component A does not directly depend on view component C, view component A depends on view component C indirectly so.
In the associating draft of MVC, in network abstract layer (NAL) unit header, also there is inter_view_flag, whether its instruction photo current is for the inter-view prediction of the picture for other views.In this draft, inter-view prediction supports, only have the sample value of reconstruction can be used to inter-view prediction by texture prediction only, and only has the picture of putting identical reconstruction with the output time of photo current to be used to inter-view prediction.After first byte of NAL unit (NALU), what follow is NAL unit header expansion (3 byte).The expansion of NAL unit header comprises the syntactic element of describing the attribute of NAL unit in the context of MVC.
As the coding tools in JMVM, motion skip predicted macroblock (MB) pattern and the motion vector from inter-view reference picture, and it is only only applicable to non-anchor some view.During encoding, in the time that anchor point picture is encoded, estimate global difference motion-vector (GDMV), and leading-out needle is the weighted average from the GDMV of two contiguous anchor point pictures to the GDMV of non-anchor some picture to make this GDMV for non-anchor some picture then.GDMV is 16 pixel precisions, that is, for any MB in photo current (, the picture that is being encoded or is decoding), the respective regions shifting in inter-view reference picture according to GDMV covers just what a MB in inter-view reference picture.
Based on this GDMV, for each non-anchor some picture, GDMV is carried out to convergent-divergent.For each MB, if this MB utilizes motion skip, indicate so this locality skew of difference motion vector.At decoder place, if use motion skip mode, use so final difference motion vector to find the motion vector in picture between view, and from picture reproduction motion vector between view.
3D video has obtained remarkable concern recently.In addition,, along with obtaining and the progress of Display Technique, by using different application chance, 3D video just becomes a reality in consumer field.Consider the maturation of be sure oing of catching with Display Technique, and add the help of MVC technology, many different 3D Video Applications of being imagined are just becoming more feasible.It should be noted in the discussion above that and conventionally 3D Video Applications can be divided into three kinds: free viewpoint video; 3D TV (video); And immersion (immersive) videoconference.The requirement of these application is can be completely different, and realizes the 3D Video Applications of every type and all have the challenge of himself.
In the time transmitting 3D content based on 2D image, limit bandwidth becomes problem, and therefore needs powerful compressor reducer to use the only view of fair amount to encode to 3D content.But at client device place, for instance, user may watch the experience of 3D content at any angle (for example,, with view navigation or automatic stereo video).Therefore, wish that decoder presents view as much as possible and carries out as far as possible continuously this point.View is synthetic can be by transmitting the view of fair amount and at renderer place, other views being carried out to interpolation and deal with this limit bandwidth simultaneously.Among MPEG video subgroup, the exploration of carrying out in 3D Video coding is tested (3DV EE) to learn similar application scenarios.What claim equally is view to be synthesized and has potential help with depth map video for each view.
In addition, MPEG has also specified the form for adhering to depth map for the conventional video flowing in MPEG-C part 3.At " Text of ISO/IEC FDIS 23002-3Representation of Auxiliary Video and Supplemental the Information " (N8768 of ISO/IEC JTC 1/SC 29/WG 11, Marrakech, Morocco, in January, 2007) in this specification has been described.
In MPEG-C part 3, so-called auxiliary video can be depth map or disparity map.Texture video is made up of three components conventionally, i.e. a luminance component Y, and two chromatic component U and V, but depth map only has the one-component of the distance between representative object pixel and camera.Conventionally, represent texture video with YUV 4:2:0,4:2:2 or 4:4:4 form, wherein, respectively for each 4,2 or 1 luma samples, a chroma sample (U or V) is encoded.Depth map is regarded as according to the video of the only brightness of YUV 4:0:0 form.Can by depth map similarly interframe encode become the texture picture of the only brightness of interframe encode, and therefore the depth map of coding can have motion vector.In the time representing depth map, it is providing flexibility aspect the amount of bits for representing each depth value.For example, the resolution of depth map can be 1/4 wide of for example associated picture and 1/2 height.
Final result it should be noted in the discussion above that and determine that it is application problem which Video Codec has been used, although can be for example monochrome video (4:0:0) by depth map Video coding.For example, can be the H.264/AVC bit stream only with luminance component by depth map encoding.Alternatively, can be by depth map encoding the auxiliary video of definition in H.264/AVC.In H.264/AVC, be independent of main picture, auxiliary picture is encoded, and therefore between the main code picture for sample value and the auxiliaring coding picture for depth value, do not have prediction.
In the time that the depth information of the each picture (, depth map video) for view is provided, synthetic being improved of view presenting for 3D video.Because depth map video can consume for the major part in whole bandwidth of whole bit stream (particularly in the time that each view is associated with depth map), so should be enough efficiently to save bandwidth to the coding of depth map video.
Conventionally, as previously discussed, to depth map video (if existence) coding independently.But, between texture video and its depth map being associated, can there is correlation.For example, the motion vector in coding depth figure and the motion vector in encoding texture video may be similar.Can predict, the sample prediction between depth map and texture is inefficient and is almost useless, but the motion prediction between depth map image and texture image is favourable.
For multi-view video content, MVC is " state of the art " coding standard.Based on MVC standard, can not in a MVC bit stream, encode to depth map video and texture video, and meanwhile support the motion prediction between depth map image and texture image.
Summary of the invention
Different execution modes provide the combined coding to depth map video and texture video.According to different execution modes, the motion vector according to the corresponding motion vector prediction of depth map video for texture video, vice versa.When only presented a view in bit stream time, consider to obey the scene of SVC.In this scene, by depth map Video coding for basic layer and texture video is encoded to enhancement layer.Additionally, can be with inter-layer motion prediction to carry out the motion in predicted texture video according to the motion in depth map video.Alternatively, texture video is encoded to basic layer, is enhancement layer by depth map Video coding, and carrys out the motion in predetermined depth figure according to the motion in texture video.When present more than one view in bit stream time, can be applied in the inter-layer prediction between texture video and corresponding depth map video to each view.In another scene with multiple views, depth map video is considered to monochromatic camera view and can be according to predicting each other.If allow JMVM coding tools, can be with motion skip between view to carry out the motion vector of predicted texture image according to depth map image.In another scene, when predicting between application view between the view in identical dependence layer, and when (motion) predicted between application layer between layer in identical view, utilize scalable multi-view video coding (SMVC).
When together with accompanying drawing, by following detailed description, these and other advantage and feature of the present invention, and their tissue and mode of operation, will become apparent, and wherein, in some accompanying drawings described below, identical element has identical numeral.
Brief description of the drawings
By reference to accompanying drawing, the execution mode of different execution modes is described, in accompanying drawing:
Fig. 1 shows traditional MVC decoding order;
Fig. 2 shows traditional MVC time and the example of inter-view prediction structure;
Fig. 3 shows the block diagram for example components and the exemplary process flow of 3D video system;
Fig. 4 show for according to different execution modes to the encode flow chart of the example process of carrying out of Media Stream;
Fig. 5 a-Fig. 5 d is according to different execution modes, for the expression of different MVC scenes of 3D content with the degree of depth and texture video coding/decoding;
Fig. 6 is the expression that is applicable to the universal multimedia communication system of different execution modes of the present invention;
Fig. 7 is the perspective view of the electronic equipment that can use together with the realization of different execution modes of the present invention; And
Fig. 8 can be included in schematically illustrating of circuit in the electronic equipment of Fig. 7.
Embodiment
Figure 3 illustrates 3D video system 300.At grabber 310 places, using 3D video content as multiple video sequences, (N view) catches.Grabber 310 also can be caught the degree of depth for the subset of each view or view, but alternatively or additionally, can in preprocessor 320, estimate the degree of depth.Preprocessor 320 is responsible for how much and is corrected and color calibration.In addition, preprocessor 320 can be carried out estimation of Depth so that depth map image is associated with video sequence.At encoder 330 places, for example, be bit stream by MVC encoder by video sequence coding.If content is supplied with or is associated with depth map image/picture together with depth map image/picture, can encode to them so, for example, be encoded to the H.264/AVC middle auxiliary picture of supporting.The 3D of compression represents that i.e. bit stream transmits or is accommodated in memory device 340 by particular channel.If many views content is supplied with together with the degree of depth, need so the degree of depth to encode.
In the time that client 350 receives bit stream from channel or memory device 340, the decoder 352 of realizing in client 350 is decoded to N view and depth map image (if existence).Which synthetic view of having encoded decoder 352 can also need to be used for showing according to, and the subset of depth map image and N view be decoded.View synthesizer 354 can be based on N view and depth map image, uses view generation algorithm to generate more views (being called novelty or virtual view).Additionally, view synthesizer 354 can be mutual with display 356, and display 356 for example provides human interface device, such as remote controller.It should be noted in the discussion above that and view synthesizer 354 can be integrated in decoder 352, particularly for the automatic stereo application with small angle.
The combined coding of different execution mode supports to depth map video and texture video.Fig. 4 show for according to different execution modes to comprising the encode flow chart of the example process of carrying out of the Media Stream of the first view (comprising the first degree of depth picture, the second degree of depth picture, the first samples pictures and the second samples pictures).At 400 places, use the first motion vector to predict the second degree of depth picture according to the first degree of depth picture.At 410 places, use the second motion vector to predict the second samples pictures according to the first samples pictures.At 420 places, the first motion vector and the second motion vector are encoded, for example, jointly coding.It should be noted in the discussion above that combined coding can comprise the prediction to the second motion vector according to the first motion vector, vice versa, and only the difference of motion vector is encoded in bit stream.
When only present a view in bit stream time, the scene of consideration obedience SVC is wherein basic layer by depth map Video coding and texture video is encoded to enhancement layer.The scene of additionally, obeying SVC is used inter-layer motion prediction with according to the motion in the motion prediction texture video in depth map video.In another embodiment, when texture video being encoded to basic layer, being enhancement layer and while using inter-layer motion prediction with motion in predetermined depth figure by depth map Video coding, the scene of considering to obey SVC.When present more than one view in bit stream time, can be applied in the inter-layer prediction between texture video and corresponding depth map to each view.In the scene with multiple views, depth map video is considered to monochromatic camera view and can be according to predicting each other.If allow JMVM coding tools, can use between view motion skip with according to the motion vector of depth map image prediction texture image.In another scene, when predicting between application view between the view in identical dependence layer, and when (motion) predicted between application layer between layer in identical view, utilize SMVC.
At H.264 (" Advanced video coding for generic audiovisual services " of ITU-T suggestion, in November, 2007) in SVC specification has been described, it can obtain from http://www.itu.int/rec/T-REC-H.264/en.
In SVC, in NAL unit header SVC expansion, exist the picture that output identification " output_flag " is decoded to specify whether will be output.For Video coding layer (VCL) the NAL unit of basic layer that belongs to AVC compatibility, comprise output_flag in relevant prefix NAL unit.
SVC has also introduced the inter-layer prediction for space and SNR scalability based on texture, remnants and motion.When with the scalable solutions of other of inter-layer prediction that only utilize texture relatively time, this is the main point of the novelty of SVC.This inter-layer prediction provides macro block (MB) level to adapt to, and each MB can carry out rate-distortion optimisation (RDO) pattern between prediction in normal layer in inter-layer prediction and enhancement layer.Spatial scalability in SVC has been summarized as any resolution ratio between two layers, therefore may support basic layer (having the aspect ratio pictures of 4: 3) from having SDTV to the scalability of enhancement layer (having the aspect ratio pictures of 16: 9) with HDTV.By encoding for inter-layer prediction to having with the enhancement layer of its basic layer equal resolution, realize SNR scalability, and by enhancement layer being encoded to the meticulousr quantization parameter (QP) of prediction residue application, it is being described below in more detail.Current, for SNR scalability supporting process granularity scalability (CGS) and medium granularity scalability (MGS).Difference between MGS and CGS is that MGS allows in any access unit place conversion the transporting and decoding of different MG S layer, and CGS layer only can be converted in some fixed point, and the picture of the layer being wherein transformed into is IDR picture.Additionally, can use more flexibly refer-mechanisms so that MGS key picture provides balance between error deviation and enhancement layer coding efficiency.
In SVC, the interlayer coding dependence layering by syntactic element dependency_id identification for spatial scalability and CGS, and by the layering of syntactic element quality_id identification MGS dependence.As performed to temporal_id, these two syntactic elements also indicate in NAL unit header SVC expansion.Position at any time, can have according to the picture inter-layer prediction with less dependency_id value the picture of larger dependency_id value.But in CGS, position and for equal dependency_id value at any time, has picture that quality_id value equals QL and can only use and have gross picture that quality_id value equals QL-1 for inter-layer prediction.It is MGS layer that quality_id is greater than these quality enhancement layer of 0.
If enhancement layer has the resolution (that is, it be CGS or MGS layer) identical with basic layer, texture, remnants or motion can be directly used in to inter-layer prediction.Otherwise before for inter-layer prediction, basic layer is by oversampling (upsample) (for texture or remnants) or convergent-divergent (for motion).These inter-layer prediction methods are below discussed.
Use the coding mode of inter-layer texture prediction in SVC, to be called as " IntraBL " (in BL) pattern.In order to support single loop decoding, only have following MB can use this pattern, for these MB, only in basic layer, the MB for the common location of inter-layer prediction is carried out intraframe coding with limitation.By the MB of intraframe coding with limitation by intraframe coding and without with reference to from any sample of MB between contiguous frames.For spatial scalability, the resolution ratio based between two-layer carries out oversampling to texture.In enhancement layer, to primary signal and possibly the difference between the basic layer texture of oversampling encode, it is the motion compensation residual in the interframe MB in single layer coding seemingly.
If MB is instructed to use residual prediction, in basic layer, the MB for the common location of inter-layer prediction must be interframe MB, and can carry out oversampling to its remnants according to resolution ratio.Then the residue signal of the oversampling possibly of the basic layer of use is with the remnants of prediction enhancement layer.Difference between the remnants of the remaining and basic layer to enhancement layer is encoded.
When the MB in enhancement layer or MB subregion are supported inter-layer motion prediction, and when meanwhile the reference key of basic layer and enhancement layer is identical, basic layer of motion vector that can the common location of convergent-divergent is with the motion vector generation forecast device of the MB in enhancement layer.Have a kind of MB type that is called fundamental mode, it sends a mark for each MB.If this is masked as true and corresponding basic layer MB is not in frame, motion vector, compartment model and reference key are all derived from basic layer so.
According to the depth coding scene of obeying SVC, can consider to have a view of two videos.The first video can be the texture video that comprises texture image.The second video can be the depth map video that comprises depth map image.Depth map image can have the resolution identical or lower with texture image.In such scene, use the scalable method of combination to support the combined coding to the degree of depth and texture image, its operation is as follows.
About basic layer, texture video is encoded to the 4:2:0 view more high chroma sample format of three color components (or have) and output_flag equals 1.About enhancement layer, texture view is encoded to 4:0:0 view (only having luminance component), and output_flag is set to 0.Additionally, if texture image has the resolution identical with depth map image, so CGS or MGS are combined with from different chroma sample forms.Alternatively, if texture image has the resolution higher than depth map image, so spatial scalability is combined with from different chroma sample forms.
Alternatively and about basic layer, be that 4:0:0 view (only having luminance component) and output_flag equal 0 by depth map Image Coding.About enhancement layer, be the 4:2:0 view more high chroma sample format of three color components (or have) by Texture Encoding, and output_flag is set to 1.Additionally, if texture image has the resolution identical with depth map image, so CGS or MGS are combined with chroma sample scalability.Alternatively, if texture image has the resolution higher than depth map image, so spatial scalability is combined with chroma sample scalability.
Moreover, in example before, only utilized inter-layer motion prediction, and IntraBL and residual prediction disabled.In addition, supplemental enhancement information (SEI) message is indicated to indicate in the manner described above SVC bit stream is encoded.H.264/AVC, in SVC and MVC, bit stream can comprise SEI message.Decoding for the sample value in picture to output does not need SEI message, but SEI message is assisted relevant process, such as picture output timing, present, error detection, error concealing and resource reservation.Much SEI message is all H.264/AVC, specify in SVC and MVC standard, and user data SEI message supporting tissue and company are for they self use and specify SEI message.H.264/AVC, SVC and the MVC standard syntax and semantics that comprises the SEI message being used to specify is learned, but is not defined for the process in decoder processing messages.Therefore, in the time that encoder creates SEI message, require encoder in accordance with H.264/AVC, SVC and MVC standard, and and do not require in accordance with H.264/AVC, the decoder of SVC and MVC standard in order to export sequence consensus treatment S EI message.H.264/AVC, SVC and MVC standard comprise a reason that the syntax and semantics of SEI message is learned be allow system specification same explain side information and therefore interactive operation.Intend: system specification may be required in coding side and decoding end is all used specific SEI message, and can specify for the process in receiver treatment S EI message for the application in system specification.
From the visual angle of decoder, if receive such message, consider three execution modes with obtain depth map video and texture video the two.In the first embodiment, carry out many ring decodings, that is, basic layer and enhancement layer are rebuild completely.According to the second execution mode, extract the different subsets of bit stream, and carry out two examples of monocycle decoder., extract only comprise the bit stream subset of basic layer and then to its decode (by H.264/AVC, SVC or MVC decoder) first to obtain depth map video.Subsequently, whole bit stream is decoded to obtain texture video.According to the 3rd execution mode, single loop decoding device is modified, to depend in order to show or view is synthetic whether needs basic layer picture and the basic layer of output picture selectively.If in order to show or the synthetic layer picture substantially that do not need of view, can use traditional single loop decoding, and the basic layer picture of coding is only as the prediction reference for corresponding enhancement-layer pictures.It should be noted in the discussion above that according to these three execution modes, only participate in the synthetic view of view for those and rebuild depth map image.
It was suggested that different mechanisms/scheme is used SVC with the each view for MVC content.For example, it was suggested MVC scheme, wherein used SVC scheme (being embodied as the MVC expansion of SVC standard) to encode to each view.The feature of these proposed schemes comprises the codec design that support is encoded to any view in many views bit stream in scalable mode.Provide the design of reference picture mark and reference picture list structural design to represent for inter-view prediction early than any dependence of any other view when front view to support to use in comfortable view order.Additionally and for the dependence for inter-view prediction represent, proposed reference picture mark design and reference picture list structural design allow the basic representation representing for the dependence of inter-view prediction or strengthen the choice for use representing.The enhancing that dependence represents represents to originate from the decoding to MGS layer represents or fine particles scalability (FGS) layer represents.According to this proposed scalable multi-view video coding (SMVC) scheme, comprise the field in the NAL unit header of field in the NAL unit header of SVC and many views in the NAL of SMVC unit header.For example, in the NAL of SMVC unit header, view_id and dependency_id have been presented.
Therefore,, for multiple view bit streams, consider to obey the depth coding scene of MVC.In the time that multiple views exist, wherein each has depth map video and texture video, can apply the depth coding of obeying MVC with the inter-view prediction between support depth map video and the inter-view prediction between texture video.The mode that many views content is encoded is according to the mode of obeying MVC.According to the first execution mode, as shown in Fig. 5 a, in the time of inter-view prediction (being indicated by arrow) between the auxiliary picture of supporting in different views, be auxiliary picture by whole depth map Image Codings.For example, Fig. 5 a shows from the depth map video of view 1 and can utilize depth map video from view 0 and view 2 as prediction reference.
According to the second execution mode, be normal 4:0:0 view by whole depth map Image Codings, and to the new view identifier of each depth map video distribution.Give an example, and as shown in Fig. 5 b, consider to have the scene of three views, texture video is encoded to view 0 to view 2, and is view N, view N+1 and view N+2 by depth map Video coding.Apply in this embodiment motion prediction between the view between depth map video and the texture video in each view.It should be noted in the discussion above that the motion skip (indicating by the diagonal arrow between each depth map and texture video) of applying does not come into force in this embodiment in JMVM.Again show inter-view prediction at all the other arrows shown in Fig. 5 b.In this example, introduce SEI message to support that the view identifier of depth map is mapped to the texture video that it is associated by renderer.In Fig. 5 b, between view, motion prediction is from depth map video to texture video, and alternatively, between view, motion prediction can be carried out from texture video to depth map.
Obeying in the depth coding scene of JMVM, can between the depth map video of identical view and texture video, support motion prediction.Refer again to Fig. 5 b and show this scene.But in this JMVM scene, motion skip will come into force.As mentioned above, be normal 4:0:0 view by depth map Video coding, and to the new view identifier of each depth map video distribution.If depth map video and texture video belong to identical view, support motion skip from depth map video to texture video so.In this example, global disparity is always denoted as 0 and be always marked as 0 (if using motion skip mode) for partial error's similarities and differences sample of MB.Therefore, can reduce coder complexity.Alternatively, also can support motion skip from texture video to depth map and still global disparity and local difference are denoted as to 0 simultaneously.In this example, introduce SEI message to support that the view identifier of depth map is mapped to the texture video that it is associated by renderer.It should be noted in the discussion above that, in order to support this scene, if when depth map video and texture video are interrelated, they should have identical resolution.Alternatively, also can carry out motion skip process from texture video to depth map video instead of carry out to texture video from depth map video.
Fig. 5 c shows SMVC depth coding scene, and wherein each view has two dependence layers.According to an execution mode, lower dependence layer is corresponding to the basic layer of obeying MVC, uses H.264/AVC coding tools and inter-view prediction to encode to it.The basic layer of each view is corresponding to the depth map video of particular figure, and it encoded with monochromatic mode.With coding tools H.264/AVC, from share identical dependence layer other views inter-view prediction and from the inter-layer motion prediction (use from the depth map video of different views and indicate to the arrow of texture video) of the basic layer of identical view, the higher dependence layer of each view is encoded.This layer is MVC dependence layer, and it is corresponding to the texture video of particular figure.If texture image has the resolution higher than depth map image, so spatial scalability is combined with from different chroma sample forms.
Alternatively and according to another execution mode, lower dependence layer is corresponding to the basic layer of obeying MVC, uses H.264/AVC coding tools and inter-view prediction to encode to it.Basic layer is corresponding to the texture video of particular figure.With monochromatic mode, utilize coding tools H.264/AVC, from share identical dependence layer other views inter-view prediction and from the inter-layer motion prediction (use from the depth map video of different views and indicate to the arrow of texture video) of the basic layer of identical view, the higher dependence layer of each view is encoded.This layer is MVC dependence layer, and it is corresponding to the depth map video of particular figure.
It should be noted in the discussion above that and require complete decoding that synthesize for view, to the depth map in each view (basic layer or enhancement layer).Require for complete decoding (will show) desirable camera view, to texture video (top or basic layer).
Alternatively, as shown in Fig. 5 d, can be applied in another execution mode of realizing depth coding in SMVC, wherein use depth map video to encode to some view, and do not use depth map video to encode to some view.In this example, some view, for example view 1, only have a dependence layer (texture video), and other views can have two dependence layers (depth map and texture video).In addition, this execution mode can also utilize inter-layer prediction from texture video to depth map video in a view.
About Fig. 4 and described above, in the first embodiment, Media Stream can be the SVC bit stream with the basic layer that comprises the first and second degree of depth pictures, wherein according to the first motion vector for example, to the second motion vector encode flat (, using inter-layer motion prediction).Alternatively, in the second execution mode, Media Stream can be the SVC bit stream with the basic layer that comprises the first and second samples pictures, wherein according to the second motion vector for example, to the first motion vector encode flat (, using inter-layer motion prediction).Additionally, layer (in the first execution mode) or enhancement layer (in the second execution mode) are encoded to monochrome video substantially, are wherein MGS, CGS or spatial enhancement layer by enhancement layer coding.Should be noted that, basic layer (in the first execution mode) or enhancement layer (in the second execution mode) are instructed to " not to be output as target ", wherein encode to indicate Media Stream to comprise the basic layer (in the first execution mode) of depth map image to SEI message.Also can encode to indicate Media Stream to comprise the enhancement layer (in the second execution mode) of depth map image to SEI message.
Similarly, as described above, Media Stream can comprise the second view, and the second view comprises degree of depth picture and samples pictures, wherein the 3rd degree of depth picture in the second view is encoded and use inter-view prediction between the second degree of depth picture and the 3rd degree of depth picture.In this embodiment, can be auxiliary picture by depth map slice encode.
In another embodiment, the differential movement between the first coded views that comprises depth map image and the second coded views of comprising texture image is designated as zero and view between motion skip mode for the predictability coding of the second motion vector.
The bitstream format of the different scenes of instruction more than this describes.For the depth coding of obeying SVC, the example of the SEI message syntax that instruction use combined depth and texture video are encoded to SVC is as follows.
joint_depth_coding_SVC(payloadSize){ | ||
view_info_pre_flag | 5 | u(1) |
if(view_info_pre_flag) | ||
view_id | 5 | ue(v) |
} |
If present, the SVC bit stream of this SEI message instruction coding has one or more dependence layers (depth map video) of 4:0:0 form, and from thering are two dependence layers of different chroma sampling forms, only allow inter-layer motion prediction.The semantics of SVC combined depth coding SEI message comprises " view_info_pre_flag ", in the time that it equals 1, indicates the corresponding view identifier of this SVC bit stream designated." view_info_pre_flag " equals 0 instruction does not have given view identifier.Additionally, the video of " view_id " instruction decoding and the view identifier of the corresponding view of depth map.
The execution mode being associated for the above-mentioned depth coding scene with obeying MVC, the grammer of exemplary MVC depth views maps identifiers SEI message is as follows:
depth_id_map_mvc(payloadSize){ | ||
num_depth_views_minus1 | 5 | ue(v) |
for(i=0;i<=num_depth_views_minus1;i++){ | ||
sample_view_id[i] | 5 | ue(v) |
depth_view_id[i] | 5 | ue(v) |
} | ||
} |
MVC depth views maps identifiers SEI message semantics comprises " num_depth_views_minus1 " parameter, the quantity of the view that its instruction is encoded with depth map video.Additionally, " sample_view_id[i] " parameter is indicated the view_id of the texture video of i the view of encoding with depth map video.The view_id of the depth map video of i the view that in addition, " depth_view_id[i] " instruction is encoded with depth map video.
In the time considering to obey the depth coding scene of JMVM, exemplary SEI message syntax can be with described identical for the depth coding of obeying MVC above.For mapping SEI message semantics, as the semantics of MVC depth views maps identifiers SEI message, comprise following: " num_depth_views_minus1 " parameter of the quantity of the view that instruction is encoded with depth map video; " sample_view_id[i] " parameter of the view_id of the texture video of instruction i the view of encoding with depth map video, and " the depth_view_id[i] " parameter of indicating the view_id of the depth map video of i the view of encoding with depth map video.Additionally, in the time existing, the bit stream support depth map video that the view identifier of " depth_view_id[i] " and " sample_view_id[i] " is right from having and the motion skip of texture video.The view that equals " depth_view_id[i] " value from having " view_id " value is set to zero to the difference the indicating motion that has " view_id " value and equal the view of " sample_view_id[i] " value.
About SMVC depth coding, when a view does not have the degree of depth and another is while having the degree of depth, relevant grammer allows basic layer to have the dependency_id that is not equal to 0, to support the inter-view prediction from the texture video in different views (having identical dependency_id value).
It should be further understood that, although the text comprising at this and example description encoding process clearly it will be understood by those skilled in the art that identical concept and principle are also applicable to corresponding decode procedure, vice versa.For example and with regard to Fig. 4, decoder can be decoded to the Media Stream of coding, and the Media Stream of coding has by using the first and second motion vectors to predict according to the first degree of depth picture and the first samples pictures the above-mentioned attribute that the second degree of depth picture and the second samples pictures realize respectively, wherein motion vector has been carried out to for example combined coding.
Fig. 6 is the diagrammatic representation of the universal multimedia communication system that can realize therein for different execution modes.As shown in Figure 6, data source 600 provides the source signal of simulation, not compressed digital or compressed digital form, or the source signal of any combination of these forms.Source signal is encoded to coded media bit stream by encoder 610.It should be noted in the discussion above that can from be positioned at remote equipment among the network of any type almost receive directly or indirectly will be decoded bit stream.Additionally, can receive bit stream from local hardware or software.Encoder 610 can be encoded to more than one video type (such as Voice & Video), or can require more than one encoder 610 to encode to the different media types of source signal.Encoder 610 also can obtain the synthetic input producing, and such as figure and text, or it can produce the coded bit stream of synthetic media.In the following, for simplified characterization, only consider a medium type coded media bit stream to process.But it should be noted that typical real-time broadcast service comprises some stream (at least one audio frequency, video and text are joined caption stream conventionally).The system of should also be noted that can comprise many encoders, but for simplified characterization without loss of generality in the situation that, only shows an encoder 610 in Fig. 6.It should be further understood that, although the text comprising at this and example description encoding process clearly it will be understood by those skilled in the art that identical concept and principle are also applicable to corresponding decode procedure, vice versa.
Transmit coded media bit stream to storage 620.Storage 620 massage storage that can comprise any type are with memory encoding media bit stream.The form of the coded media bit stream in storage 620 can be basic self-contained bitstream format, or one or more coded media bit stream can be encapsulated in container file.Some system " scene " operation, omits storage and coded media bit stream is directly transmitted to transmitter 630 from encoder 610.Then when needed coded media bit stream is transmitted to transmitter 630 (also referred to as server).The form using in transmission can be basic self-contained bitstream format, packet stream format, or one or more coded media bit stream can be encapsulated in container file.Encoder 610, storage 620 and server 630 can be arranged in same physical device, or they can be included in different equipment.Encoder 610 and server 630 can operation site implementation contents, coded media bit stream is not conventionally by permanent storage in this example, but cushion for short time in content encoder 610 and/or in server 630, with the variation in Processing for removing delay, propagation delay and encoded media bit rate.
Server 630 uses communication protocol stack to send coded media bit stream.This stack can include but not limited to real time transport protocol (RTP), User Datagram Protoco (UDP) (UDP) and Internet protocol (IP).When communication protocol stack is that server 630 is encapsulated into coded media bit stream in grouping towards when grouping.For example, in the time using RTP, server 630 is encapsulated into coded media bit stream in RTP grouping according to RTP pay(useful) load form.Conventionally, each media formats has special RTP pay(useful) load form.Should again note, system can comprise more than one server 630, but for simplicity, a server 630 is only considered in ensuing description.
Server 630 can be connected to gateway 640 or be free of attachment to gateway 640 by communication network.Gateway 640 can be carried out dissimilar function, such as translating the merging of another communication protocol stack, data flow and fork according to the stream of packets of a communication protocol stack and according to down link and/or receiver capability operation data flow, such as the bit rate of the stream forwarding according to the down link network condition control being dominant.The example of gateway 640 comprises the IP wrapper in gateway between the visual telephone of MCU, circuit conversion and packet switched, push-to-talk over cellular (PoC) server, digital video broadcasting-hand-hold type (DVB-H) system, or broadcasting transmitting is carried out to the local Set Top Box forwarding to family wireless network.In the time using RTP, gateway 640 is called as RTP blender or RTP translater, and conventionally serves as the end points that RTP connects.
System comprises one or more receivers 650, and it can receive transmitted signal conventionally, demodulation or be descapsulated into coded media bit stream.Transmit coded media bit stream to record storage 655.Record storage 655 massage storage that can comprise any type are with memory encoding media bit stream.Record storage 655 can alternatively or additionally comprise computing store, such as random access storage device.The form of the coded media bit stream in record storage 655 can be basic self-contained bitstream format, or one or more coded media bit stream can be encapsulated in container file.If there are multiple coded media bit stream associated with each other, such as audio stream and video flowing, so conventionally, use container file and receiver 650 to comprise container file maker or be attached to container file maker, container file maker produces container file from inlet flow.Some system " scene " operation, omits record storage 655 and coded media bit stream is directly transmitted to decoder 660 from receiver 650.In some system, in record storage 655, only keep the forefield of the stream recording, for example, within nearest 10 minutes of the stream recording, collect, and from record storage 655, abandon any early data of record.
Transmit coded media bit stream from record storage 655 to decoder 660.If exist associated with each other and be packaged into the many coded media bit stream in container file, such as audio stream and video flowing, can use so paper analyzer (not shown in FIG.) with from the each coded media bit stream of container file decapsulation.Record storage 655 or decoder 660 can comprise paper analyzer, or paper analyzer is attached to record storage 655 or decoder 660.
Coded media bit stream is further processed by decoder 660 conventionally, and decoder 660 is output as one or more not compressed media stream.Finally, renderer 670 can be reproduced not compressed media stream with for example loud speaker or display.Receiver 650, record storage 655, decoder 660 and renderer 670 can be arranged in same physical device or they can be included in different equipment.It should be noted in the discussion above that can from be positioned at remote equipment among the network of any type almost receive will be decoded bit stream.Additionally, can receive bit stream from local hardware or software.
According to different execution modes, transmitter 630 can be arranged to for many reasons and select the layer or the view that transmit, the situation that is dominant of the network transmitting thereon such as the request in response to receiver 650 or for bit stream.Can be for example for layer or the variation of view for showing from the request of receiver, or the request of comparing the variation of the display device with different abilities from upper one.
Communication equipment of the present invention can use different transmission technologys to communicate, and this transmission technology includes but not limited to code division multiple access (CDMA), global system for mobile communications (GSM), Universal Mobile Telecommunications System (UMTS), time division multiple access (TDMA), frequency division multiple access (FDMA), TCP/IP (TCP/IP), sending and receiving short messages service (SMS), Multimedia Message transmitting-receiving service (MMS), Email, instant message transrecieving service (IMS), bluetooth, IEEE 802.11 etc.Communication equipment can use different medium to communicate, and this medium includes but not limited to that radio, infrared ray, laser, cable connect etc.
Fig. 7 and Fig. 8 show a representational electronic equipment 12 can realizing therein different execution modes.But should be appreciated that the equipment that is not intended to different execution modes to be restricted to a particular type.The electronic equipment 12 of Fig. 7 and Fig. 8 comprises shell 30, takes the display 32 of liquid crystal display form, keypad 34, microphone 36, receiver 38, battery 40, infrared port 42, antenna 44, take smart card 46, card reader 48, radio interface circuit 52, codec circuit 54, controller 56 and the memory 58 of UICC form according to an execution mode.Independent circuit and element are all type well known in the art.
Different execution mode described here has been described in the general context of method step or process, in one embodiment, can realize the method step or process by computer program, this computer program is presented as computer-readable medium, comprise computer executable instructions, such as program code, carried out by the computer in networked environment.Computer-readable medium can comprise detachable and non-dismountable memory device, it includes but not limited to, read-only memory (ROM), random access storage device (RAM), compressed disk (CD), digital universal video disc (DVD) etc.Conventionally, program module can comprise and carries out particular task or realize routine, program, object, assembly, data structure of particular abstract data type etc.Computer executable instructions, the data structure being associated and program module represent the example of the program code of the step for carrying out method disclosed herein.The particular sequence representative of such executable instruction or the data structure being associated is for realizing the corresponding actions of the function of describing in above-mentioned steps or process.
Can be in software, hardware, applied logic, or in the combination of software, hardware and applied logic, realize different execution modes.Software, applied logic and/or hardware can be positioned on for example chipset, mobile device, desktop computer, laptop computer or server.The software of different execution modes and web realize and can realize with following: there is the standardization program designing technique of rule-based logic, and for realizing other logics of disparate databases search step or process, correlation step or process, comparison step or process and deciding step or process.Also can among network element or module, completely or partially realize different execution modes.It should be noted in the discussion above that the word " assembly " that uses in this and claims of following and " module " are intended to comprise uses a line or multirow software code and/or hardware to realize and/or for receiving the realization of equipment of artificial input.
Independent and the specific structure of describing in above-mentioned example is appreciated that the representational structure that is configured for the device of carrying out the specific function described in ensuing claims, although the restriction in claims is not appreciated that formation " device adds function " restriction the in the situation that of not using term " device " in claims.Additionally, should not be used to explain that to the use of term " step " any specific restriction in claims forms " step adds function " restriction in the foregoing description.The independent reference of publishing with the patent that comprises mandate described here or that otherwise mention, patent application and non-patent is limited, and such reference is not intended to, and should not be interpreted as limiting the scope of ensuing claims.
Foregoing description for execution mode provides for the purpose of illustration and description.Foregoing description be not limit or be intended to different execution modes to be limited in disclosed accurate form, and consider above instruction, amendment and modification are possible, or can from the practice of different execution modes, obtain them.Selecting and describing execution mode discussed herein is principle and character and the practical application thereof for different execution modes are described, to make those skilled in the art can utilize different execution modes, and different amendments is suitable for desired specific use.The feature of execution mode described here can be combined in the whole possible combination of method, device, module, system and computer program.
Claims (12)
1. the method for the Media Stream that comprises the first depth map picture being associated with the first texture picture and the second depth map picture being associated with the second texture picture is encoded, wherein said the first depth map picture belongs to the first view, and described the second depth map picture belongs to the second view, described method comprises:
Use the first motion vector to belong to the described second depth map picture of described the second view according to described the first depth map picture prediction that belongs to described the first view;
Use the second motion vector to predict described the second texture picture according to described the first texture picture;
Described the first motion vector is encoded in bit stream;
At least described the second motion vector based on showing as predicted motion vector, predicts described the first motion vector; And
Difference between described the first motion vector and described predicted motion vector is encoded in described bit stream.
2. method according to claim 1, wherein said the first texture picture and the second texture picture belong to the first view, and described Media Stream also comprises the second view, described the second view comprises the 3rd depth map picture being associated with texture picture and the 4th depth map picture being associated with the 4th texture picture, and described method also comprises:
Use the 3rd motion vector to predict described the 4th depth map picture according to described the 3rd depth map picture;
Use the 4th motion vector to predict described the 4th texture picture according to described texture picture;
Described the 3rd motion vector and described the 4th motion vector are encoded.
3. method according to claim 2, also comprises:
Between described the second depth map picture and described the 4th depth map picture, between execution view, predicting.
4. the equipment for the Media Stream that comprises the first depth map picture being associated with the first texture picture and the second depth map picture being associated with the second texture picture is encoded, wherein said the first depth map picture belongs to the first view, and described the second depth map picture belongs to the second view, described equipment comprises:
For using the first motion vector to belong to the device of the described second depth map picture of described the second view according to described the first depth map picture prediction that belongs to described the first view;
For using the second motion vector to predict the device of described the second texture picture according to described the first texture picture;
For described the first motion vector being encoded to the device of bit stream;
Be used at least described the second motion vector based on showing as predicted motion vector, predict the device of described the first motion vector; And
For the difference between described the first motion vector and described predicted motion vector being encoded to the device of described bit stream.
5. equipment according to claim 4, wherein said the first texture picture and the second texture picture belong to the first view, and described Media Stream also comprises the second view, described the second view comprises the 3rd depth map picture being associated with texture picture and the 4th depth map picture being associated with the 4th texture picture, and described equipment also comprises:
For using the 3rd motion vector to predict the device of described the 4th depth map picture according to described the 3rd depth map picture;
For using the 4th motion vector to predict the device of described the 4th texture picture according to described texture picture;
For the device that described the 3rd motion vector and described the 4th motion vector are encoded.
6. equipment according to claim 5, also comprises for the device predicting between execution view between described the second depth map picture and described the 4th depth map picture.
7. the method for the Media Stream that comprises the first depth map picture being associated with the first texture picture and the second depth map picture being associated with the second texture picture is decoded, wherein said the first depth map picture belongs to the first view, and described the second depth map picture belongs to the second view, described method comprises:
By the first motion vector from bit stream decoding;
The residual value of fetching by the prediction to based on described the first motion vector with from described bit stream is sued for peace, and the second motion vector is decoded;
Described the second depth map picture that belongs to described the second view is decoded, and wherein said the first motion vector is for predicting described the second depth map picture according to described the first depth map picture that belongs to described the first view; And
Described the second texture picture is decoded, and wherein said the second motion vector is for predicting described the second texture picture according to described the first texture picture.
8. method according to claim 7, wherein said the first texture picture and the second texture picture belong to the first view, and described Media Stream also comprises the second view, described the second view comprises the 3rd depth map picture being associated with texture picture and the 4th depth map picture being associated with the 4th texture picture, and described method also comprises:
The 3rd motion vector and the 4th motion vector are decoded;
Described the 4th depth map picture is decoded, and wherein said the 3rd motion vector is for predicting described the 4th depth map picture according to described the 3rd depth map picture; And
Described the 4th texture picture is decoded, and wherein said the 4th motion vector is for predicting described the 4th texture picture according to described texture picture.
9. method according to claim 8, also comprises:
Between described the second depth map picture and described the 4th depth map picture, between execution view, predicting.
10. the equipment for the Media Stream that comprises the first depth map picture being associated with the first texture picture and the second depth map picture being associated with the second texture picture is decoded, wherein said the first depth map picture belongs to the first view, and described the second depth map picture belongs to the second view, described equipment comprises:
For the device from bit stream decoding by the first motion vector;
Sue for peace for the residual value of fetching by the prediction to based on described the first motion vector with from described bit stream, the device that the second motion vector is decoded;
For the device of decoding to belonging to the described second depth map picture of described the second view, wherein said the first motion vector is for predicting described the second depth map picture according to described the first depth map picture that belongs to described the first view; And
For the device that described the second texture picture is decoded, wherein said the second motion vector is for predicting described the second texture picture according to described the first texture picture.
11. equipment according to claim 10, wherein said the first texture picture and the second texture picture belong to the first view, and described Media Stream also comprises the second view, described the second view comprises the 3rd depth map picture being associated with texture picture and the 4th depth map picture being associated with the 4th texture picture, and described equipment also comprises:
For the device that the 3rd motion vector and the 4th motion vector are decoded;
For the device that described the 4th depth map picture is decoded, wherein said the 3rd motion vector is for predicting described the 4th depth map picture according to described the 3rd depth map picture; And
For the device that described the 4th texture picture is decoded, wherein said the 4th motion vector is for predicting described the 4th texture picture according to described texture picture.
12. equipment according to claim 11, also comprise:
Be used at the device of predicting between execution view between described the second depth map picture and described the 4th depth map picture.
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10650308P | 2008-10-17 | 2008-10-17 | |
US61/106,503 | 2008-10-17 | ||
PCT/FI2009/050834 WO2010043773A1 (en) | 2008-10-17 | 2009-10-16 | Sharing of motion vector in 3d video coding |
Publications (2)
Publication Number | Publication Date |
---|---|
CN102257818A CN102257818A (en) | 2011-11-23 |
CN102257818B true CN102257818B (en) | 2014-10-29 |
Family
ID=42106276
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN200980151318.7A Active CN102257818B (en) | 2008-10-17 | 2009-10-16 | Sharing of motion vector in 3d video coding |
Country Status (4)
Country | Link |
---|---|
US (3) | US9973739B2 (en) |
EP (1) | EP2338281A4 (en) |
CN (1) | CN102257818B (en) |
WO (1) | WO2010043773A1 (en) |
Families Citing this family (110)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2009130561A1 (en) | 2008-04-21 | 2009-10-29 | Nokia Corporation | Method and device for video coding and decoding |
CN102106152A (en) * | 2008-07-24 | 2011-06-22 | 皇家飞利浦电子股份有限公司 | Versatile 3-D picture format |
CN102792699A (en) * | 2009-11-23 | 2012-11-21 | 通用仪表公司 | Depth coding as an additional channel to video sequence |
CN102986214A (en) * | 2010-07-06 | 2013-03-20 | 皇家飞利浦电子股份有限公司 | Generation of high dynamic range images from low dynamic range images |
KR102185765B1 (en) | 2010-08-11 | 2020-12-03 | 지이 비디오 컴프레션, 엘엘씨 | Multi-view signal codec |
US8994792B2 (en) | 2010-08-27 | 2015-03-31 | Broadcom Corporation | Method and system for creating a 3D video from a monoscopic 2D video and corresponding depth information |
JPWO2012029884A1 (en) | 2010-09-03 | 2013-10-31 | ソニー株式会社 | Encoding apparatus, encoding method, decoding apparatus, and decoding method |
EP2630799A4 (en) * | 2010-10-20 | 2014-07-02 | Nokia Corp | Method and device for video coding and decoding |
US10057559B2 (en) * | 2010-12-03 | 2018-08-21 | Koninklijke Philips N.V. | Transferring of 3D image data |
RU2480941C2 (en) | 2011-01-20 | 2013-04-27 | Корпорация "Самсунг Электроникс Ко., Лтд" | Method of adaptive frame prediction for multiview video sequence coding |
CN105100822B (en) * | 2011-01-28 | 2018-05-11 | 华为技术有限公司 | Method for loading auxiliary video supplementary information, processing method, device and system |
EP2485495A3 (en) * | 2011-02-03 | 2013-08-28 | Broadcom Corporation | Method and system for creating a 3D video from a monoscopic 2D video and corresponding depth information |
US9565449B2 (en) | 2011-03-10 | 2017-02-07 | Qualcomm Incorporated | Coding multiview video plus depth content |
US20120236934A1 (en) * | 2011-03-18 | 2012-09-20 | Qualcomm Incorporated | Signaling of multiview video plus depth content with a block-level 4-component structure |
US8761990B2 (en) * | 2011-03-30 | 2014-06-24 | Microsoft Corporation | Semi-autonomous mobile device driving with obstacle avoidance |
US9485517B2 (en) | 2011-04-20 | 2016-11-01 | Qualcomm Incorporated | Motion vector prediction with motion vectors from multiple views in multi-view video coding |
WO2012171477A1 (en) | 2011-06-15 | 2012-12-20 | Mediatek Inc. | Method and apparatus of texture image compression in 3d video coding |
KR20240017975A (en) * | 2011-06-16 | 2024-02-08 | 지이 비디오 컴프레션, 엘엘씨 | Context initialization in entropy coding |
JP5830993B2 (en) * | 2011-07-14 | 2015-12-09 | ソニー株式会社 | Image processing apparatus and image processing method |
US9363535B2 (en) * | 2011-07-22 | 2016-06-07 | Qualcomm Incorporated | Coding motion depth maps with depth range variation |
US9521418B2 (en) | 2011-07-22 | 2016-12-13 | Qualcomm Incorporated | Slice header three-dimensional video extension for slice header prediction |
US20130188013A1 (en) * | 2011-07-22 | 2013-07-25 | Qualcomm Incorporated | Mvc based 3dvc codec supporting inside view motion prediction (ivmp) mode |
US11496760B2 (en) | 2011-07-22 | 2022-11-08 | Qualcomm Incorporated | Slice header prediction for depth maps in three-dimensional video codecs |
US9288505B2 (en) | 2011-08-11 | 2016-03-15 | Qualcomm Incorporated | Three-dimensional video with asymmetric spatial resolution |
JP5155462B2 (en) * | 2011-08-17 | 2013-03-06 | 株式会社スクウェア・エニックス・ホールディングス | VIDEO DISTRIBUTION SERVER, VIDEO REPRODUCTION DEVICE, CONTROL METHOD, PROGRAM, AND RECORDING MEDIUM |
US10165267B2 (en) * | 2011-08-30 | 2018-12-25 | Intel Corporation | Multiview video coding schemes |
CN103891291A (en) * | 2011-08-30 | 2014-06-25 | 诺基亚公司 | An apparatus, a method and a computer program for video coding and decoding |
TWI595770B (en) | 2011-09-29 | 2017-08-11 | 杜比實驗室特許公司 | Frame-compatible full-resolution stereoscopic 3d video delivery with symmetric picture resolution and quality |
WO2013049179A1 (en) | 2011-09-29 | 2013-04-04 | Dolby Laboratories Licensing Corporation | Dual-layer frame-compatible full-resolution stereoscopic 3d video delivery |
CN102387368B (en) * | 2011-10-11 | 2013-06-19 | 浙江工业大学 | Fast selection method of inter-view prediction for multi-view video coding (MVC) |
KR20130046534A (en) * | 2011-10-28 | 2013-05-08 | 삼성전자주식회사 | Method and apparatus for encoding image and method and apparatus for decoding image |
WO2013068548A2 (en) | 2011-11-11 | 2013-05-16 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Efficient multi-view coding using depth-map estimate for a dependent view |
IN2014KN00990A (en) | 2011-11-11 | 2015-10-09 | Fraunhofer Ges Forschung | |
US9485503B2 (en) * | 2011-11-18 | 2016-11-01 | Qualcomm Incorporated | Inside view motion prediction among texture and depth view components |
WO2013072484A1 (en) | 2011-11-18 | 2013-05-23 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Multi-view coding with efficient residual handling |
US10154276B2 (en) * | 2011-11-30 | 2018-12-11 | Qualcomm Incorporated | Nested SEI messages for multiview video coding (MVC) compatible three-dimensional video coding (3DVC) |
US9288506B2 (en) | 2012-01-05 | 2016-03-15 | Qualcomm Incorporated | Signaling view synthesis prediction support in 3D video coding |
WO2013107931A1 (en) * | 2012-01-19 | 2013-07-25 | Nokia Corporation | An apparatus, a method and a computer program for video coding and decoding |
CN104081780A (en) * | 2012-01-31 | 2014-10-01 | 索尼公司 | Image processing apparatus and image processing method |
KR101669524B1 (en) | 2012-02-01 | 2016-11-09 | 노키아 테크놀로지스 오와이 | Method and apparatus for video coding |
CN103392189B (en) | 2012-02-23 | 2017-10-31 | 史克威尔·艾尼克斯控股公司 | dynamic image transmission server, dynamic image playing device and control method |
KR20130098122A (en) | 2012-02-27 | 2013-09-04 | 세종대학교산학협력단 | Device and method for encoding/decoding |
WO2013129822A1 (en) * | 2012-02-27 | 2013-09-06 | 세종대학교산학협력단 | Image encoding and decoding apparatus, and image encoding and decoding method |
US10447990B2 (en) | 2012-02-28 | 2019-10-15 | Qualcomm Incorporated | Network abstraction layer (NAL) unit header design for three-dimensional video coding |
US9445076B2 (en) | 2012-03-14 | 2016-09-13 | Qualcomm Incorporated | Disparity vector construction method for 3D-HEVC |
US9525861B2 (en) | 2012-03-14 | 2016-12-20 | Qualcomm Incorporated | Disparity vector prediction in video coding |
US9503720B2 (en) | 2012-03-16 | 2016-11-22 | Qualcomm Incorporated | Motion vector coding and bi-prediction in HEVC and its extensions |
US10200709B2 (en) | 2012-03-16 | 2019-02-05 | Qualcomm Incorporated | High-level syntax extensions for high efficiency video coding |
KR101536501B1 (en) | 2012-04-12 | 2015-07-13 | 신라 테크놀로지스, 인크. | Moving image distribution server, moving image reproduction apparatus, control method, recording medium, and moving image distribution system |
CN104365104B (en) * | 2012-04-16 | 2018-11-06 | 三星电子株式会社 | For multiple view video coding and decoded method and apparatus |
US9549180B2 (en) * | 2012-04-20 | 2017-01-17 | Qualcomm Incorporated | Disparity vector generation for inter-view prediction for video coding |
US20150117514A1 (en) * | 2012-04-23 | 2015-04-30 | Samsung Electronics Co., Ltd. | Three-dimensional video encoding method using slice header and method therefor, and three-dimensional video decoding method and device therefor |
US20130287093A1 (en) * | 2012-04-25 | 2013-10-31 | Nokia Corporation | Method and apparatus for video coding |
WO2013176485A1 (en) * | 2012-05-22 | 2013-11-28 | 엘지전자 주식회사 | Method and device for processing video signal |
EP2670146A1 (en) * | 2012-06-01 | 2013-12-04 | Alcatel Lucent | Method and apparatus for encoding and decoding a multiview video stream |
US9307252B2 (en) | 2012-06-04 | 2016-04-05 | City University Of Hong Kong | View synthesis distortion model for multiview depth video coding |
US20130336405A1 (en) * | 2012-06-15 | 2013-12-19 | Qualcomm Incorporated | Disparity vector selection in video coding |
US9872041B2 (en) | 2012-08-10 | 2018-01-16 | Lg Electronics Inc. | Method and apparatus for transceiving image component for 3D image |
KR101754999B1 (en) * | 2012-08-29 | 2017-07-06 | 브이아이디 스케일, 인크. | Method and apparatus of motion vector prediction for scalable video coding |
KR102160242B1 (en) * | 2012-09-09 | 2020-09-25 | 엘지전자 주식회사 | Image decoding method and apparatus using same |
WO2014047351A2 (en) | 2012-09-19 | 2014-03-27 | Qualcomm Incorporated | Selection of pictures for disparity vector derivation |
JP6000463B2 (en) | 2012-09-21 | 2016-09-28 | 聯發科技股▲ふん▼有限公司Mediatek Inc. | Method and apparatus for virtual depth value of 3D video encoding |
SG11201502194QA (en) | 2012-09-21 | 2015-04-29 | Nokia Technologies Oy | Method and apparatus for video coding |
US9392268B2 (en) * | 2012-09-28 | 2016-07-12 | Qualcomm Incorporated | Using base layer motion information |
MY186413A (en) * | 2012-09-28 | 2021-07-22 | Sony Corp | Image processing device and method |
KR101835358B1 (en) | 2012-10-01 | 2018-03-08 | 지이 비디오 컴프레션, 엘엘씨 | Scalable video coding using inter-layer prediction contribution to enhancement layer prediction |
CN104704837A (en) * | 2012-10-03 | 2015-06-10 | 联发科技股份有限公司 | Method and apparatus for inter-component motion prediction in three-dimensional video coding |
WO2014053099A1 (en) * | 2012-10-03 | 2014-04-10 | Mediatek Inc. | Method and apparatus for motion information inheritance in three-dimensional video coding |
US9699450B2 (en) * | 2012-10-04 | 2017-07-04 | Qualcomm Incorporated | Inter-view predicted motion vector for 3D video |
KR20140048783A (en) * | 2012-10-09 | 2014-04-24 | 한국전자통신연구원 | Method and apparatus for deriving motion information by sharing depth information value |
CN104782123A (en) * | 2012-10-22 | 2015-07-15 | 数码士控股有限公司 | Method for predicting inter-view motion and method for determining inter-view merge candidates in 3d video |
EP2919463A4 (en) * | 2012-11-07 | 2016-04-20 | Lg Electronics Inc | Method and apparatus for processing multiview video signal |
US20150304676A1 (en) * | 2012-11-07 | 2015-10-22 | Lg Electronics Inc. | Method and apparatus for processing video signals |
CN104838658B (en) * | 2012-12-14 | 2018-07-20 | 高通股份有限公司 | Interior views motion prediction in texture and depth views component with asymmetric spatial resolution |
US9247256B2 (en) | 2012-12-19 | 2016-01-26 | Intel Corporation | Prediction method using skip check module |
US9584792B2 (en) | 2013-01-04 | 2017-02-28 | Qualcomm Incorporated | Indication of current view dependency on reference view in multiview coding file format |
US9762905B2 (en) * | 2013-03-22 | 2017-09-12 | Qualcomm Incorporated | Disparity vector refinement in video coding |
JP2014187580A (en) * | 2013-03-25 | 2014-10-02 | Kddi Corp | Video encoder, video decoder, video encoding method, video decoding method, and program |
WO2014163468A1 (en) * | 2013-04-05 | 2014-10-09 | 삼성전자 주식회사 | Interlayer video encoding method and apparatus for using view synthesis prediction, and video decoding method and apparatus for using same |
WO2014166096A1 (en) * | 2013-04-11 | 2014-10-16 | Mediatek Singapore Pte. Ltd. | Reference view derivation for inter-view motion prediction and inter-view residual prediction |
CN105103543B (en) * | 2013-04-12 | 2017-10-27 | 寰发股份有限公司 | Compatible depth relies on coding method |
KR20160021222A (en) * | 2013-06-18 | 2016-02-24 | 브이아이디 스케일, 인크. | Inter-layer parameter set for hevc extensions |
WO2015000108A1 (en) * | 2013-07-01 | 2015-01-08 | Mediatek Singapore Pte. Ltd. | An improved texture merging candidate in 3dvc |
WO2015007159A1 (en) * | 2013-07-15 | 2015-01-22 | Mediatek Singapore Pte. Ltd. | Method of disparity derived depth coding in 3d video coding |
WO2015006922A1 (en) * | 2013-07-16 | 2015-01-22 | Mediatek Singapore Pte. Ltd. | Methods for residual prediction |
WO2015006924A1 (en) * | 2013-07-16 | 2015-01-22 | Mediatek Singapore Pte. Ltd. | An additional texture merging candidate |
WO2015006883A1 (en) * | 2013-07-18 | 2015-01-22 | Qualcomm Incorporated | Motion vector inheritance techniques for depth coding |
CN105393539B (en) * | 2013-07-24 | 2019-03-29 | 高通股份有限公司 | The sub- PU motion prediction decoded for texture and depth |
ES2906238T3 (en) | 2013-07-24 | 2022-04-13 | Qualcomm Inc | Simplified Advanced Motion Prediction for 3D-HEVC |
EP3025498B1 (en) * | 2013-08-13 | 2019-01-16 | HFI Innovation Inc. | Method of deriving default disparity vector in 3d and multiview video coding |
US9813736B2 (en) * | 2013-09-27 | 2017-11-07 | Qualcomm Incorporated | Inter-view dependency type in MV-HEVC |
US10382752B2 (en) * | 2013-10-15 | 2019-08-13 | Sony Corporation | Image processing device and method |
US10045048B2 (en) | 2013-10-18 | 2018-08-07 | Lg Electronics Inc. | Method and apparatus for decoding multi-view video |
US9948950B2 (en) * | 2014-01-03 | 2018-04-17 | Qualcomm Incorporated | Disparity vector and/or advanced residual prediction for video coding |
WO2015109598A1 (en) * | 2014-01-27 | 2015-07-30 | Mediatek Singapore Pte. Ltd. | Methods for motion parameter hole filling |
CN106105212A (en) | 2014-03-07 | 2016-11-09 | 高通股份有限公司 | Sub-predicting unit (SUB PU) kinematic parameter simplifying inherits (MPI) |
WO2015135137A1 (en) * | 2014-03-11 | 2015-09-17 | Mediatek Singapore Pte. Ltd. | A method of motion information sharing in multi-view and 3d video coding |
CN105187824A (en) | 2014-06-10 | 2015-12-23 | 杭州海康威视数字技术股份有限公司 | Image coding method and device, and image decoding method and device |
CN107409214B (en) * | 2015-01-21 | 2021-02-02 | 三星电子株式会社 | Method and apparatus for decoding inter-layer video and method and apparatus for encoding inter-layer video |
US10455242B2 (en) * | 2015-03-04 | 2019-10-22 | Qualcomm Incorporated | Signaling output indications in codec-hybrid multi-layer video coding |
EP3316575A1 (en) * | 2016-10-31 | 2018-05-02 | Thomson Licensing | Method for providing continuous motion parallax effect using an auto-stereoscopic display, corresponding device, computer program product and computer-readable carrier medium |
US10916041B2 (en) | 2018-03-30 | 2021-02-09 | Samsung Electronics Co., Ltd. | Method for depth image di coding |
US10623791B2 (en) | 2018-06-01 | 2020-04-14 | At&T Intellectual Property I, L.P. | Field of view prediction in live panoramic video streaming |
US10812774B2 (en) | 2018-06-06 | 2020-10-20 | At&T Intellectual Property I, L.P. | Methods and devices for adapting the rate of video content streaming |
US10616621B2 (en) | 2018-06-29 | 2020-04-07 | At&T Intellectual Property I, L.P. | Methods and devices for determining multipath routing for panoramic video content |
US10708494B2 (en) | 2018-08-13 | 2020-07-07 | At&T Intellectual Property I, L.P. | Methods, systems and devices for adjusting panoramic video content |
US11019361B2 (en) | 2018-08-13 | 2021-05-25 | At&T Intellectual Property I, L.P. | Methods, systems and devices for adjusting panoramic view of a camera for capturing video content |
WO2021087800A1 (en) * | 2019-11-06 | 2021-05-14 | Oppo广东移动通信有限公司 | Information processing method, encoding apparatus, decoding apparatus, system, and storage medium |
CN111885386B (en) * | 2020-07-29 | 2022-07-26 | 北京市商汤科技开发有限公司 | Image compression method, image decompression method, image compression device, image decompression device, electronic equipment and storage medium |
US20220279204A1 (en) * | 2021-02-26 | 2022-09-01 | Qualcomm Incorporated | Efficient video encoder architecture |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2002163678A (en) * | 2000-09-13 | 2002-06-07 | Monolith Co Ltd | Method and device for generating pseudo three- dimensional image |
US7728878B2 (en) * | 2004-12-17 | 2010-06-01 | Mitsubishi Electric Research Labortories, Inc. | Method and system for processing multiview videos for view synthesis using side information |
MY159176A (en) * | 2005-10-19 | 2016-12-30 | Thomson Licensing | Multi-view video coding using scalable video coding |
KR101101965B1 (en) * | 2006-10-16 | 2012-01-02 | 노키아 코포레이션 | System and method for using parallelly decodable slices for multi-view video coding |
KR101885790B1 (en) * | 2007-04-12 | 2018-08-06 | 돌비 인터네셔널 에이비 | Tiling in video encoding and decoding |
US20100284466A1 (en) * | 2008-01-11 | 2010-11-11 | Thomson Licensing | Video and depth coding |
WO2009130561A1 (en) * | 2008-04-21 | 2009-10-29 | Nokia Corporation | Method and device for video coding and decoding |
-
2009
- 2009-10-16 CN CN200980151318.7A patent/CN102257818B/en active Active
- 2009-10-16 US US13/124,641 patent/US9973739B2/en active Active
- 2009-10-16 EP EP09820314A patent/EP2338281A4/en not_active Ceased
- 2009-10-16 WO PCT/FI2009/050834 patent/WO2010043773A1/en active Application Filing
-
2018
- 2018-05-14 US US15/979,121 patent/US10306201B2/en active Active
-
2019
- 2019-05-23 US US16/420,729 patent/US10715779B2/en active Active
Also Published As
Publication number | Publication date |
---|---|
CN102257818A (en) | 2011-11-23 |
EP2338281A1 (en) | 2011-06-29 |
WO2010043773A1 (en) | 2010-04-22 |
US10306201B2 (en) | 2019-05-28 |
EP2338281A4 (en) | 2012-08-15 |
US20180262742A1 (en) | 2018-09-13 |
US9973739B2 (en) | 2018-05-15 |
US20190281270A1 (en) | 2019-09-12 |
US20110216833A1 (en) | 2011-09-08 |
US10715779B2 (en) | 2020-07-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102257818B (en) | Sharing of motion vector in 3d video coding | |
CN111543060B (en) | Apparatus, method and computer program for video encoding and decoding | |
CN101627634B (en) | System and method for using parallelly decodable slices for multi-view video coding | |
Chen et al. | Overview of the MVC+ D 3D video coding standard | |
CN107852532B (en) | Method and apparatus for video encoding | |
KR101682999B1 (en) | An apparatus, a method and a computer program for video coding and decoding | |
CN101548548B (en) | System and method for providing picture output indications in video coding | |
KR102224703B1 (en) | An apparatus, a method and a computer program for video coding and decoding | |
CN106105220B (en) | For Video coding and decoded method and apparatus | |
CN103907347B (en) | Multi-view video coding and decoding | |
CN104769948B (en) | A kind of decoding method, device and readable storage medium storing program for executing | |
CN102918836B (en) | Frame for asymmetric stereo video encapsulates | |
US9942558B2 (en) | Inter-layer dependency information for 3DV | |
CN101558652B (en) | System and method for implementing low-complexity multi-view video coding | |
CN108702503A (en) | For Video coding and decoded device, method and computer program | |
CN109155861A (en) | Method and apparatus and computer program for coded media content | |
JP2018534824A (en) | Video encoding / decoding device, method, and computer program | |
EP3018908B1 (en) | Method and apparatus for decoding video including a plurality of layers | |
KR20100074280A (en) | Motion skip and single-loop encoding for multi-view video content | |
CN105027567A (en) | Method and apparatus for video coding and decoding | |
CN105325003A (en) | An apparatus, a method and a computer program for video coding and decoding | |
WO2012039936A1 (en) | Coding stereo video data | |
CN105027569A (en) | An apparatus, a method and a computer program for video coding and decoding | |
Miličević et al. | EXTENSIONS OF HIGH EFFICIENCY VIDEO CODING STANDARD: AN OVERVIEW |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
C41 | Transfer of patent application or patent right or utility model | ||
TR01 | Transfer of patent right |
Effective date of registration: 20160215 Address after: Espoo, Finland Patentee after: Technology Co., Ltd. of Nokia Address before: Espoo, Finland Patentee before: Nokia Oyj |