WO2012116638A2 - 3d格式描述信息的获取方法和装置 - Google Patents

3d格式描述信息的获取方法和装置 Download PDF

Info

Publication number
WO2012116638A2
WO2012116638A2 PCT/CN2012/071767 CN2012071767W WO2012116638A2 WO 2012116638 A2 WO2012116638 A2 WO 2012116638A2 CN 2012071767 W CN2012071767 W CN 2012071767W WO 2012116638 A2 WO2012116638 A2 WO 2012116638A2
Authority
WO
WIPO (PCT)
Prior art keywords
format
video
information
description information
acquiring
Prior art date
Application number
PCT/CN2012/071767
Other languages
English (en)
French (fr)
Other versions
WO2012116638A3 (zh
Inventor
张园园
惠宇
石腾
张楚雄
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Priority to EP12752211.8A priority Critical patent/EP2675181A4/en
Publication of WO2012116638A2 publication Critical patent/WO2012116638A2/zh
Publication of WO2012116638A3 publication Critical patent/WO2012116638A3/zh
Priority to US14/014,767 priority patent/US20140002593A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/172Processing image signals image signals comprising non-image signal components, e.g. headers or format information
    • H04N13/178Metadata, e.g. disparity information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/435Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/63Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
    • H04N21/643Communication protocols
    • H04N21/6437Real-time Transport Protocol [RTP]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/816Monomedia components thereof involving special video data, e.g 3D video
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/84Generation or processing of descriptive data, e.g. content descriptors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/434Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams, extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
    • H04N21/4347Demultiplexing of several video streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/434Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams, extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
    • H04N21/4348Demultiplexing of additional data and video streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8543Content authoring using a description language, e.g. Multimedia and Hypermedia information coding Expert Group [MHEG], eXtensible Markup Language [XML]

Definitions

  • the present invention relates to the field of information technology, and in particular, to a method and an apparatus for acquiring 3D format description information.
  • 3D (Three Dimensional; 3D) video can be used in different formats for transmission or storage.
  • Common 3D format types include Frame Packing (FP), Two Dimensional Video Plus Auxiliary Video (2DA), and Simultaneous Broadcasting. (Simulcast; SC) and so on.
  • the client In order to properly process 3D video, the client needs to obtain some information, which allows the client to determine which format the 3D video uses, so that the client can adjust to the state of the received 3D video format, and then perform 3D video. Processing, for example, processing the 3D video to obtain a left view for projection to the left eye of the person, and a right view for projection to the right eye of the person.
  • the information that the client needs to obtain is collectively referred to as 3D format description information.
  • the 3D format description information describing the frame packing format is encapsulated into a frame packing arrangement SB message, and then the frame packing arrangement SB message is encapsulated into the video. Transmitted in the bitstream.
  • the video bitstream is the encoded video.
  • the client After receiving the video bit stream, the client obtains the frame packing arrangement supplementary enhancement information message from the video bit stream, and then obtains the 3D format description information describing the frame packing format.
  • the multimedia system is usually a heterogeneous system, there may be both a 2D client and a 3D client; all are 3D clients, and some clients may support a frame wrap format, and some clients support a 2D plus auxiliary video format; Both are 3D clients that support the frame wrap format. Some clients may only support side-by-side and top-and-bottom frame wrappers, but not the checkerboard format. They are 3D clients that support 2D plus auxiliary video formats. Some clients do not support auxiliary video as a depth map. This may lead to some cases where the client does not support the 3D format adopted by a 3D video.
  • the frame packing arrangement supplemental enhancement information is carried by the video bit stream and is rotated in the transmission mode, that is, the transmission is performed at intervals, this may cause the client of the multimedia system to receive the video bit stream.
  • the frame packing arrangement supplemental enhancement information can be obtained, and then the 3D format description information describing the frame packing format can be obtained, and then the 3D format adopted by the received video can be determined.
  • the existing method for transmitting the 3D format description information may cause the user to wait for a certain period of time to determine that the client does not support the 3D format adopted by the 3D video, so that the 3D video cannot be correctly processed and displayed, which affects the user experience.
  • the overhead of receiving and processing video is increased, which increases power consumption, and in particular increases the burden on power-sensitive mobile clients.
  • the embodiment of the invention provides a method and a device for acquiring 3D format description information, which can solve the defect that the client can receive the 3D format description information after receiving the video for a period of time in the prior art, and shorten the client to determine whether to support the video adoption.
  • the time of the 3D format is not limited to a period of time in the prior art.
  • An embodiment of the present invention provides a method for acquiring 3D format description information, including:
  • the embodiment of the invention further provides a method for acquiring 3D format description information, including:
  • the metadata portion of the 3D video file carrying 3D format description information; parsing a metadata portion of the 3D video file, and acquiring the 3D format description information from the metadata portion.
  • the embodiment of the invention further provides a device for acquiring 3D format description information, including:
  • the receiving module is configured to receive an outband message that is sent by the sending end and that carries the 3D format description information, where the receiving module receives the outband message before the client joins and participates in the multimedia session initiated by the sending end;
  • a parsing module configured to parse the out-of-band message received by the receiving module, and obtain the 3D format description information from the out-of-band message.
  • the embodiment of the present invention further provides another apparatus for acquiring 3D format description information, including: an obtaining module, configured to acquire a 3D video file, where a metadata part of the 3D video file carries 3D format description information;
  • a parsing module configured to parse a metadata portion of the 3D video file acquired by the obtaining block, and obtain the 3D format description information from the metadata portion.
  • the client can Obtain the 3D format description information before obtaining the video, so that the client can determine whether to support the 3D format adopted by the 3D video before receiving the video, and determine the 3D format adopted by the client to support the 3D video, and then obtain the video.
  • the time for the client to determine the 3D format used by the video is shortened, the overhead of receiving and processing the video is reduced, power consumption is reduced, and the burden on the receiving device is reduced.
  • FIG. 1 is a flowchart of a first embodiment of a method for acquiring 3D format description information provided by the present invention
  • FIG. 2 is a flowchart of a second embodiment of a method for acquiring 3D format description information provided by the present invention
  • FIG. 4 is a flowchart of a fourth embodiment of a method for acquiring 3D format description information provided by the present invention
  • FIG. 5 is a flowchart of 3D format description information provided by the present invention
  • FIG. 6 is a flowchart of a sixth embodiment of a method for acquiring 3D format description information provided by the present invention
  • FIG. 7 is a first embodiment of a device for acquiring 3D format description information provided by the present invention; Schematic diagram of the structure;
  • FIG. 8 is a schematic structural diagram of a second embodiment of an apparatus for acquiring 3D format description information provided by the present invention.
  • FIG. 1 is a flowchart of a first embodiment of a method for acquiring 3D format description information provided by the present invention. As shown in FIG. 1, the method includes:
  • S101 Receive an outband message that is sent by the sending end and that carries the 3D format description information.
  • the client receives the outband message before adding and participating in the multimedia session initiated by the sending end.
  • the outband message carrying the 3D format description information is a message that is received by the receiving device from the multimedia session initiated by the sending end.
  • the client receives the out-of-band message before joining and participating in the multimedia session initiated by the sender. That is, the receiving device can receive the out-of-band message before receiving the media stream transmitted by the transmitting end.
  • the out-of-band message can be various messages transmitted between the transmitting device and the receiving device.
  • the outband message may be a Session Description Protocol (SDP) file.
  • SDP Session Description Protocol
  • the sender needs to send the SDP file to the client first. Send a video to the client. Therefore, the sending end can carry the 3D format description information in the SDP file, so that the client obtains the 3D format description information carried in the S D P file before joining and participating in the multimedia session.
  • the 3D format description information may be included in the attribute of the S D P file.
  • the electronic program guide (EPG) metadata of the receiving device of the client since the electronic program guide (EPG) metadata of the receiving device of the client first acquires the content according to the EPG metadata, the media stream is started to be received. Therefore, the out-of-band message may be an EPG. Metadata, the sender can carry the 3D format description information in the EPG metadata. Specifically, the 3D format description information may be included in an Extensible Markup Language (XML) element or attribute of the EPG metadata.
  • XML Extensible Markup Language
  • the notification message associated with the program content is sent slightly ahead of the program content. Therefore, the out-of-band message may also be a notification message, and the sender carries the 3D format description information through the notification message, and specifically includes the 3D format description information in the notification message payload.
  • This embodiment only enumerates specific types of out-of-band messages in several systems or business processes, but is not intended to be a limitation of the present invention.
  • the 3D format description information may be used to indicate the 3D format type identification information, and the 3D format type identification information may further include a 3D format type identifier, and may also include a component type identifier. In addition, the 3D format description information may further include 3D video processing parameter information.
  • the client can obtain the 3D format description information through the outband information before acquiring the video, so that the client can receive the video before receiving the video.
  • determine whether to support the 3D format adopted by 3D video determine the 3D format used by the client to support 3D video before acquiring the video. It shortens the time for the client to determine whether to support the 3D format adopted by 3D video, reduces the overhead of receiving and processing video, reduces power consumption, and reduces the burden on the receiving device.
  • FIG. 2 is a flowchart of a second embodiment of a method for acquiring 3D format description information provided by the present invention. As shown in FIG. 2, this embodiment is applicable to a multimedia service process, and carries a band of 3D format description information in a multimedia service flow.
  • the external information is an SDP file.
  • the sender can carry the 3D format description information in the SDP file, and the client receives the SDP file before joining and participating in the media session initiated by the sender. Therefore, the client can know which 3D format the video uses before receiving the video, so that the client can determine whether the client supports the 3D format adopted by the video before receiving the video.
  • the method includes:
  • S201 Receive a session description protocol SDP file sent by the sending end, where the SDP file carries 3D format description information.
  • the execution body of the foregoing step is a receiving device of the client.
  • the sender When the sender sends a 3D video to the client, it first sends an SDP file to the client, and carries the 3D format description information in the SDP file.
  • the specific method is to carry the 3D format description information in the attributes of the SDP file.
  • the client After receiving the SDP file, the client parses the SDP file, and determines whether the SDP file carries the 3D format description information.
  • the SDP file can be used to determine whether the SDP file contains the attributes of the 3D format description information, and the 3D format description information is obtained by parsing the attributes. .
  • the 3D format description information may include 3D format type identification information.
  • the 3D format type identification information includes a 3D format type identifier, and the 3D format type identifier indicates a format type adopted by the 3D video.
  • the 3D format type identification information may further include a component type identifier indicating a type of a video component constituting the 3D video.
  • the component type identifier indicates that the type of the video component is side by side (SBS), top and bottom (TAB), and line interleaved (LID, ⁇ ' J interlaced (CIL), Chessboard (CHB) frame sequence (SEQ); SEQ (frame sequence; SEQ) or any other frame-arranged video.
  • SBS side by side
  • TAB top and bottom
  • LID line interleaved
  • CIL Chessboard
  • SEQ frame sequence
  • SEQ frame sequence; SEQ
  • the component class Type identification indicates that the type of video component is 2D video, depth map, disparity map, hole data, including
  • the 2D video further includes a depth map, including both the 2D video and the disparity map, and the like; further, the component type identifier may further indicate whether the 2D video is carried in the left view, the right view, or the middle view. Any one. If the 3D format type is simulcast, the component type identifier indicates that the type of the video component is any one of a video carrying a left view and a video carrying a right view.
  • This embodiment provides an implementation manner of carrying 3D format type identification information by using an attribute in an SDP file, specifically:
  • the sender can carry the 3D format type identification information through the attribute 3dFormatType in the SDP file, and the attribute is a media level attribute.
  • the specific format is:
  • parameter ⁇ 30 ⁇ 0177131 1 ⁇ 6> is a 30-format type identifier
  • 1 ⁇ 6> is the component type identifier.
  • the value of ⁇ 3d format type> includes but is not limited to FP, 2DA, SC, etc., respectively, representing 3D format type as frame packing, 2D plus auxiliary video, simulcast;
  • ⁇ 3d format type is FP, ⁇ component type>
  • the values include, but are not limited to, SBS, TAB, LIL, CIL, CHB, SEQ, etc., respectively, representing the component of the video composing the 3D video as frame-packed video in the form of side by side, up and down, line interleaving, column interleaving, checkerboard, and frame sequence;
  • the value of ⁇ component type> includes but is not limited to 2d, D, P, 2dD, 2dP, etc., respectively, representing the component type of the video composing the 3D video as 2D video, depth map , disparity map, hole data, including 2D video and depth map, including 2D video and disparity map;
  • Each video component constituting the 3D video can indicate the 3D format type adopted by the 3D video composed of the video component and the type of the video component through the attribute 3dFormatType.
  • the video component constituting the 3D video does not pass the attribute 3dFormatType to indicate the 3D format type and the component type. Then, the 3D format type is 2D plus auxiliary video, and the component type is 2D video.
  • This embodiment also provides another type of identifier that carries the 3D format type by using an attribute in the SDP file.
  • the implementation of the information specifically:
  • the sender can carry the 3D format type identification information through the attribute fmpt, and the attribute fmpt is a media level attribute.
  • the specific format is:
  • ⁇ payload type> ⁇ 3d format type >[ ⁇ component type>]
  • ⁇ payload type> is the RTP payload type carrying 3D video.
  • the parameter ⁇ 3d format 1 ⁇ 6> is the 30 format type identifier; the optional parameter 3 ⁇ 4 ⁇ component type> is the component type identifier.
  • Each video component constituting the 3D video can indicate the 3D format type adopted by the 3D video composed of the video component and the component type of the video through the attribute fmtp.
  • the video component constituting the 3D video does not pass the attribute fmtp to indicate the 3D format type and the component type. Then, the 3D format type is 2D plus auxiliary video, and the component type is 2D video.
  • the 3D format description information may further include 3D video processing parameter information in addition to the 3D format type identification information.
  • the 3D video processing parameter information includes, but is not limited to, parameter information of a sampling type when performing frame packing, and identification information of a frame placement order when performing a frame packing operation;
  • the format type is 2D plus auxiliary video and the auxiliary video is a depth map
  • the 3D video processing parameter information includes but is not limited to a horizontal offset of the depth sample in the spatial sampling grid of the 2D video, a vertical offset, and a depth
  • the value range indication information that is, the maximum distance after the screen, the maximum distance before the screen, and the like;
  • the 3D format type is 2D plus auxiliary video and the auxiliary video is a disparity map
  • the 3D video processing parameter information includes However, it is not limited to the horizontal offset of the parallax sample in the spatial sampling grid of the 2D video, the vertical offset, the value representing the zero parallax, the scaling for defining the range of the disparity value, the reference viewing distance, and the reference screen.
  • Parameter information such as
  • the 3D video processing information can also be carried by an attribute in the SDP file.
  • This embodiment provides an implementation manner of carrying 3D video processing parameter information by using an attribute in an SDP file, specifically:
  • the 3D format type When the 3D format type is frame wrapper, it can be carried by the property FramePackingParameters. With corresponding 3D video processing parameter information.
  • the property FramePackingParameters is a media-level property.
  • the specific format is:
  • the ⁇ sampling type> indicates the sampling type when the frame is wrapped.
  • the values of the Sampling type include, but are not limited to, none, interleaved, quincunx, etc., which represent no sampling, alternating sampling, and five-point plum sampling.
  • the parameter ⁇ content interpretation type> indicates the frame placement order when the frame is wrapped.
  • the value is LFirst or RFirst, which means that the video frame corresponding to the left view is placed in front of the video frame corresponding to the front and right views.
  • the 3D format type is 2D plus auxiliary video and the auxiliary video is a depth map, you can pass the attribute.
  • DepthParameters carries the corresponding 3D video processing parameter information.
  • the property DepthParameters is a media-level property.
  • the specific format is:
  • the parameter ⁇ position offset 11> indicates the horizontal offset of the depth sample in the spatial sampling grid of the 2D video
  • the parameter ⁇ position offset v> indicates the vertical offset of the auxiliary video sample in the spatial sampling grid of the 2D video
  • the parameters ⁇ 111 ⁇ 31 and ⁇ 111 ⁇ 11631 are used to indicate the range of values for the depth samples, ⁇ nkfar; ⁇ is the maximum distance after the screen, and ⁇ nknear> indicates the maximum distance before the screen.
  • the 3D format type is 2D plus auxiliary video and the auxiliary video is a disparity map, you can pass the attribute.
  • ParallaxParameters carry the corresponding 3D video processing parameter information. Attributes ParallaxParameters are media-level attributes. The specific format is:
  • the parameter ⁇ position offset 11> indicates the horizontal offset of the disparity sample in the spatial sampling grid of the 2D video
  • the parameter ⁇ position offset ⁇ indicates the vertical offset of the disparity sample in the spatial sampling grid of the 2D video
  • the parameters ⁇ parallax zero>, ⁇ parallax scale>, ⁇ dref>, ⁇ wref> indicate which value represents zero parallax, the scaling used to define the range of disparity values, the reference viewing distance, and the reference screen width, respectively.
  • the above is only one possible implementation manner of carrying 3D video processing parameter information through attributes in the SDP file, but is not intended to limit the present invention.
  • 3D video may be composed of multiple video components, and different video components may be carried by different media streams.
  • a plurality of media streams may be described in one SDP file, and each media stream carries different video components, for example: SDP files describe media stream 1, media stream 2, media stream 3, and media stream 4.
  • SDP files describe media stream 1, media stream 2, media stream 3, and media stream 4.
  • the video component carried by the media stream 1 and the video component carried by the media stream 2 constitute a 3D video; the video component carried by the media stream 3 and the video component carried by the media stream 4 constitute another 3D video. Therefore, you need to tell the client which media streams constitute 3D video.
  • This embodiment provides an implementation manner of informing the client which media streams constitute a 3D video by using the attribute group and mid attributes in the SDP file, specifically:
  • the media stream identifier is defined by the attribute mid for different media streams that make up the 3D video.
  • the attribute mid is the media level attribute, which is the identifier of the media stream and is unique in an SDP file.
  • the specific format is:
  • the different media streams that make up the 3D video are grouped together by the attribute group.
  • the attribute group is a session-level attribute and is used to group several media streams identified by the attribute mid.
  • the specific format is:
  • 3D video is composed of different media streams that are grouped together.
  • the details of an SDP file are as follows, where: the video component in the media stream identified as 1 and the video component in the media stream identified as 2 constitute a 3D video.
  • the video component in the media stream identified as 1 is the video of the left view in the 3D video of the simulcast format
  • the video component in the media stream identified as 2 is the video of the right view in the 3D video carrying the simulcast format.
  • the video component in the media stream identified as 3 and the video component in the media stream identified as 4 constitute a 3D video.
  • the video component in the media stream identified as 3 is the auxiliary video in the 2D plus auxiliary video format 3D video and the auxiliary video is a depth map.
  • An SDP file can be as follows:
  • the 3D format description information carried in the SDP file by the sending end may include 3D format type identification information, and may further include 3D video processing parameter information, and the receiving device of the client may obtain the 3D format from the acquired SDP file.
  • the type identifies the information to determine the format in which the 3D video is used.
  • the client may also obtain 3D video processing parameter information from the obtained SDP file to perform corresponding processing on the subsequently received 3D video.
  • FIG. 3 is a flowchart of a third embodiment of a method for acquiring 3D format description information provided by the present invention.
  • the method provided in this embodiment is mainly applicable to a Real-time Transport Protocol (RTP).
  • RTP Real-time Transport Protocol
  • the multimedia system the receiving device of the client will start to obtain the video from the media stream after receiving the media stream for a period of time. Therefore, the sender can place the 3D video
  • the parameter information is carried in the media stream, and carries the 3D format type identification information and the indication message of the 3D video processing parameter information in the SDP file, so that the client can obtain the 3D format type identification information from the SDP file, and from the media stream. Get 3D video processing parameter information.
  • the method includes:
  • S301 Receive a session description protocol SDP file sent by the sending end, where the SDP file carries
  • 3D format description information where the 3D format description information is 3D format type identification information and 3D video processing parameter information indication information, where the indication information is used to identify a location of the 3D video processing parameter information in the media stream;
  • the SDP file is parsed, and the indication information of the 3D format type identification information and the 3D video processing parameter information is obtained from the SDP file.
  • the media stream sent by the sending end is received, and the 3D video processing parameter information is obtained from the media stream according to the indication information.
  • the execution body of the above step is the receiving device of the client.
  • the sending end When sending the 3D video to the client, the sending end first sends an SDP file to the client, and carries the 3D format type identification information and the indication information of the 3D video processing parameter information in the SDP file.
  • the client After receiving the SDP file, the client parses the SDP file to determine whether the SDP file carries the 3D format, 3D format type identifier information, and 3D video processing parameter information. If carried, obtain the indication information of the 3D format type identification information and the 3D video processing parameter information.
  • the client obtains a media stream constituting the 3D video, and acquires 3D video processing parameter information from a corresponding position in the media stream according to the indication information.
  • the 3D format type identifier information is carried by the attribute in the SDP file, and the client obtains the 3D format type identifier information by parsing the attribute.
  • the media stream is in the RTP packet as the transmission unit, and the RTP packet is divided into an RTP header and an RTP payload.
  • the RTP header is divided into two parts: fixed header and extended header. Therefore, the sender can carry the 3D video processing parameter information through the payload portion of the RTP packet or through the extended packet header portion of the RTP packet.
  • the transmitting end carries the RTP packet of the 3D video processing parameter through the payload part, and transmits the same RTP packet for carrying the corresponding 3D video component through the same media stream.
  • the sender can carry the indication information of the 3D video processing parameter information through the attributes in the SDP file.
  • the indication information indicates an RTP payload type number carrying 3D video processing parameter information.
  • an implementation manner of carrying 3D video processing parameter information through an RTP payload is provided, specifically:
  • the message carrying the corresponding 3D video processing parameter information may be encapsulated into the RTP payload, and the message carrying the corresponding 3D video processing parameter information may be sei_rbsp().
  • the message carrying the corresponding 3D video processing parameter information can be encapsulated into the RTP payload.
  • the message carrying the corresponding 3D video processing parameter information may specifically be si-rbsp()
  • the embodiment provides an implementation manner of carrying the indication information of the 3D video processing parameter information by using the attribute in the SDP file, where the load part of the RTP packet carries the 3D video processing parameter information, where the indication information is specific The type number of the RTP payload. specific:
  • the RTP payload type number carrying the 3D video processing parameter information can be indicated by the attribute rtpmap.
  • the attribute rtpmap is a media-level attribute that identifies the meaning of the payload format represented by the payload type number.
  • the specific format is:
  • the semantics of the parameter ⁇ encoding name is the 3D video processing parameter carried by the RTP payload, for example, the value is 3d Parameters, and the corresponding parameter ⁇ payload type> takes the value.
  • the RTP payload type number of the 3D video processing parameter information is the RTP payload type number of the 3D video processing parameter information.
  • the client first obtains the SDP file, and can obtain the RTP payload type number carrying the 3D video processing parameter information from the attribute rtpmap in the SDP file.
  • the client obtains the corresponding media stream
  • the RTP packet with the same PT (payload type) segment value as the obtained payload type is obtained from the media stream according to the obtained load type number, from the The RTP packet payload portion of the RTP packet obtains the 3D video processing parameter information.
  • the sending end may also carry the 3D video processing parameter information in the RTP extension header of the RTP packet for carrying the corresponding 3D video component.
  • the sending end carries the indication of the 3D video processing parameter information by using the attribute in the SDP file.
  • Information the indication information is used to indicate an identifier of an extension item carrying 3D video processing parameter information.
  • This embodiment provides an implementation manner of carrying 3D video processing parameter information by using an RTP extended packet header, specifically: If the 3D format type is a frame wrapper, the corresponding 3D video processing parameter information can be carried by an extension item. Specifically, the sei-rbsp() message carrying the corresponding 3D video processing parameter information may be encapsulated into the extension, and the extension is encapsulated into an RTP extension header of the RTP packet carrying the corresponding 3D video component.
  • the corresponding 3D video processing parameter information can be carried by an extension item.
  • the si-rbsp() message carrying the corresponding 3D video processing parameter information may be encapsulated into the extension, and the extension is encapsulated into an RTP extension header of an RTP packet carrying the corresponding 3D video component.
  • This embodiment also provides an implementation manner of carrying 3D video processing parameter information by using an RTP extended packet header, specifically:
  • the parameter information such as the identification information of the sampling type in the frame wrapping and the identification information of the frame placement sequence in the frame packing operation may be respectively carried in the corresponding 3D video processing parameter information by using different extension items;
  • the extension is encapsulated into an RTP extension header of an RTP packet carrying the corresponding 3D video component.
  • the horizontal offset and vertical of the depth samples in the corresponding 3D video processing parameter information in the spatial sampling grid of the 2D video may be respectively carried by different extension items.
  • the offset, the range of the depth sample indicates parameter information such as the maximum distance after the screen, the maximum distance before the screen, and the like; and the extension is encapsulated into the RTP extension of the RTP packet carrying the corresponding 3D video component. In the header.
  • the horizontal offset and vertical of the disparity samples in the corresponding 3D video processing parameter information in the spatial sampling grid of the 2D video may be respectively carried by different extension items.
  • the extension is encapsulated into an RTP extension header that carries the RTP packet of the corresponding 3D video component.
  • only the extensions carrying the 3D video processing parameter information may be encapsulated into the RTP extension header of the RTP packet carrying the key frame of the corresponding 3D video component.
  • an implementation manner of carrying indication information of 3D video processing parameter information through an attribute in an SDP file is provided in the case where the RTP extended packet header carries the 3D video processing parameter information. specifically is:
  • the identifier of the extension carrying the 3D video processing parameter information is indicated by the attribute extmap.
  • Attributes An extmap can be a media-level attribute or a session-level attribute. Correspondence between the identifier and meaning of the extension item. The specific format is:
  • the extension carries the 3D video processing parameter information, for example, the value is urn: Example:params:3dParameters, the parameter ⁇ value is the identifier of the extension that carries the 3D video processing parameter information.
  • the client first obtains the SDP file, which can be obtained from the attribute extmap in the SDP file.
  • the RTP packet containing the extended packet header is first obtained, which may be an RTP packet with an X (extension) extension of the RTP header.
  • the RTP packet containing the extended packet header is first obtained, which may be an RTP packet with an X (extension) extension of the RTP header.
  • the 3D video processing parameter information is obtained in the extension item.
  • the sending end may further carry the 3D video processing parameter information in the video bitstream. Specifically, it can be:
  • the enhanced information message is carried by the frame wrap in the video bitstream to carry the 3D video processing parameter information.
  • the 3D video processing parameter information is carried by the si-rbsp() message in the video bitstream.
  • the transmitting end may carry the indication information of the 3D video processing parameter information by using an attribute in the SDP file, where the indication information indicates the type of the video bit stream message carrying the 3D video processing parameter information.
  • This embodiment provides an implementation manner of carrying indication information of 3D video processing parameter information through attributes in an SDP file. specifically is:
  • the media-level attribute is used.
  • FramePackingArrangementS ⁇ PresentFlag indicates that the type of the video bitstream message carrying the 3D video processing parameter information is a frame packing supplemental enhancement information message.
  • the specific format is:
  • ⁇ value> 1 indicates that the video bitstream contains a frame packing supplemental enhancement information message carrying the corresponding 3D video processing parameter information.
  • a value of 0 means no.
  • the 3D format type is 2D plus auxiliary video
  • the type of the video bitstream message carrying the 3D video processing parameter information is indicated by the media level attribute SiRbspPresentFlag as a si-rbsp message.
  • the specific format is:
  • ⁇ value> is 1 indicates that the video bitstream contains a si-rbsp message carrying the corresponding 3D video processing parameter information.
  • a value of 0 means no.
  • FIG. 4 is a flowchart of a fourth embodiment of a method for acquiring 3D format description information provided by the present invention.
  • the 1" column is applicable to a television system, and the electronic program (EPG) metadata is used.
  • EPG electronic program
  • the user or the receiving device can browse and select the program through the EPG metadata, and then join the multimedia session corresponding to the participating program to obtain the program content including the video. Therefore, the transmitting end can be in the EPG element.
  • the data carries the 3D format description information, so that the receiving device of the client can obtain the 3D format description information before acquiring the video, thereby determining more quickly whether the receiving device matches the format adopted by the 3D video.
  • the method includes:
  • S401 Receive an EPG metadata of an electronic program menu sent by the sending end, where the EPG metadata carries 3D format description information;
  • the EPG metadata is metadata for generating an electronic program menu, and the user or the receiving device can browse and select the program through the electronic program menu. Then join the multimedia session corresponding to the participating program to obtain the program content.
  • the EPG metadata includes metadata describing channel information, metadata describing on-demand program information, and metadata describing live program information.
  • the transmitting end may carry 3D format description information in the EPG metadata describing the channel information, the on-demand program information, and the live program information, respectively, and provide 3D format description information for the channel content, the live program content, and the 3D video in the on-demand program content.
  • EPG metadata can be in the form of Extensible Markup Language (XML). It can extend the EPG metadata, add XML elements or attributes, and carry 3D format description information through new XML elements or attributes.
  • XML Extensible Markup Language
  • the 3D format description information includes 3D format type identification information.
  • This embodiment provides a method for carrying 3D by extending EPG metadata and adding XML elements or attributes.
  • the implementation of the format type identification information is specifically:
  • the XML element or attribute FramePackingType may be further added to indicate the frame wrapper type used by the frame wrapper video component constituting the 3D video.
  • AuxVideoType indicates the auxiliary video type used by the auxiliary video component constituting the 3D video.
  • an XML element or attribute StereolD may be further added to indicate the view identifier of the 2D video component constituting the 3D video.
  • the 3D format description information may further include 3D video processing parameter information.
  • This example shows an implementation method of carrying 3D video processing parameter information by extending EPG metadata and adding XML elements or attributes, specifically:
  • the 3D format type is frame wrapper, add the XML element FramePackingParameters to carry Corresponding 3D video processing parameter information.
  • the FramePackingParameters may include an XML element or attributes SamplingType, Contentl nterpretationType, respectively indicating the sampling type when the frame is wrapped, and the frame placement order when performing the frame wrapping operation;
  • the 3D format type is 2D plus auxiliary video and the auxiliary video type is depth map
  • the Depth Parameters may include an XML element or an attribute posit ion_offset_h, position_offset_v, nkfar, nknear, respectively indicating a horizontal offset, a vertical offset of the depth sample in the spatial sampling grid of the 2D video, The maximum distance after the screen, the maximum distance before the screen;
  • ParallaxParameters carries the corresponding 3D video processing parameter information.
  • ParallaxParameters may include XML elements or attributes position-offset-h, position-offset-v, parallax-zero, parallax-scale, dref, wref, respectively indicating the horizontal offset of the disparity samples in the spatial sampling grid of the 2D video, The vertical offset, the value representing zero parallax, the scaling used to define the range of disparity values, the reference viewing distance, and the reference screen width.
  • the 3D format description information can be obtained from the EPG metadata. If the EPG metadata includes an XML element for carrying 3D format description information, the XM L element may be parsed to obtain the 3D format description information.
  • FIG. 5 is a flowchart of a fifth embodiment of a method for acquiring 3D format description information provided by the present invention. As shown in FIG. 5, the embodiment is applicable to a television system. Before sending a 3D video to a client, the sender sends the client to the client first. The notification message related to the program content is sent. Therefore, the sender can carry the 3D format description information in the notification message, so that the receiving device of the client can quickly determine whether the format adopted by the receiving device and the 3D video matches.
  • the method includes:
  • the sender Before the sender sends the 3D video to the client, the sender first sends a notification message related to the program content.
  • the payload of the notification message carries 3D format description information.
  • the 3D format description information includes 3D format type identification information, and may further include 3D video processing parameter information.
  • the payload of the notification message may be an XML element, and the 3D format description information may be carried by adding an XML element or attribute.
  • the receiving device of the client can parse the XML element from the payload of the notification message to obtain the 3D format description information.
  • the sending end before the sending end sends the 3D video to the client, the sending end sends a notification message carrying the readable text to the terminal, and the receiving device receives the notification message and presents the notification message to the user.
  • the readable text can be used to prompt the user to wear 3D glasses or to prompt the user to view the optimal viewing distance of the 3D program.
  • FIG. 6 is a flowchart of a sixth embodiment of a method for acquiring 3D format description information provided by the present invention.
  • the client may obtain a 3D video file from a storage medium such as an optical disk or a mobile hard disk, or output from a sending end.
  • the 3D format description information can be carried in the metadata part of these 3D video files.
  • the 3D format description information is obtained from the metadata portion of the 3D video file.
  • the method provided in this embodiment includes:
  • the client After obtaining the 3D video file, the client obtains the metadata item carrying the 3D format description information from the metadata part of the 3D video file, parses the metadata item, and obtains the 3D format description information.
  • the 3D format description information corresponding to different 3D format types is carried by different metadata items.
  • the 3D format type is indicated by the type or name of the metadata item.
  • the corresponding 3D format description information of other 3D format types can be carried as the content of the metadata item.
  • the type or name of the metadata item and the content of the metadata item can be carried by different boxes.
  • the type or name of the metadata item can be carried via the Item Info Box; the content of the metadata item is carried via the Item Data Box.
  • the Box is then packaged into a Metadata Box, then the Metadata Box is packaged into a 3D video track box, and the 3D video track box is then packaged into a Movie Box; Movie Box is packaged into a file.
  • the corresponding 3D format description information may be encapsulated into a corresponding track box of the frame wrapper as a metadata item of the type fpdt; then the track box is encapsulated Go to the Movie Box; finally package the Movie Box into a file.
  • the type of the metadata item identified in the Item Info Box is fpdt, and the S ⁇ message carrying the corresponding 3D format description information of the frame wrapping format is packaged into the Item Data Box; and the Item Info Box and the Item Data Box are encapsulated into the Metadata Box. Medium; package the Metadata Box into the Track Box.
  • the corresponding 3D format description information may be encapsulated into a corresponding track box of the auxiliary video as a metadata item of the type sirp; then the corresponding track box of the auxiliary video is packaged into the Movie Box. Finally, the Movie Box is packaged into a file.
  • the type of metadata item identified in the Item Info Box is sirp; the si-rbsp() package is placed in the Item Data Box; the Item Info Box and the Item Data Box are encapsulated into the Metadata Box; The Metadata Box is packaged into the Track Box.
  • the corresponding 3D format description information may be encapsulated as a metadata item of type stvw into a corresponding track box of the video carrying the left view and a corresponding track box of the video carrying the right view; The above track box is packaged into the Movie Box; finally the Movie Box is packaged into a file.
  • a possible implementation manner of encapsulating the corresponding 3D format description information of the simulcast format type as a metadata item of the type stvw into a corresponding track box of the video carrying the left view or the video carrying the right view is:
  • the metadata item is identified in the Item Info Box as stvw; the stem_view-lnfo() package is placed in the Item Data Box; the Item Info Box and the Item Data Box are packaged in the Metadata Box; and the Metadata Box is packaged into the Track Box.
  • the stereo_id is used to indicate whether the carried view is left or right.
  • Reference— track— id Indicates the identity of the video track carrying another view.
  • Information After obtaining the 3D video file, the client obtains the box carrying the 3D format description information from the metadata part of the 3D video file, parses the box, and obtains the 3D format description information.
  • the embodiment further provides a specific implementation manner of carrying the 3D format description information in the box of the metadata part of the 3D video file, which specifically includes:
  • the type of Box indicates the type of 3D format.
  • the box of type spif is used to carry the 3D format description type corresponding to the 2d plus auxiliary video format type.
  • the type of stif is used to carry the corresponding 3D format description information of the simulcast format type.
  • Class FramePackingDataBox extends Box('fpdt') ⁇ Unsigned int(8) frame— packing— arrangement— type;
  • Class SupplementallnfoBox extends Box('spif ) ⁇
  • the client can obtain the 3D format description information through the metadata in the 3D video file before acquiring the video, so that the client can determine whether to support the 3D video before acquiring the video.
  • the 3D format is adopted to determine the 3D format used by the client to support 3D video before the video is acquired. Shortened the client's decision to support 3D video adoption The time of the 3D format reduces the overhead of receiving and processing video, reduces power consumption, and reduces the burden on the receiving device.
  • the storage medium may be a magnetic disk, an optical disk, a read-only memory (ROM), or a random access memory (RAM).
  • Figure 7 is a schematic structural diagram of a first embodiment of a device for acquiring 3D format description information provided by the present invention. As shown in Figure 7, the device includes: a receiving module 1 1 and a parsing module 12;
  • the receiving module 1 1 is configured to receive an outband message carrying 3D format description information, where the outband message is acquired before joining and participating in the multimedia session;
  • the parsing module 12 is configured to parse the out-of-band message acquired by the receiving module 1 to obtain the 3D format description information from the out-of-band message.
  • the 3D format description information carried in the outband message obtained by the receiving module 1 1 includes 3D format type identifier information; or the 3D format description information includes 3D format type identifier information and 3D video processing parameter information.
  • the 3D format type identification information may include a 3D format type identifier.
  • the 3D format type identification information may further include a 3D format type identifier and a component type identifier.
  • the out-of-band message acquired by the receiving module 11 is an SDP file.
  • the SDP file received by the receiving module 11 carries the 3D format type identification information and the indication information of the 3D video processing parameter information, where the indication information is used to indicate the 3D video processing parameter information, and is in the media stream. position.
  • the receiving module 1 1 may be further configured to: receive a media stream sent by the sending end, and obtain the 3D video processing parameter information from the media stream according to the indication information.
  • the indication information carried in the SDP file received by the receiving module 1 is an RTP payload type number
  • the receiving module 1 is further configured to: obtain the 3D video processing parameter information from a corresponding RTP load of the media stream according to the RTP payload type number.
  • the indication information carried in the SDP file received by the receiving module 11 may also be The identifier of the extension of the RTP header;
  • the receiving module 1 is further configured to: obtain the 3D video processing parameter information from a corresponding RTP packet header of the media stream according to an identifier of the extended item of the RTP packet header.
  • the out-of-band message received by the receiving module 11 may also be EPG metadata.
  • the out-of-band message received by the receiving module 11 may also be a notification message in the television system.
  • the device for acquiring the 3D format description information provided in this embodiment corresponds to the method for obtaining the 3D format description information provided by the present invention.
  • the specific implementation manner of the function device of the method is as described above. Method embodiment to fifth embodiment, and details are not described herein again.
  • the device for acquiring 3D format description information can enable the client to carry the 3D format description information from the outband message before acquiring the video, so that the client can determine whether to support the 3D used by the 3D video before receiving the video.
  • Format determine the 3D format used by the client to support 3D video before acquiring the video. It shortens the time for the client to determine whether to support the 3D format adopted by 3D video, reduces the overhead of receiving and processing video, reduces power consumption, and reduces the burden on the receiving device.
  • Figure 8 is a schematic structural diagram of a second embodiment of a device for acquiring 3D format description information provided by the present invention. As shown in Figure 8, the device includes: an obtaining module 21 and a parsing module 22;
  • the obtaining module 21 is configured to acquire a 3D video file, where the metadata part of the 3D video file carries 3D format description information;
  • the parsing module 22 is configured to parse the metadata part of the 3D video file acquired by the obtaining block 21, and obtain the 3D format description information from the metadata part.
  • the 3D format description information carried in the metadata part of the 3D video file acquired by the acquiring module 21 includes 3D format type identification information; or the 3D format description information includes 3D format type identification information and 3D video. Process parameter information.
  • the 3D format type identifier information includes a 3D format type identifier; or the 3D format type identifier information includes a 3D format type identifier and a component type identifier.
  • the apparatus for acquiring the 3D format description information provided in this embodiment corresponds to the sixth embodiment of the method for acquiring the 3D format description information provided by the present invention.
  • the specific implementation manner can be referred to the foregoing method. Six embodiments are not described herein.
  • the device for acquiring 3D format description information provided by the embodiment of the present invention can be obtained by the client
  • the video obtains the 3D format description information through the metadata in the 3D video file, so that the client can determine whether to support the 3D format adopted by the 3D video before receiving the video, and determine that the client supports the 3D format adopted by the 3D video. video. It shortens the time for the client to determine whether to support the 3D format adopted by 3D video, reduces the overhead of receiving and processing video, reduces power consumption, and reduces the burden on the receiving device.

Description

3D格式描述信息的获取方法和装置
本申请要求于 2011 年 03 月 02 日提交中国专利局、 申请号为 201110050253.6、 发明名称为 "3D格式描述信息的获取方法和装置" 的中国 专利申请的优先权, 其全部内容通过引用结合在本申请中。
技术领域
本发明涉及信息技术领域, 特别涉及一种 3D格式描述信息的获取方法 和装置。
背景技术
3D ( Three Dimensional; 3D )视频在传输或存储时可以采用不同格式, 常见的 3D格式类型包括帧包装( Frame Packing; FP )、二维加辅助视频( Two Dimensional video plus Auxiliary video; 2DA )、联播 ( Simulcast; SC )等。
为了正确处理 3D视频, 客户端需要获取一些信息, 这些信息可以让客 户端确定 3D视频采用的是哪种格式,从而使客户端调整到与接收的 3D视频 格式匹配的状态, 进而对 3D视频进行处理, 例如, 对 3D视频进行处理, 得 到供投影到人的左眼的左视图、 供投影到人的右眼的右视图。 客户端需要获 取的这些信息统称为 3D格式描述信息。
现有的 3D格式描述信息的传输方法中,将描述帧包装格式的 3D格式描述 信息封装到帧包装排列补充增强信息消息 (frame packing arrangement SB message )中, 然后将 frame packing arrangement SB message封装到视频 比特流中传输。 视频比特流即编码后形式的视频。 客户端接收到视频比特流 后, 从视频比特流中获取帧包装排列补充增强信息消息, 然后从中获取描述 帧包装格式的 3D格式描述信息。
由于多媒体系统通常是一个异构系统, 可能既有 2D客户端, 又有 3D客户 端; 都是 3D客户端, 可能有的客户端支持帧包装格式, 有的客户端支持 2D加 辅助视频格式; 都是支持帧包装格式的 3D客户端, 可能有的客户端只支持并 排和上下格式的帧包装, 而不支持棋盘格式的帧包装; 都是支持二维加辅助 视频格式的 3D客户端, 可能有的客户端不支持辅助视频为深度图。 这样就可 能出现一些客户端不支持某个 3D视频采用的 3D格式的情况。而由于帧包装排 列补充增强信息通过视频比特流携带并且在传输时采用轮播方式, 即每隔一 段时间进行一次传输, 这就可能造成多媒体系统的客户端接收视频比特流一 段时间后才能获取帧包装排列补充增强信息, 进而获得所述描述帧包装格式 的 3D格式描述信息, 然后才能确定是否支持接收到的视频采用的 3D格式。
因此,现有的 3D格式描述信息的传输方法,可能导致用户点击播放后, 等待一定时间才能确定客户端不支持 3D视频采用的 3D格式,从而不能正确 处理和显示 3D视频, 影响用户体验。 另一方面增加了接收和处理视频的开 销, 增加了电力消耗, 尤其会增加对电力敏感的移动客户端的负担。
发明内容
本发明实施例提供了一种 3D格式描述信息的获取方法和装置, 以解决现 有技术中客户端接收视频一段时间后才能接收到 3D格式描述信息的缺陷, 缩 短了客户端确定是否支持视频采用的 3D格式的时间。
本发明实施例提供一种 3D格式描述信息的获取方法, 包括:
接收发送端发送的携带 3D格式描述信息的带外消息, 客户端在加入和参 与所述发送端发起的多媒体会话前接收所述带外消息;
解析所述带外消息, 从所述带外消息中获取所述 3D格式描述信息。
本发明实施例还提供一种 3D格式描述信息的获取方法, 包括:
获取 3D视频文件, 所述 3D视频文件的元数据部分携带 3D格式描述信息; 解析所述 3D视频文件的元数据部分,从所述元数据部分中获取所述 3D格 式描述信息。
本发明实施例还提供一种 3D格式描述信息的获取装置, 包括:
接收模块, 用于接收发送端发送的携带 3D格式描述信息的带外消息, 所 述接收模块在客户端加入和参与所述发送端发起的多媒体会话前接收所述带 外消息;
解析模块, 用于解析所述接收模块接收的所述带外消息, 从所述带外消 息中获取所述 3D格式描述信息。
本发明实施例还提供另一种 3D格式描述信息的获取装置, 包括: 获取模块,用于获取 3D视频文件,所述 3D视频文件的元数据部分携带 3D 格式描述信息;
解析模块, 用于解析所述获取块获取的所述 3D视频文件的元数据部分, 从所述元数据部分中获取所述 3 D格式描述信息。
本发明实施例提供的 3D格式描述信息的获取方法和装置, 客户端能够在 获取视频之前获取 3D格式描述信息, 从而使客户端接收到视频前便可以确定 是否支持 3D视频采用的 3D格式, 确定客户端支持 3D视频采用的 3D格式后, 才获取视频。 缩短了客户端确定视频采用的 3D格式的时间, 降低了接收和处 理视频的开销, 减少了电力消耗, 减轻了接收设备的负担。
附图说明
为了更清楚地说明本发明实施例或现有技术中的技术方案, 下面将对实 施例或现有技术描述中所需要使用的附图作一筒单地介绍, 显而易见地, 下 面描述中的附图是本发明的一些实施例, 对于本领域普通技术人员来讲, 在 不付出创造性劳动性的前提下, 还可以根据这些附图获得其他的附图。
图 1为本发明提供的 3D格式描述信息的获取方法第一实施例的流程图; 图 2为本发明提供的 3D格式描述信息的获取方法第二实施例的流程图; 图 3为本发明提供的 3D格式描述信息的获取方法第三实施例的流程图; 图 4为本发明提供的 3D格式描述信息的获取方法第四实施例的流程图; 图 5为本发明提供的 3D格式描述信息的获取方法第五实施例的流程图; 图 6为本发明提供的 3D格式描述信息的获取方法第六实施例的流程图; 图 7为本发明提供的 3D格式描述信息的获取装置第一实施例的结构示 意图;
图 8为本发明提供的 3D格式描述信息的获取装置第二实施例的结构示 意图。
具体实施方式
为使本发明实施例的目的、 技术方案和优点更加清楚, 下面将结合本发 明实施例中的附图, 对本发明实施例中的技术方案进行清楚、 完整地描述, 显然, 所描述的实施例是本发明一部分实施例, 而不是全部的实施例。 基于 本发明中的实施例, 本领域普通技术人员在没有作出创造性劳动前提下所获 得的所有其他实施例, 都属于本发明保护的范围。
图 1为本发明提供的 3 D格式描述信息的获取方法第一实施例的流程图, 如图 1所示, 该方法包括:
S101、接收发送端发送的携带 3D格式描述信息的带外消息,客户端在加 入和参与所述发送端发起的多媒体会话前接收所述带外消息。
S102、解析所述带外消息,从所述带外消息中获取所述 3D格式描述信息。 以上步骤的执行主体为客户端的接收设备。
其中, 携带 3D格式描述信息的带外消息是一种接收设备从发送端发起的 多媒体会话外获取的消息。 本发明实施例中, 客户端在加入和参与发送端发 起的多媒体会话前接收带外消息。 即, 接收设备能够在接收发送端发送的媒 体流之前接收该带外消息。 而在不同的系统或不同的应用场景中, 带外消息 可以为发送设备和接收设备之间传输的各种消息。
具体的, 在多媒体业务流程中, 带外消息可以是会话描述协议(Session Description Protocol; SDP )文件, 由于 SDP文件中通常携带视频获取信息, 发送端需要先将 SDP文件发送至客户端后, 才向客户端发送视频。 因此, 发 送端可以在 SDP文件中携带 3D格式描述信息, 以使客户端加入和参与多媒体 会话前获取 S D P文件中携带的 3 D格式描述信息。 具体可以通过 S D P文件的属 性中包含 3D格式描述信息。
在电视系统中, 由于客户端的接收设备要先获取的电子节目菜单 ( Electronic Program Guide; EPG )元数据, 根据 EPG元数据选择内容后, 才会开始接收媒体流, 因此, 带外消息可以为 EPG元数据, 发送端可以将 3D 格式描述信息携带在 EPG元数据中。 具体可以在 EPG元数据的可扩展标记语 言 ( Extensible Markup Language; XML )元素或属性中包含所述 3D格式描 述信息。
在电视系统中, 由于与节目内容相关的通知消息会略提前于节目内容进 行下发。 因此, 带外消息还可以是通知消息, 发送端通过通知消息携带 3D格 式描述信息, 具体可以在通知消息负载 (payload ) 中包含所述 3D格式描述 信息。
本实施例仅列举了几种系统或业务流程中的带外消息的具体类型, 但并 不以此作为本发明的限制。
3D格式描述信息可以为 3D格式类型标识信息, 用于表示 3D视频采用哪 种格式, 其中, 3D格式类型标识信息又可以进一步包括 3D格式类型标识, 还 可以包括组件类型标识。 另外, 3D格式描述信息还可以包括 3D视频处理参数 信息。
本发明实施例提供的 3D格式描述信息的获取方法, 客户端能够在获取视 频之前通过带外信息获取 3D格式描述信息, 从而使客户端接收到视频前便可 以确定是否支持 3D视频采用的 3D格式, 确定客户端支持 3D视频采用的 3D格 式后,才获取视频。缩短了客户端确定是否支持 3D视频采用的 3D格式的时间, 降低了接收和处理视频的开销, 减少了电力消耗, 减轻了接收设备的负担。
图 2为本发明提供的 3D格式描述信息的获取方法第二实施例的流程图, 如图 2所示, 本实施例适用于多媒体业务流程,在多媒体业务流程中,携带 3D 格式描述信息的带外信息为 SDP文件。发送端可以在 SDP文件中携带 3D格式 描述信息, 客户端在加入和参与发送端发起的媒体会话前接收该 SDP文件。 从而可以使客户端接收到视频之前获知视频采用哪种 3D格式, 进而可以使客 户端接收到视频前便可以确定客户端是否支持视频采用的 3D格式。 该方法包 括:
5201、接收发送端发送的会话描述协议 SDP文件,所述 SDP文件中携带 3D格式描述信息;
5202、 解析所述 SDP文件, 从所述 SDP文件中获取所述 3D格式描述信 以上步骤的执行主体为客户端的接收设备。
发送端在向客户端发送 3D视频时, 首先向客户端发送 SDP文件, 并在 SDP文件中携带 3D格式描述信息。具体方法是通过 SDP文件的属性中携带 3D 格式描述信息。
客户端接收 SDP文件后, 解析 SDP文件, 判断 SDP文件中是否携带 3D格 式描述信息, 具体可以判断 S D P文件中是否包含携带 3 D格式描述信息的属性 , 通过解析所述属性来获取 3D格式描述信息。
本发明提供的实施例中, 3D格式描述信息可以包括 3D格式类型标识信息。 所述 3D格式类型标识信息包括 3D格式类型标识, 所述 3D格式类型标识指示 3D视频采用的格式类型。 此外, 所述 3D格式类型标识信息还可以包括组件类 型标识, 所述组件类型标识指示组成 3D视频的视频组件的类型。
具体的, 若 3D格式类型为帧包装, 则所述组件类型标识指示视频组件的 类型是按并排 ( side by side; SBS ) 、 上下 ( top and bottom; TAB ) 、 行 交织 ( line interleaved; LID 、 歹' J交织 ( column interleaved; CIL ) 、 棋盘 ( Chessboard; CHB ) 帧序列 (frame sequential; SEQ )等帧包装排列方 式排列的视频中的任意一种。 若 3D格式类型为 2D加辅助视频, 则所述组件类 型标识指示视频组件的类型为 2D视频、 深度图、 视差图、 空洞数据、 既包括
2D视频又包括深度图、 既包括 2D视频又包括视差图等中的任意一种; 此外, 所述组件类型标识还可以进一步指示其中的 2D视频携带的是左视图、 右视图 还是中间视图中的任意一种。 若 3D格式类型为联播, 则所述组件类型标识指 示视频组件的类型是携带左视图的视频、 携带右视图的视频中的任意一种。
本实施例给出一种通过 SDP文件中的属性携带 3D格式类型标识信息的 实施方式, 具体是:
发送端可以通过 SDP文件中的属性 3dFormatType来携带 3D格式类型标 识信息, 该属性为媒体级属性。 具体格式为:
a=3dFormatType: <3d format type > [<component type>]
其中, 参数 <30^0177131 1卩6>为30格式类型标识, 可选参数< component
1卩6>为组件类型标识。
<3d format type>的取值包括但不限于 FP、 2DA和 SC等, 分别代表 3D格 式类型为帧包装、 2D加辅助视频、 联播; <3d format type 取值为 FP时, <component type>的取值包括但不限于 SBS、 TAB, LIL、 CIL、 CHB、 SEQ 等,分别代表组成 3D视频的视频的组件类型为并排、上下、行交织、列交织、 棋盘、 帧序列形式的帧包装视频; <3d format type;^ 取值为 2DA时, <component type>的取值包括但不限于 2d、 D、 P、 2dD、 2dP等, 分别代表 组成 3D视频的视频的组件类型为 2D视频、 深度图、 视差图、 空洞数据、 既包 括 2D视频又包括深度图、 既包括 2D视频又包括视差图等; <3d format type> 的取值为 SC时, component type>的取值包括但不限于 L、 R, 分别代表组 成 3 D视频的视频的组件类型为左视图视频、 右视图视频。
组成 3D视频的每个视频组件都可以通过属性 3dFormatType来指示由该 视频组件构成的 3D视频采用的 3D格式类型和该视频组件的类型。
如果组成 3D视频的视频组件没有通过属性 3dFormatType来指示所述 3D 格式类型和所述组件类型。 则可以默认所述 3D格式类型为 2D加辅助视频, 所 述组件类型为 2D视频。
以上仅是通过属性 3dFormatType来携带 3D格式类型标识信息的一种可 行的实施方式, 但并不以此作为对本发明的限制。
本实施例还给出另一种用通过 SDP文件中的属性携带 3D格式类型标识 信息的实施方式, 具体是:
发送端可以通过属性 fmpt来携带 3D格式类型标识信息, 该属性 fmpt为媒 体级属性。 具体格式为:
a=fmtp: <payload type><3d format type >[<component type>] 其中, 参数 <payload type>为携带 3D视频的 RTP负载类型。 参数 <3d format 1卩6>为30格式类型标识; 可选参 ¾<component type>为组件类型标 识。
组成 3D视频的每个视频组件都可以通过属性 fmtp来指示由该视频组件 构成的 3 D视频采用的 3 D格式类型和该视频的组件类型。
如果组成 3D视频的视频组件没有通过属性 fmtp来指示所述 3D格式类型 和所述组件类型。 则可以默认所述 3D格式类型为 2D加辅助视频, 所述组件类 型为 2D视频。
以上仅是通过属性 fmtp来携带 3D格式类型标识信息的一种可行的实施 方式, 但并不以此作为对本发明的限制。
本发明提供的实施例中,所述 3D格式描述信息除了包括 3D格式类型标识 信息以外, 还可以进一步包括 3 D视频处理参数信息。
具体的, 若 3D格式类型为帧包装, 则所述 3D视频处理参数信息包括但不 限于进行帧包装时采样类型的标识信息、 进行帧包装操作时帧放置顺序的标 识信息等参数信息; 若 3D格式类型为 2D加辅助视频且辅助视频为深度图, 则 所述 3D视频处理参数信息包括但不限于深度样本在 2D视频的空间采样网格 中的水平偏移量、 垂直偏移量, 深度的取值范围指示信息, 也即在屏幕之后 的最大距离、在屏幕之前的最大距离等参数信息; 若 3D格式类型为 2D加辅助 视频且辅助视频为视差图, 则所述 3D视频处理参数信息包括但不限于视差样 本在 2D视频的空间采样网格中的水平偏移量、垂直偏移量、代表零视差的值、 用于定义视差值取值范围的缩放比例、 参考观看距离以及参考屏幕宽度等参 数信息。
同样, 所述 3D视频处理信息也可以通过 SDP文件中的属性携带。
本实施例给出一种通过 SDP文件中的属性携带 3D视频处理参数信息的 实施方式, 具体是:
当 3D格式类型为帧包装时, 可以通过属性 FramePackingParameters携 带相应 3D视频处理参数信息。属性 FramePackingParameters为媒体级属性, 具体格式为:
a=FramePackingParameters: <sampling type>=<value>; < content interpretation type >=<vslue>。
其中, 参数 <sampling type>指示进行帧包装时的采样类型; Sampling type的取值包括但不限于 none、 interleaved , quincunx等, 分别代表不进行 采样、 交替采样、 五点梅花采样等。 参数 <content interpretation type>指示 进行帧包装时的帧放置顺序, 取值为 LFirst或 RFirst, 分别代表左视图对应的 视频帧放置在前、 右视图对应的视频帧放置在前。
当 3D格式类型为 2D加辅助视频且辅助视频为深度图时, 可以通过属性
DepthParameters携带相应 3D视频处理参数信息。 属性 DepthParameters为 媒体级属性, 具体格式为:
a=DepthParameters: <position offset h>=<value>; < position offset v>=<value>; <nkfar>=<value>;<nknear>=<value>
其中, 参数 <position offset 11>指示深度样本在 2D视频的空间采样网格中 的水平偏移量; 参数 <position offset v>指示辅助视频样本在 2D视频的空间采 样网格中的垂直偏移量;参数<111^31 和<111<11631 用于指示深度样本的取值范 围, <nkfar;^ 示在屏幕之后的最大距离, <nknear>指示在屏幕之前的最大距 离。
当 3D格式类型为 2D加辅助视频且辅助视频为视差图时, 可以通过属性
ParallaxParameters携 带 相 应 3D视 频 处 理 参数信 息 。 属 性 ParallaxParameters为媒体级属性, 具体格式为:
a=ParallaxParameters: <position offset h>=<value>; < position offset v>=<value>; <parallax zero>=<value>;<parallax scale>=<value>; <dref>=<value>; <wref>=<value>
其中, 参数 <position offset 11>指示视差样本在 2D视频的空间采样网格中 的水平偏移量; 参数 <position offset ν;^ 示视差样本在 2D视频的空间采样网 格中的垂直偏移量; 参数 <parallax zero>、 <parallax scale>, <dref>、 <wref> 分别指示哪个值代表零视差、 用于定义视差值取值范围的缩放比例、 参考观 看距离以及参考屏幕宽度。 以上仅是通过 SDP文件中的属性来携带 3D视频处理参数信息的一种可 行的实施方式, 但并不以此作为对本发明的限制。
需要说明的是, 3D视频可能由多个视频组件组成, 不同视频组件可以通 过不同媒体流携带。 一个 SDP文件中可能描述多个媒体流, 各媒体流携带不 同的视频组件, 例如: SDP文件描述了媒体流 1、 媒体流 2、 媒体流 3、 媒体 流 4。 媒体流 1携带的视频组件和媒体流 2携带的视频组件构成一个 3D视频; 媒体流 3携带的视频组件和媒体流 4携带的视频组件构成另一个 3D视频。 因此 需要告知客户端哪些媒体流构成 3D视频。
本实施例给出一种通过 SDP文件中的属性 group和 mid属性告知客户端 哪些媒体流构成 3D视频的实施方式, 具体是:
通过属性 mid为组成 3D视频的不同媒体流定义媒体流标识。属性 mid为媒 体级属性, 为媒体流的标识并在一个 SDP文件中唯一。 具体格式为:
a=mid: <meidia stream identifer>
通过属性 group将组成 3D视频的不同媒体流归为一组。属性 group为会话 级属性, 用于将几个用属性 mid标识的媒体流归为一组。 具体格式为:
a=group: <semantics> <media stream identifier 1 > <meidia stream identifier 2> ... <media stream identifier n>
<36^1311 3>为 S3D时, 代表归为一组的不同媒体流组成了 3D视频。 一个 SDP文件的具体如下所示, 其中: 标识为 1的媒体流中的视频组件和 标识为 2的媒体流中的视频组件组成 3D视频。 其中, 标识为 1的媒体流中的视 频组件为携带联播格式的 3D视频中的左视图的视频, 标识为 2的媒体流中的 视频组件为携带联播格式的 3D视频中的右视图的视频。 标识为 3的媒体流中 的视频组件和标识为 4的媒体流中的视频组件组成 3D视频。 其中, 标识为 3的 媒体流中的视频组件为 2D加辅助视频格式的 3D视频中的辅助视频且辅助视 频为深度图。 一个 SDP文件具体可如下所示:
ν=0
o=Alice 292742730 29277831 IN IP4 131 .1 63.72.4
s=The technology of 3D -TV
c=IN IP4 131 .164.74.2 t=0 0
a=group:S3D 1 2
m=video 49170 RTP/AVP 99
a=rtpmap:99 H264/90000
a=3dFormatType: SC L
a=mid:1
m=video 49172 RTP/AVP 1 01
a=rtpmap:1 01 H264/90000
a=3dFormatType:SC R
a=mid:2
a=group: S3D 3 4
m=video 49170 RTP/AVP 103
a=rtpmap:1 03 H264/90000
a=3dFormatType: 2DA 2D
a=mid:3
m=video 49172 RTP/AVP 1 05
a=rtpmap:1 05 H264/90000
a=3dFormatType: 2DA D
a=mid:4
m=audio 52890 RTP/AVP 98
a=rtpmap:98 L16/16000/2
如前所述,发送端携带在 SDP文件中的 3D格式描述信息可以包括 3D格式 类型标识信息, 还可以进一步包括 3D视频处理参数信息, 则客户端的接收设 备可以从获取的 SDP文件中获取 3D格式类型标识信息,以确定 3D视频采用的 格式。对应的,客户端还可能从获取的 SDP文件中获取 3D视频处理参数信息, 以对后续接收到的 3D视频进行相应处理。
图 3为本发明提供的 3 D格式描述信息的获取方法第三实施例的流程图, 如图 3所示, 本实施例提供的方法主要适用于基于实时传输协议(Real-time Transport Protocol; RTP ) 的多媒体系统, 客户端的接收设备在接收媒体流 一段时间后, 才会开始从媒体流中获取视频。 因此, 发送端可以将 3D视频处 理参数信息携带在媒体流中, 并在 SDP文件中携带 3D格式类型标识信息以及 3D视频处理参数信息的指示消息,使客户端能够从 SDP文件中获取 3D格式类 型标识信息, 并且从媒体流中获取 3D视频处理参数信息。
该方法包括:
S301、接收发送端发送的会话描述协议 SDP文件, 所述 SDP文件中携带
3D格式描述信息, 所述 3D格式描述信息为 3D格式类型标识信息和 3D视频处 理参数信息的指示信息, 所述指示信息用于标识所述 3D视频处理参数信息在 媒体流中的位置;
5302、 解析 SDP文件, 从所述 SDP文件中获取所述 3D格式类型标识信 息和 3D视频处理参数信息的指示信息。
5303、 接收发送端发送的媒体流, 根据所述指示信息从所述媒体流中获 取所述 3D视频处理参数信息。
其中, 以上步骤的执行主体为客户端的接收设备。
发送端在向客户端发送 3D视频时, 首先向客户端发送 SDP文件, 并在 SDP文件中携带 3D格式类型标识信息和 3D视频处理参数信息的指示信息。
客户端接收 SDP文件后, 解析 SDP文件, 判断 SDP文件中是否携带 3D格 式、 3D格式类型标识信息和 3D视频处理参数信息的指示信息。 如果携带, 获 取 3 D格式类型标识信息和 3 D视频处理参数信息的指示信息。
客户端获取组成 3D视频的媒体流, 根据所述指示信息从媒体流中的相应 位置获取 3D视频处理参数信息。
本实施例中, 通过 SDP文件中的属性携带 3D格式类型标识信息, 客户端 通过解析所述属性获取 3D格式类型标识信息。
在基于 RTP的多媒体系统中, 媒体流以 RTP包为传输单位, RTP包分为 RTP包头 ( header )和 RTP负载 ( payload )两部分。 而 RTP包头分为固定包 头和扩展包头两部分。 因此, 发送端可以通过 RTP包的负载部分, 或者通过 RTP包的扩展包头部分携带 3D视频处理参数信息。
发送端通过负载部分携带 3D视频处理参数的 RTP包,与用于携带相应 3D 视频组件的 RTP包通过同一媒体流传输。在这种情况下,发送端可以通过 SDP 文件中的属性携带 3D视频处理参数信息的指示信息。 所述指示信息指示携带 3D视频处理参数信息的 RTP负载类型号。 本实施例中给出了一种通过 RTP负载携带 3D视频处理参数信息的实施 方式, 具体是:
若 3D格式类型为帧包装,则可以将携带相应 3D视频处理参数信息的消息 封装到 RTP负载中, 携带相应 3D视频处理参数信息的消息具体可以是 sei— rbsp()。
若 3D格式类型为 2D加辅助视频, 则可以将携带相应 3D视频处理参数信 息的消息封装到 RTP负载中。 携带相应 3D视频处理参数信息的消息具体可以 是 si— rbsp()
相应的, 本实施例给出一种通过 RTP包的负载部分携带 3D视频处理参数 信息的情况下, 通过 SDP文件中的属性携带 3D视频处理参数信息的指示信息 的实施方式, 其中, 指示信息具体为 RTP负载类型号。 具体的:
可以通过属性 rtpmap指示携带 3D视频处理参数信息的 RTP负载类型号。 属性 rtpmap为媒体级属性, 用于标识负载类型号代表的负载格式的含义。 具 体格式为:
a=rtpmap: <payload type> <encoding name> I <clock rate> I [/<encoding parameters>]
若参数 <encoding name 取值的语义为 RTP负载携带的是 3D视频处 理参数, 比如取值为 3d Parameters, 相应的参数 <payload type>的取值为带
3D视频处理参数信息的 RTP负载类型号。
客户端首先获取 SDP文件, 可以从 SDP文件中的属性 rtpmap中获取携带 3D视频处理参数信息的 RTP负载类型号。 当客户端获取相应的媒体流时, 便 可以根据获取的负载类型号从媒体流中获取包头中 PT ( Payload Type, 负载 类型)段取值与获取的负载类型号相同的 RTP包, 从所述 RTP包的 RTP包负 载部分获取所述 3D视频处理参数信息。
发送端还可以将 3D视频处理参数信息携带在用于携带相应 3D视频组件 的 RTP包的 RTP扩展包头中, 在这种情况下, 发送端通过 SDP文件中的属性 携带 3D视频处理参数信息的指示信息,该所述指示信息用于指示携带 3D视频 处理参数信息的扩展项的标识。
本实施例给出一种通过 RTP扩展包头携带 3D视频处理参数信息的实施 方式, 具体是: 若 3D格式类型为帧包装,则可以通过一个扩展项携带相应 3D视频处理参 数信息。 具体可以是将携带相应 3D视频处理参数信息的 sei— rbsp()消息封装 到所述扩展项中, 将所述扩展项封装到携带相应 3D视频组件的 RTP包的 RTP 扩展包头中。
若 3D格式类型为 2D加辅助视频, 则可以通过一个扩展项携带相应 3D视 频处理参数信息。 具体可以是将携带相应 3D视频处理参数信息的 si— rbsp()消 息封装到所述扩展项中, 将所述扩展项封装到携带相应 3D视频组件的 RTP包 的 RTP扩展包头中。
本实施例还给出一种通过 RTP扩展包头携带 3D视频处理参数信息的实 施方式, 具体是:
若 3D格式类型为帧包装,则可以通过不同扩展项分别携带相应 3D视频处 理参数信息中的进行帧包装时采样类型的标识信息、 进行帧包装操作时帧放 置顺序的标识信息等参数信息; 将所述扩展项封装到携带相应 3D视频组件的 RTP包的 RTP扩展包头中。
若 3D格式类型为 2D加辅助视频且辅助视频为深度图,则可以通过不同扩 展项分别携带相应 3D视频处理参数信息中的深度样本在 2D视频的空间采样 网格中的水平偏移量、 垂直偏移量、 深度样本的取值范围指示信息中的在屏 幕之后的最大距离、 在屏幕之前的最大距离等参数信息; 将所述扩展项封装 到携带相应 3 D视频组件的 RTP包的 RTP扩展包头中。
若 3D格式类型为 2D加辅助视频且辅助视频为视差图,则可以通过不同扩 展项分别携带相应 3D视频处理参数信息中的视差样本在 2D视频的空间采样 网格中的水平偏移量、 垂直偏移量、 代表零视差的值、 用于定义视差值取值 范围的缩放比例、 参考观看距离以及参考屏幕宽度等参数信息。 将所述扩展 项封装到携带相应 3 D视频组件的 RTP包的 RTP扩展包头中。
本实施例中, 可以只将携带 3D视频处理参数信息的扩展项封装到携带相 应 3D视频组件的关键帧的 RTP包的 RTP扩展包头中。
相应地, 本实施例中给出一种通过 RTP扩展包头携带 3D视频处理参数信 息的情况下, 通过 SDP文件中的属性携带 3D视频处理参数信息的指示信息的 实施方式。 具体是:
通过属性 extmap指示携带 3D视频处理参数信息的扩展项的标识。 属性 extmap可以是媒体级属性, 也可以是会话级属性。 用于标识扩展项的标识和 含义的对应关系。 具体格式为:
a=extmap:<value>["/"<direction>] <URI> <extensionattributes> 若参数<1^1>取值的语义为扩展项携带的是 3D视频处理参数信息, 例如 取值为 urn:example:params:3dParameters,参数 <value 取值为携带 3D视 频处理参数信息的扩展项的标识。
客户端首先获取 SDP文件, 可以从 SDP文件中的属性 extmap中获取携带
3D视频处理参数信息的扩展项的标识。 当客户端获取到相应的媒体流后, 首 先获取包含扩展包头的 RTP包, 具体可以是获取 RTP包头的 X ( extension , 扩展)段为 1的 RTP包。 从所述 RTP包中获取扩展包头, 解析所述扩展包头, 从中获取扩展项标识等于所述获取的携带 3D视频处理参数信息的扩展项的 标识的扩展项, 解析所述扩展项, 从所述扩展项中获取所述 3D视频处理参数 信息。
本实施例中作为另一种较佳的实施方式, 发送端还可以将 3D视频处理 参数信息携带在视频比特流中。 具体可以是:
若 3D格式类型为帧包装, 通过视频比特流中的帧包装补充增强信息消息 携带 3D视频处理参数信息。
若 3D格式类型为 2D加辅助视频,通过视频比特流中的 si— rbsp()消息携带 3D视频处理参数信息。
在这种情况下, 发送端可以通过 SDP文件中的属性携带 3D视频处理参数 信息的指示信息, 所述指示信息指示携带 3D视频处理参数信息的视频比特流 消息的类型。
本实施例给出一种通过 SDP文件中的属性携带 3D视频处理参数信息的 指示信息的实施方式。 具体是:
若 3D 格 式 类 型 为 帧 包 装 , 通 过 媒 体 级 属 性
FramePackingArrangementS曰 PresentFlag指示携带 3D视频处理参数信息 的视频比特流消息的类型为帧包装补充增强信息消息。 具体格式为:
a=FramePackingArrangementSEI PresentFlag: <value>
其中 <value> 1表示视频比特流中包含携带相应 3D视频处理参数信息 的帧包装补充增强信息消息。 为 0则表示不包含。 若 3D格式类型为 2D加辅助视频, 通过媒体级属性 SiRbspPresentFlag指 示携带 3D视频处理参数信息的视频比特流消息的类型为 si— rbsp消息。具体格 式为:
a=SiRbspPresentFlag: <value>
其中, <value>为1表示视频比特流中包含携带相应 3D视频处理参数信息 的 si— rbsp消息。 为 0则表示不包含。
图 4为本发明提供的 3D格式描述信息的获取方法第四实施例的流程图, ^口图 4所示,本实施 1"列适用于电视系统, ( Electronic Program Guide; EPG ) 元数据是用于生成电子节目菜单的元数据, 用户或者接收设备可以通过 E PG 元数据浏览和选择节目, 然后加入和参与节目相应的多媒体会话, 获取包括 视频在内的节目内容因此,发送端可以在 EPG元数据中携带 3D格式描述信息, 从而使客户端的接收设备能够在获取视频之前获取 3D格式描述信息, 从而更 快速确定接收设备与 3D视频采用的格式是否匹配。
该方法包括:
S401、接收发送端发送的电子节目菜单 EPG元数据, 所述 EPG元数据中 携带 3D格式描述信息;
S402、 解析所述 EPG元数据, 从所述 EPG元数据中获取所述 3D格式描 述信息。
其中, EPG元数据是用于生成电子节目菜单的元数据, 用户或者接收设 备可以通过电子节目菜单浏览和选择节目。 然后加入和参与节目相应的多媒 体会话, 获取节目内容。 EPG元数据中包括描述频道信息的元数据、 描述点 播节目信息的元数据以及描述直播节目信息的元数据。 发送端可以在描述频 道信息、 点播节目信息、 直播节目信息的 EPG元数据中分别携带 3D格式描述 信息, 分别为频道内容、 直播节目内容、 点播节目内容中的 3D视频提供 3D格 式描述信息。
EPG元数据可以为可扩展标记语言( Extensible Markup Language; XML ) 形式, 可以通过扩展 EPG元数据, 增加 XML元素或属性, 通过新增的 XML元 素或属性来携带 3D格式描述信息。
本实施例中, 所述 3D格式描述信息包括 3D格式类型标识信息。
本实施例给出一种通过扩展 EPG元数据,增加 XML元素或属性来携带 3D 格式类型标识信息的实施方式, 具体是:
增加 XML元素或属性指示 3D格式类型。
若 3D格式类型为帧包装, 则还可以进一步增加 XML元素或属性 FramePackingType指示构成 3D视频的帧包装视频组件采用的帧包装类型。
若 3D格式类型为 2D加辅助视频, 则还可以进一步增加 XML元素或属性 AuxVideoType指示构成 3D视频的辅助视频组件采用的辅助视频类型。
若 3D格式类型为联播, 则还可以进一步增加 XML元素或属性 StereolD来 指示构成 3 D视频的 2 D视频组件的视图标识。
其中, XML元素或属性的具体定义如下表所示:
Figure imgf000018_0001
本实施例中, 所述 3D格式描述信息还可以包括 3D视频处理参数信息。 本实例给出一种通过扩展 EPG元数据,增加 XML元素或属性来携带 3D视 频处理参数信息的实施方式, 具体是:
若 3D格式类型为帧包装, 增加 XML元素 FramePackingParameters携带 相应的 3D视频处理参数信息。 FramePackingParameters可以包括 XML元素 或属性 SamplingType、 Contentl nterpretationType , 分别指示进行帧包装时 的采样类型、 进行帧包装操作时帧放置顺序;
若 3D格式类型为 2D加辅助视频且辅助视频类型为深度图,增加 XML元素 DepthParameters携带相应的 3D视频处理参数信息。 Depth Parameters可以 包括 XML元素或属性 posit ion— offset— h、 position— offset— v、 nkfar、 nknear, 分别指示深度样本在 2D视频的空间采样网格中的水平偏移量、 垂直偏移量, 在屏幕之后的最大距离、 在屏幕之前的最大距离;
若 3D格式类型为 2D加辅助视频且辅助视频为视差图, 增加 XML元素 ParallaxParameters携带相应的 3D视频处理参数信息。 ParallaxParameters 可以包括 XML元素或属性 position— offset— h、 position— offset— v、 parallax— zero、 parallax— scale、 dref、 wref, 分别指示视差样本在 2D视频的空间采样网格中 的水平偏移量、 垂直偏移量、 代表零视差的值、 用于定义视差值取值范围的 缩放比例、 参考观看距离以及参考屏幕宽度。
其中, 所述 XML元素或属性的具体定义如下表所示:
Figure imgf000019_0001
dref 参考观看距离
wref 参考屏幕宽度
客户端的接收设备获取 EPG元数据后,可以从 EPG元数据中获取 3D格式 描述信息。 如果 EPG元数据中包括用于携带 3D格式描述信息的 XML元素, 则 可以解析 XM L元素来获取 3 D格式描述信息。
图 5为本发明提供的 3D格式描述信息的获取方法第五实施例的流程图, 如图 5所示, 本实施例适用于电视系统, 发送端向客户端发送 3D视频前, 首 先向客户端发送节目内容相关的通知消息, 因此, 发送端可以在通知消息中 携带 3D格式描述信息, 从而使客户端的接收设备能够快速的确定接收设备与 3D视频采用的格式是否匹配。 该方法包括:
S501、 接收发送端发送的与节目内容相关的通知消息, 所述通知消息中 携带 3D格式描述信息;
S502、解析所述通知消息,从所述通知消息中获取所述 3 D格式描述信息。 发送端向客户端发送 3D视频前, 首先向客户端发送节目内容相关的通知 消息。 所述通知消息的负载(payload ) 中携带 3D格式描述信息。 所述 3D格 式描述信息包括 3D格式类型标识信息,还可以进一步包括 3D视频处理参数信 息。 所述通知消息的负载可以是 XML元素, 可以通过增加 XML元素或属性来 携带 3D格式描述信息。 客户端的接收设备接收到通知消息后, 可以从通知消 息的负载中解析 XML元素, 获取 3D格式描述信息。
需要说明的是, 本实施例中涉及的通知消息, 可以是发送端向客户端发 送 3D视频前, 向终端发送负载中携带可读文本的通知消息, 接收设备接收所 述通知消息并展现给用户。 所述可读文本可以用于提示用户戴上 3D眼镜或者 提示用户观看 3D节目的最佳观看距离。
图 6为本发明提供的 3D格式描述信息的获取方法第六实施例的流程图, 如图 6所示, 客户端可以从光盘、 移动硬盘等存储介质中获取 3D视频文件, 或者从发送端出接收 3D视频文件,可以在这些 3D视频文件的元数据部分携带 3D格式描述信息。 客户端获取 3D视频文件后, 从 3D视频文件的元数据部分 中获取 3D格式描述信息。
因此, 本实施例提供的方法包括:
S601、 获取 3D视频文件, 所述 3D视频文件的元数据部分, 携带 3D格式 描述信息; S602、解析所述 3D视频文件的元数据部分,从所述元数据部分获取所述 3D格式描述信息。
客户端获取 3 D视频文件后, 从 3 D视频文件的元数据部分中获取携带 3 D 格式描述信息的元数据项, 解析所述元数据项, 获取 3D格式描述信息。
本实施例中给出一种通过元数据项携带 3D格式描述信息的实施方式。 具 体为:
通过不同元数据项携带不同 3D格式类型相应的 3D格式描述信息。通过元 数据项的类型或名称指示 3D格式类型。 其他 3D格式类型相应的 3D格式描述 信息可以作为元数据项的内容携带。 所述元数据项的类型或名称、 元数据项 的内容可以通过不同的 Box (箱)携带。可以通过 Item Info Box (项目信息箱) 携带元数据项的类型或名称; 通过 Item Data Box (项目数据箱 )携带元数据 项的内容。 然后将所述 Box封装到 Metadata Box (元数据箱) 中, 随后将 Metadata Box封装到 3D视频轨道箱 (track box ) 中, 然后将 3D视频轨道箱 封装到 Movie Box (影片箱) 中; 最后将 Movie Box封装到文件中。
具体的, 如果 3D格式类型为帧包装, 则可将相应的 3D格式描述信息作为 类型为 fpdt的元数据项封装到帧包装视频相应的轨道箱 (track box ) 中; 然 后将所述轨道箱封装到 Movie Box中; 最后将 Movie Box封装到文件中。
其中, 将帧包装格式类型相应的 3D格式描述信息作为类型为 fpdt的元数 据项封装到帧包装视频相应的轨道箱 (track box ) 中的一种可行的实施方式 为:
在 Item Info Box中标识元数据项的类型为 fpdt,将携带帧包装格式相应的 3D格式描述信息的 S曰消息封装放到 Item Data Box中; 再将 Item Info Box和 Item Data Box封装到 Metadata Box中;将 Metadata Box封装到 Track Box中。
如果 3D格式类型为 2D加辅助视频, 则可以将相应的 3D格式描述信息作 为类型为 sirp的元数据项封装到辅助视频相应的轨道箱中;然后将辅助视频相 应的轨道箱封装到 Movie Box中; 最后将 Movie Box封装到文件中。
其中, 将 2D加辅助视频格式相应的 3D格式描述信息作为类型为 sirp的元 数据项封装到辅助视频相应的轨道箱中的一种可行的实施方式为:
在 Item Info Box中标识元数据项的类型为 sirp; 将 si— rbsp()封装放到 Item Data Box中; 将 Item Info Box和 Item Data Box封装到 Metadata Box中; 将 Metadata Box封装到 Track Box中。
如果 3D格式类型为联播, 则可以将相应的 3D格式描述信息作为类型为 stvw的元数据项分别封装到携带左视图的视频相应的轨道箱和携带右视图的 视频相应的轨道箱中; 然后将上述轨道箱封装到 Movie Box中; 最后将 Movie Box封装到文件中。
其中, 将联播格式类型相应的 3D格式描述信息作为类型为 stvw的元数据 项封装到携带左视图的视频或携带右视图的视频相应的轨道箱中的一种可行 的实施方式为:
在 Item Info Box中标识元数据项的类型为 stvw;将 stem— view—lnfo()封装 放到 Item Data Box中;将 Item Info Box和 Item Data Box封装到 Metadata Box 中; 将 Metadata Box封装到 Track Box中。
其中, stereo— view— info()结构的定义
Figure imgf000022_0001
其中, stereo— id用于指示携带的视图是左还是右。 Reference— track— id 指示携带另一视图的视频轨道的标识。 述信息。客户端获取 3D视频文件后,从 3D视频文件的元数据部分中获取携带 3D格式描述信息的箱, 解析所述箱, 获取 3D格式描述信息。
本实施例还提供了一种通过 3D视频文件的元数据部分的箱携带 3D格式 描述信息的具体实施方式, 具体包括:
通过不同类型的 box携带不同 3D格式类型相应的 3D格式描述信息, 通过
Box的类型指示 3D格式的类型。 : 类型为 spif的 box用于携带 2d加辅助视频格式类型相应的 3D格式描述信 类型为 stif的 box用于携带联播格式类型相应的 3D格式描述信息。
所述 box的定义如下所示:
class FramePackingDataBox extends Box('fpdt'){ unsigned int(8) frame— packing— arrangement— type;
unsigned int(8) sampling— type;
unsigned int(8) content— interpretation— type;
}
class SupplementallnfoBox extends Box('spif ){
unsigned int(8) aux— video— type;
unsigned int(8) position— offset— h;
unsigned int(8) position— offset— v;
if(aux— video— type=0){
unsigned int(8) nkar;
unsigned int(8) nknear;
}
else if(aux— video— type=1 ){
unsigned int(1 6) parallax— zero;
unsigned int(1 6) parallax— scale;
unsigned int(1 6) dref;
unsigned int(1 6) wref;
}
}
class StereoViewl nfoBox extends Box('stif'){
unsigned int(8) stereo— id;
unsigned int(8) reference— track— id;
}
将所述 Box封装封装到 Sample Description Box (样本描述箱) 中, 将所 述样本描述箱封装到相应的轨道箱中,将所述轨道箱封装到 Movie Box中; 最 后将 Movie Box封装到文件中。
本发明实施例提供的 3D格式描述信息的获取方法, 客户端能够在获取视 频之前通过 3D视频文件中的元数据获取 3D格式描述信息,从而使客户端获取 到视频前便可以确定是否支持 3D视频采用的 3D格式, 确定客户端支持 3D视 频采用的 3D格式后, 才获取视频。 缩短了客户端确定是否支持 3D视频采用的 3D格式的时间, 降低了接收和处理视频的开销, 减少了电力消耗, 减轻了接 收设备的负担。
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程, 是可以通过计算机程序来指令相关的硬件来完成, 所述的程序可存储于一计 算机可读取存储介质中, 该程序在执行时, 可包括如上述各方法的实施例的 流程。 其中, 所述的存储介质可为磁碟、 光盘、 只读存储记忆体( Read-Only Memory, ROM )或随机存储记忆体 ( Random Access Memory, RAM )等。
图 7为本发明提供的 3D格式描述信息的获取装置第一实施例的结构示意 图, 如图 7所示, 该装置包括: 接收模块 1 1和解析模块 12;
接收模块 1 1 , 用于接收携带 3D格式描述信息的带外消息, 所述带外消息 在加入和参与多媒体会话前获取;
解析模块 12, 用于解析所述接收模块 1 1获取的所述带外消息, 从所述带 外消息中获取所述 3D格式描述信息。
其中,接收模块 1 1获取的带外消息中携带的 3 D格式描述信息包括 3 D格式 类型标识信息; 或者, 所述 3D格式描述信息包括 3D格式类型标识信息和 3D 视频处理参数信息。
所述 3D格式类型标识信息可以包括 3D格式类型标识; 或者, 所述 3D格 式类型标识信息还可以包括 3D格式类型标识和组件类型标识。
作为一个较佳的实施例, 所述接收模块 1 1获取的所述带外消息为 SDP文 件。
所述接收模块 1 1接收的所述 SDP文件中携带 3D格式类型标识信息以及 3D视频处理参数信息的指示信息,所述指示信息用于指示所述 3D视频处理参 数信 , ¾在媒体流中的位置。
进一步的, 所述接收模块 1 1还可以用于: 接收发送端发送的媒体流, 根 据所述指示信息从所述媒体流中获取所述 3D视频处理参数信息。
其中, 所述接收模块 1 1接收的所述 SDP文件中携带的所述指示信息为 RTP负载类型号;
相应的, 所述接收模块 1 1还用于: 根据所述 RTP负载类型号从所述媒体 流的对应 RTP负载中获取所述 3D视频处理参数信息。
或者, 所述接收模块 1 1接收的所述 SDP文件中携带的所述指示信息还可 以为 RTP包头的扩展项的标识;
则相应的, 所述接收模块 1 1还用于: 根据所述 RTP包头的扩展项的标识 从所述媒体流的对应 RTP包头中获取所述 3D视频处理参数信息。
作为另一个较佳的实施例, 所述接收模块 1 1接收的所述带外消息还可以 为 EPG元数据。
作为又一个较佳的实施例, 所述接收模块 1 1接收的所述带外消息也可以 为电视系统中的通知消息。
本实施例提供的 3D格式描述信息的获取装置,与本发明提供的 3D格式描 述信息的获取方法实施例〜第五实施例相对应, 为实现该方法的功能设备, 其 具体实现方式可参见前面的方法实施例〜第五实施例, 在此不再赘述。
本发明实施例提供的 3D格式描述信息的获取装置, 客户端能够在获取视 频之前从带外消息中携带 3D格式描述信息, 从而使客户端接收到视频前便可 以确定是否支持 3D视频采用的 3D格式, 确定客户端支持 3D视频采用的 3D格 式后,才获取视频。缩短了客户端确定是否支持 3D视频采用的 3D格式的时间, 降低了接收和处理视频的开销, 减少了电力消耗, 减轻了接收设备的负担。
图 8为本发明提供的 3D格式描述信息的获取装置第二实施例的结构示意 图, 如图 8所示, 该装置包括: 获取模块 21和解析模块 22;
获取模块 21 ,用于获取 3D视频文件,所述 3D视频文件的元数据部分携带 3D格式描述信息;
解析模块 22,用于解析所述获取块 21获取的所述 3D视频文件的元数据部 分, 从所述元数据部分中获取所述 3D格式描述信息。
其中,所述获取模块 21获取的 3D视频文件的元数据部分中携带的所述 3D 格式描述信息包括 3D格式类型标识信息;或者,所述 3D格式描述信息包括 3D 格式类型标识信息和 3 D视频处理参数信息。
进一步的, 所述 3D格式类型标识信息包括 3D格式类型标识; 或者, 所 述 3D格式类型标识信息包括 3D格式类型标识和组件类型标识。
本实施例提供的 3D格式描述信息的获取装置,与本发明提供的 3D格式 描述信息的获取方法第六实施例相对应, 为实现该方法的功能设备, 其具体 实现方式可参见前面的方法第六实施例, 在此不再赘述。
本发明实施例提供的 3D格式描述信息的获取装置, 客户端能够在获取 视频之前通过 3D视频文件中的元数据获取 3D格式描述信息,从而使客户端 接收到视频前便可以确定是否支持 3D视频采用的 3D格式,确定客户端支持 3D视频采用的 3D格式后, 才获取视频。 缩短了客户端确定是否支持 3D视 频采用的 3D格式的时间,降低了接收和处理视频的开销,减少了电力消耗, 减轻了接收设备的负担。
最后应说明的是: 以上实施例仅用以说明本发明的技术方案, 而非对其 限制; 尽管参照前述实施例对本发明进行了详细的说明, 本领域的普通技术 人员应当理解: 其依然可以对前述各实施例所记载的技术方案进行修改, 或 者对其中部分技术特征进行等同替换; 而这些修改或者替换, 并不使相应技 术方案的本质脱离本发明各实施例技术方案的精神和范围。

Claims

权 利 要 求
1、 一种 3D格式描述信息的获取方法, 其特征在于, 包括:
接收发送端发送的携带 3D格式描述信息的带外消息, 客户端在加入和参 与所述发送端发起的多媒体会话前接收所述带外消息;
解析所述带外消息, 从所述带外消息中获取所述 3D格式描述信息。
2、 根据权利要求 1所述的方法, 其特征在于, 所述 3D格式描述信息包括 3D格式类型标识信息; 或者, 所述 3D格式描述信息包括 3D格式类型标识信 息和 3 D视频处理参数信息。
3、 根据权利要求 2所述的方法, 其特征在于, 所述 3D格式类型标识信息 包括 3D格式类型标识; 或者, 所述 3D格式类型标识信息包括 3D格式类型标 识和组件类型标识。
4、 根据权利要求 1 -3任一项所述的方法, 其特征在于, 所述带外消息为 SDP文件。
5、 根据权利要求 4所述的方法, 其特征在于, 所述 SDP文件中携带 3D格 式类型标识信息以及 3D视频处理参数信息的指示信息, 所述指示信息用于指 示所述 3D视频处理参数信息在媒体流中的位置。
6、 根据权利要求 5所述的方法, 其特征在于, 还包括:
接收所述发送端发送的媒体流, 根据所述指示信息从所述媒体流中获取 所述 3D视频处理参数信息。
7、 根据权利要求 6所述的方法, 其特征在于, 所述指示信息为实时传输 协议 RTP负载类型号, 则根据所述指示信息从所述媒体流中获取所述 3D视频 处理参数信息具体为:
根据所述 RTP负载类型号从所述媒体流的对应 RTP负载中获取所述 3 D 视频处理参数信息。
8、 根据权利要求 6所述的方法, 其特征在于, 所述指示信息为 RTP包头 的扩展项的标识, 则根据所述指示信息从所述媒体流中获取所述 3 D视频处理 参数信息具体为:
根据所述 RTP包头的扩展项的标识从所述媒体流的对应 RTP包头中获取 所述 3D视频处理参数信息。
9、 根据权利要求 1 -3任一项所述的方法, 其特征在于, 所述带外消息为 电子节目菜单 EPG元数据。
10、 根据权利要求 1 -3任一项所述的方法, 其特征在于, 所述带外消息为 电视系统中的通知消息。
1 1、 一种 3D格式描述信息的获取方法, 其特征在于, 包括:
获取 3D视频文件, 所述 3D视频文件的元数据部分携带 3D格式描述信息; 解析所述 3D视频文件的元数据部分,从所述元数据部分中获取所述 3D格 式描述信息。
12、 根据权利要求 1 1所述的方法, 其特征在于, 所述 3D格式描述信息包 括 3D格式类型标识信息; 或者, 所述 3D格式描述信息包括 3D格式类型标识 信息和 3D视频处理参数信息。
13、 根据权利要求 12所述的方法, 其特征在于, 所述 3D格式类型标识信 息包括 3D格式类型标识; 或者, 所述 3D格式类型标识信息包括 3D格式类型 标识和组件类型标识。
14、 一种 3D格式描述信息的获取装置, 其特征在于, 包括:
接收模块, 用于接收发送端发送的携带 3D格式描述信息的带外消息, 所 述接收模块在客户端加入和参与所述发送端发起的多媒体会话前接收所述带 外消息;
解析模块, 用于解析所述接收模块获取的所述带外消息, 从所述带外消 息中获取所述 3D格式描述信息。
15、 根据权利要求 14所述的 3D格式描述信息的获取装置, 其特征在于, 所述接收模块接收的带外消息中携带的 3D格式描述信息包括 3D格式类型标 识信息; 或者, 所述 3D格式描述信息包括 3D格式类型标识信息和 3D视频处 理参数信息。
16、 根据权利要求 15所述的 3D格式描述信息的获取装置, 其特征在于, 所述 3D格式类型标识信息包括 3D格式类型标识; 或者, 所述 3D格式类型标 识信息包括 3 D格式类型标识和组件类型标识。
17、 根据权利要求 14-16任一项所述的 3D格式描述信息的获取装置, 其 特征在于, 所述接收模块获取的所述带外消息为 S D P文件。
18、 根据权利要求 17所述的 3D格式描述信息的获取装置, 其特征在于, 所述接收模块获取的所述 SDP文件中携带 3D格式类型标识信息以及 3D视频 处理参数信息的指示信息, 所述指示信息用于指示所述 3D视频处理参数信息 在媒体流中的位置。
19、 根据权利要求 18所述的 3D格式描述信息的获取装置, 其特征在于, 所述接收模块还用于: 接收所述发送端发送的媒体流, 根据所述指示信息从 所述媒体流中获取所述 3D视频处理参数信息。
20、 根据权利要求 19所述的 3D格式描述信息的获取装置, 其特征在于, 所述接收模块接收的所述 SDP文件中携带的所述指示信息为实时传输协议 RTP负载类型号;
所述接收模块还用于: 根据所述 RTP负载类型号从所述媒体流的对应 RTP负载中获取所述 3D视频处理参数信息。
21、 根据权利要求 19所述的 3D格式描述信息的获取装置, 其特征在于, 所述接收模块接收的所述 S D P文件中携带的所述指示信息为 RTP包头的扩展 项的标识;
所述接收模块还用于: 根据所述 RTP包头的扩展项的标识从所述媒体流 的对应 RTP包头中获取所述 3D视频处理参数信息。
22、 根据权利要求 14-16任一项所述的 3D格式描述信息的获取装置, 其 特征在于, 所述接收模块接收的所述带外消息为电子节目菜单 EPG元数据。
23、 根据权利要求 14-16任一项所述的 3D格式描述信息的获取装置, 其 特征在于, 所述接收模块接收的所述带外消息为电视系统中的通知消息。
24、 一种 3D格式描述信息的获取装置, 其特征在于, 包括:
获取模块,用于获取 3D视频文件,所述 3D视频文件的元数据部分携带 3D 格式描述信息;
解析模块, 用于解析所述获取块获取的所述 3D视频文件的元数据部分, 从所述元数据部分中获取所述 3 D格式描述信息。
25、 根据权利要求 24所述的 3D格式描述信息的获取装置, 其特征在于, 所述获取模块获取的 3D视频文件的元数据部分中携带的所述 3D格式描述信 息包括 3D格式类型标识信息; 或者, 所述 3D格式描述信息包括 3D格式类型 标识信息和 3D视频处理参数信息。
26、 根据权利要求 25所述的 3D格式描述信息的获取装置, 其特征在于, 所述 3D格式类型标识信息包括 3D格式类型标识; 或者, 所述 3D格式类型标 识信息包括 3 D格式类型标识和组件类型标识。
PCT/CN2012/071767 2011-03-02 2012-02-29 3d格式描述信息的获取方法和装置 WO2012116638A2 (zh)

Priority Applications (2)

Application Number Priority Date Filing Date Title
EP12752211.8A EP2675181A4 (en) 2011-03-02 2012-02-29 METHOD AND DEVICE FOR ACQUIRING 3D FORMAT DESCRIPTION INFORMATION
US14/014,767 US20140002593A1 (en) 2011-03-02 2013-08-30 Method and apparatus for acquiring 3d format description information

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201110050253.6A CN102137298B (zh) 2011-03-02 2011-03-02 3d格式描述信息的获取方法和装置
CN201110050253.6 2011-03-02

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US14/014,767 Continuation US20140002593A1 (en) 2011-03-02 2013-08-30 Method and apparatus for acquiring 3d format description information

Publications (2)

Publication Number Publication Date
WO2012116638A2 true WO2012116638A2 (zh) 2012-09-07
WO2012116638A3 WO2012116638A3 (zh) 2012-11-01

Family

ID=44296937

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2012/071767 WO2012116638A2 (zh) 2011-03-02 2012-02-29 3d格式描述信息的获取方法和装置

Country Status (4)

Country Link
US (1) US20140002593A1 (zh)
EP (1) EP2675181A4 (zh)
CN (2) CN102137298B (zh)
WO (1) WO2012116638A2 (zh)

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5870510B2 (ja) * 2010-09-14 2016-03-01 株式会社リコー ステレオカメラ装置、校正方法およびプログラム
CN102137298B (zh) * 2011-03-02 2015-12-09 华为技术有限公司 3d格式描述信息的获取方法和装置
KR20140021231A (ko) * 2012-08-09 2014-02-20 한국전자통신연구원 증강방송 메타데이터 전송 장치, 사용자 단말, 증강방송 메타데이터 전송 방법과 재생 방법
TWI502545B (zh) * 2013-06-25 2015-10-01 儲存3d影像內容的方法
JP6599864B2 (ja) 2014-04-27 2019-11-06 エルジー エレクトロニクス インコーポレイティド 放送信号送信装置、放送信号受信装置、放送信号送信方法、及び放送信号受信方法
WO2016204481A1 (ko) * 2015-06-16 2016-12-22 엘지전자 주식회사 미디어 데이터 전송 장치, 미디어 데이터 수신 장치, 미디어 데이터 전송 방법, 및 미디어 데이터 수신 방법
KR102468763B1 (ko) * 2016-02-05 2022-11-18 삼성전자 주식회사 영상처리장치 및 그 제어방법
US10528739B2 (en) * 2016-04-20 2020-01-07 Sophos Limited Boot security
CN106408641B (zh) * 2016-09-19 2019-10-18 上海玮舟微电子科技有限公司 一种图像数据的缓存方法和装置
US11917269B2 (en) * 2022-01-11 2024-02-27 Tencent America LLC Multidimensional metadata for parallel processing of segmented media data

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
MY145716A (en) * 2005-04-08 2012-03-30 Qualcomm Inc Method and apparatus for enhanced file distribution in multicast or broadcast
CN101098470A (zh) * 2006-06-28 2008-01-02 腾讯科技(深圳)有限公司 一种基于p2p技术的流媒体节目播放/接收方法及装置
JP4347322B2 (ja) * 2006-07-14 2009-10-21 ソニー株式会社 受信装置および方法、並びにプログラム
KR100939641B1 (ko) * 2008-02-18 2010-02-03 전자부품연구원 입체형 비디오 파일 포맷 및 그에 따른 입체형 비디오파일을 기록한 컴퓨터로 판독 가능한 기록매체
KR20100002032A (ko) * 2008-06-24 2010-01-06 삼성전자주식회사 영상 생성 방법, 영상 처리 방법, 및 그 장치
CN101616334A (zh) * 2008-08-21 2009-12-30 青岛海信电器股份有限公司 视频信号的显示方法和装置
KR101500440B1 (ko) * 2008-10-17 2015-03-09 삼성전자 주식회사 영상처리장치 및 영상처리방법
US8606076B2 (en) * 2008-11-24 2013-12-10 Koninklijke Philips N.V. 3D video reproduction matching the output format to the 3D processing ability of a display
BRPI0924138A2 (pt) * 2009-01-26 2016-02-10 Thomson Licensing método, aparelho e sistema para melhorar a sintonia em receptores
CA2760100C (en) * 2009-04-27 2015-12-08 Lg Electronics Inc. Broadcast transmitter, broadcast receiver and 3d video data processing method thereof
KR20100138806A (ko) * 2009-06-23 2010-12-31 삼성전자주식회사 자동 3차원 영상 포맷 변환 방법 및 그 장치
CN101964920B (zh) * 2010-10-28 2012-03-14 四川长虹电器股份有限公司 在3d电视上兼容多种3d播放模式的方法
CN102137298B (zh) * 2011-03-02 2015-12-09 华为技术有限公司 3d格式描述信息的获取方法和装置

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
None
See also references of EP2675181A4

Also Published As

Publication number Publication date
US20140002593A1 (en) 2014-01-02
EP2675181A2 (en) 2013-12-18
CN102137298A (zh) 2011-07-27
CN102137298B (zh) 2015-12-09
EP2675181A4 (en) 2014-07-16
CN105554551A (zh) 2016-05-04
WO2012116638A3 (zh) 2012-11-01

Similar Documents

Publication Publication Date Title
WO2012116638A2 (zh) 3d格式描述信息的获取方法和装置
US10194134B2 (en) Signaling three-dimensional video information in communication networks
US9351028B2 (en) Wireless 3D streaming server
EP2498495B1 (en) Decoder and method at the decoder for synchronizing the rendering of contents received through different networks
EP2706755A2 (en) Media transmission method, media reception method, client and system thereof
US10869099B2 (en) Transmission device, transmission method, reception device, and reception method
US20120033035A1 (en) Method and system for transmitting/receiving 3-dimensional broadcasting service
EP2178306A2 (en) Image processing apparatus and image processing method
KR20130014428A (ko) 실시간으로 전송되는 기준 영상과 별도로 전송되는 부가 영상 및 콘텐츠를 연동하여 3d 서비스를 제공하기 위한 전송 장치 및 방법, 및 수신 장치 및 방법
KR20130120422A (ko) 멀티미디어 전송 시스템을 위한 데이터 송수신 방법 및 장치
US10116997B2 (en) Method and apparatus for transmitting/receiving content in a broadcast system
US9674499B2 (en) Compatible three-dimensional video communications
EP2685729A2 (en) Transmission device and method for providing program-linked stereoscopic broadcasting service, and reception device and method for same
JP5841715B2 (ja) 映像音声出力装置、および映像音声出力システム、およびマスタ装置
US9204123B2 (en) Video content generation
CN107787585B (zh) 用于处理立体视频数据的mmt装置和mmt方法
CA2824708C (en) Video content generation
KR20140004045A (ko) 실시간으로 전송되는 기준 영상과 별도로 전송되는 부가 영상 및 콘텐츠를 연동하여 3d 서비스를 제공하기 위한 전송 장치 및 방법, 및 수신 장치 및 방법

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12752211

Country of ref document: EP

Kind code of ref document: A2

NENP Non-entry into the national phase

Ref country code: DE

REEP Request for entry into the european phase

Ref document number: 2012752211

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2012752211

Country of ref document: EP