WO2015196710A1 - 一种设备能力协商的方法、装置和计算机存储介质 - Google Patents

一种设备能力协商的方法、装置和计算机存储介质 Download PDF

Info

Publication number
WO2015196710A1
WO2015196710A1 PCT/CN2014/091679 CN2014091679W WO2015196710A1 WO 2015196710 A1 WO2015196710 A1 WO 2015196710A1 CN 2014091679 W CN2014091679 W CN 2014091679W WO 2015196710 A1 WO2015196710 A1 WO 2015196710A1
Authority
WO
WIPO (PCT)
Prior art keywords
information
layer video
decoding
capability
code stream
Prior art date
Application number
PCT/CN2014/091679
Other languages
English (en)
French (fr)
Inventor
李明
吴平
尚国强
谢玉堂
Original Assignee
中兴通讯股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中兴通讯股份有限公司 filed Critical 中兴通讯股份有限公司
Priority to EP14895886.1A priority Critical patent/EP3163878A4/en
Priority to US15/322,143 priority patent/US10375408B2/en
Priority to EP20171393.0A priority patent/EP3793198A1/en
Publication of WO2015196710A1 publication Critical patent/WO2015196710A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/44Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/61Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio
    • H04L65/612Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio for unicast
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/24Negotiation of communication capabilities
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/184Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/42Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
    • H04N19/423Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation characterised by memory arrangements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234327Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by decomposing into layers, e.g. base layer and one or more enhancement layers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/63Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
    • H04N21/633Control signals issued by server directed to the network components or client
    • H04N21/6332Control signals issued by server directed to the network components or client directed to client
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/63Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
    • H04N21/637Control signals issued by the client directed to the server or network components
    • H04N21/6377Control signals issued by the client directed to the server or network components directed to server

Definitions

  • the present invention relates to the field of video codec technology, and in particular, to a method, an apparatus, and a computer storage medium for device capability negotiation.
  • the HVC High Efficiency Video Coding, HEVC
  • 3DV three-dimensional video
  • MV-HEVC HEVC Multi-view video coding extension framework
  • SHVC In Scalable high-efficiency video coding
  • profile, level (Tier), level (Level) are used to represent the decoder resource requirements required to decode the code stream, ie (P, T, L).
  • level Tier
  • Level level
  • the decoder can use (P, T, L) to illustrate its own maximum decoding capability.
  • the decoder can correctly decode the code stream when the maximum decoding power of the decoder satisfies the decoder resource requirements required to decode the code stream.
  • the (P, T, L) requirement corresponding to the accumulated resources required for decoding the entire multi-layer video coded stream is declared, and the corresponding resources required for decoding each layer are declared (P, T, L) Requirements.
  • the (P, T, L) identification index value corresponding to the cumulative resource requirement may be the same as the H.265/HEVC basic layer (ie, the H.265/HEVC Version 1 standard that does not include the extended standard) (P, T, L)
  • the identification index value is the same, but the corresponding specific parameter values are different.
  • the value of the MaxLumaPs parameter corresponding to the Level ID of 3.1 in the Main file is 983040; in the SHVC standard extension
  • the parameter values corresponding to the (P, T, L) identification index values corresponding to the resources required to decode each layer are the same as the H.265/HEVC Version 1 standard.
  • the existing method can sufficiently identify the decoder capability required to decode the multi-layer video encoded code stream, at the beginning of the communication session establishment, the terminal decoder does not have any information about the multi-layer video encoded code stream that it is about to receive, therefore, It is not possible to directly provide the required output layer information for the server to judge and select the transmitted code stream. If the server sends the detailed information of the code stream to the terminal and then selects by the terminal, the session needs to be reciprocated multiple times during the session establishment process, and the efficiency is low and the time delay is large.
  • the server needs to resend the detailed information of the code stream, and the terminal selects the best output under the current conditions.
  • the layers combine and feed back the server request stream.
  • embodiments of the present invention are expected to provide a method, an apparatus, and a computer storage medium for device capability negotiation.
  • An embodiment of the present invention provides a method for device capability negotiation, where the method includes:
  • the first device sends response information to the second device, where the response information includes the following information At least one of: decoding capability information required to decode a multi-layer video stream, decoding capability information required to decode layers in a multi-layer video stream, and decoding capability required to decode different output layer combinations in a multi-layer video stream information.
  • the multi-layer video data request sent by the second device includes decoding capability information of the second device
  • the decoding capability information of the second device includes: capability information of the decoded multi-layer video code stream that the second device has, and/or capability information of the decoded base layer video code stream that the second device has.
  • the response information is sent to the second device at one or more of the following locations:
  • the decoding capability information is a decoder processing capability indicator defined by the profile profile, the level Tier, and the level parameter of the multi-layer video coding standard.
  • the decoder processing capability indicator includes: the image includes the maximum number of luma sampling points, the maximum capacity of the encoded image buffer CPB, the maximum number of slice segments that the image can divide, and the tile partition of the image partition.
  • the embodiment of the invention further provides a method for device capability negotiation, the method comprising:
  • the second device sends a multi-layer video data request to the first device, where the request includes: capability information of the decoded multi-layer video code stream that the second device has, and/or decoding of the second device Capability information of the base layer video stream;
  • the second device receives the response information sent by the first device.
  • the response information includes at least one of the following: decoding capability information required to decode the multi-layer video stream, decoding capability information required to decode each layer in the multi-layer video stream, and decoding the multi-layer video stream.
  • decoding capability information required for the combination of different output layers.
  • the response information is sent to the second device at one or more of the following locations:
  • An embodiment of the present invention further provides an apparatus for device capability negotiation, where the apparatus is applied to a first device, where the apparatus includes:
  • a request receiving unit configured to receive a multi-layer video data request sent by the second device
  • the response sending unit is configured to send the response information to the second device, where the response information includes at least one of the following information: decoding capability information required for decoding the multi-layer video code stream, and decoding required for each layer in the multi-layer video code stream Decoding capability information, decoding capability information required to decode different output layer combinations in a multi-layer video stream.
  • the multi-layer video data request sent by the second device includes decoding capability information of the second device, and the decoding capability information of the second device includes: a decoded multi-layer video code that the second device has Capability information of the stream, and/or capability information of the decoded base layer video code stream possessed by the second device.
  • the response sending unit is configured to: send the response information to the second device by placing the response information in one or more locations:
  • the decoding capability information is a decoder processing capability indicator defined by the profile profile, the level Tier, and the level parameter of the multi-layer video coding standard.
  • the decoder processing capability indicator includes: the image includes the maximum number of luma sampling points, the maximum capacity of the encoded image buffer CPB, the maximum number of slice segments that the image can divide, and the tile partition of the image partition.
  • An embodiment of the present invention further provides an apparatus for device capability negotiation, where the apparatus is applied to a second device, where the apparatus includes:
  • a request sending unit configured to send a multi-layer video data request to the first device, where the request includes: capability information of the decoded multi-layer video code stream that the second device has, and/or the second device Having capability information for decoding a base layer video stream;
  • the response receiving unit is configured to receive the response information sent by the first device.
  • the response information includes at least one of the following: decoding capability information required to decode the multi-layer video stream, decoding capability information required to decode each layer in the multi-layer video stream, and decoding the multi-layer video stream.
  • decoding capability information required for the combination of different output layers.
  • the response information sent by the first device is located at one or more of the following locations:
  • An embodiment of the present invention further provides a computer storage medium, the storage medium comprising a set of computer executable instructions, the instructions being used to perform a method for device capability negotiation of the first device.
  • An embodiment of the present invention further provides a computer storage medium, the storage medium comprising a set of computer executable instructions, the instructions being used to perform a method for device capability negotiation of the second device.
  • a method, a device, and a computer storage medium for the device capability negotiation when the second device sends a multi-layer video data request to the first device, notifying the first device of the decoding capability information a device; when the first device sends a response to the second device, the decoding capability information required to decode the multi-layer video stream, and/or the decoding capability information required to decode each layer in the multi-layer video stream, and/or the decoding
  • the decoding capability information required for the combination of different output layers in the layer video stream is informed to the second device.
  • the device capability negotiation is completed in the foregoing manner, so that the session establishment process does not need multiple reciprocating responses, which improves the efficiency of capability negotiation, reduces the delay, reduces the network burden, and optimizes the device processing resources.
  • FIG. 1 is a flowchart of a method for device capability negotiation according to Embodiment 1 of the present invention
  • FIG. 2 is a flowchart of a method for device capability negotiation according to Embodiment 2 of the present invention
  • FIG. 3 is a schematic structural diagram of an apparatus for device capability negotiation according to Embodiment 3 of the present invention.
  • FIG. 4 is a schematic structural diagram of an apparatus for device capability negotiation according to Embodiment 4 of the present invention.
  • a method for device capability negotiation according to Embodiment 1 of the present invention is as shown in FIG. 1 , where the method includes:
  • Step 101 The first device receives the multi-layer video data request sent by the second device.
  • the multi-layer video data request sent by the second device includes decoding capability information of the second device; and the decoding capability information of the second device includes: capability information of the decoded multi-layer video code stream that the second device has And/or capability information of the decoded base layer video code stream possessed by the second device.
  • the base layer code stream refers to a code stream of HEVC Version 1.
  • Step 102 The first device sends response information to the second device, where the response information includes at least one of the following information: decoding capability information required for decoding the multi-layer video code stream, and decoding required for each layer in the multi-layer video code stream. Decoding capability information, decoding capability information required to decode different output layer combinations in a multi-layer video stream.
  • the response information is sent to the second device at one or more of the following locations:
  • Data payload of the session negotiation protocol Parameter Set; Supplemental Enhancement Information (SEI); multi-layer video code stream; system layer information unit; custom data information unit;
  • SEI Supplemental Enhancement Information
  • the first device may decode the decoding capability information required for the multi-layer video stream, and/or decode the decoding capability information required for each layer in the multi-layer video stream, and/or decode the multi-layer video code.
  • the decoding capability information required for the combination of different output layers in the stream is placed in the data payload of the session negotiation protocol and sent to the second device;
  • the first device may also decode decoding capability information required for decoding the multi-layer video stream, and/or decode decoding capability information required for each layer in the multi-layer video stream, and/or decode different outputs in the multi-layer video stream.
  • the decoding capability information required for layer combination is placed in a parameter set unit of the multi-layer video coding standard and sent to the second device;
  • the first device may also decode decoding capability information required for decoding the multi-layer video stream, and/or decode decoding capability information required for each layer in the multi-layer video stream, and/or decode different outputs in the multi-layer video stream.
  • the decoding capability information required for layer combination is placed in a Supplemental Enhancement Information (SEI) unit of the multi-layer video coding standard and sent to the second device;
  • SEI Supplemental Enhancement Information
  • the first device may also include decoding capability information required to decode the multi-layer video stream, and/or decoding capability information required to decode layers in the multi-layer video stream, and/or decode different layers of the video stream.
  • the multi-layer video code stream of the decoding capability information required by the output layer combination is sent to the second device;
  • the first device may also decode decoding capability information required for decoding the multi-layer video stream, and/or decode decoding capability information required for each layer in the multi-layer video stream, and/or decode different outputs in the multi-layer video stream.
  • the decoding capability information required for layer combination is placed in the corresponding system layer information unit, and/or the custom data information unit, and/or the external information unit, and sent to the second device.
  • the first device may also use any combination of the above to send the decoding capability information required to decode the multi-layer video stream to the second device, and/or the decoding capability required to decode each layer in the multi-layer video stream.
  • the decoding capability information may be a decoder processing capability indicator defined by a profile, a Tier, and a Level parameter of the multi-layer video coding standard.
  • the decoder processing capability indicators include, but are not limited to, the image contains the maximum number of luma sampling points, the maximum capacity of the encoded image buffer (CPB), the maximum number of slice segments that the image can divide, and the tile that the image can divide.
  • the maximum number of rows of tiles, the tiles of the image that can be divided The maximum number of columns, the maximum sampling rate of the luminance component, the maximum code rate, and the maximum compression ratio.
  • the device capability negotiation method in the first embodiment is applied to the first device, and the first device may send the multi-layer video encoded code stream to the second device.
  • a method for device capability negotiation according to Embodiment 2 of the present invention is as shown in FIG. 2, where the method includes:
  • Step 201 The second device sends a multi-layer video data request to the first device, where the request includes: capability information of the decoded multi-layer video code stream possessed by the second device, and/or decoding basic information of the second device The capability information of the layer video stream.
  • Step 202 The second device receives the response information sent by the first device.
  • the response information includes at least one of the following information: decoding capability information required for decoding the multi-layer video stream, decoding capability information required for decoding each layer in the multi-layer video stream, and different decoding of the multi-layer video stream.
  • the output layer combines the required decoding capability information.
  • the response information may be sent to the second device at one or more of the following locations:
  • the device capability negotiation method in the foregoing Embodiment 2 is applied to the second device, and the second device may receive the multi-layer video coded code stream.
  • the third embodiment of the present invention provides a first device.
  • the device includes:
  • the request receiving unit 10 is configured to receive a multi-layer video data request sent by the second device;
  • the response sending unit 20 is configured to send the response information to the second device, where the response information includes at least one of the following information: decoding capability information required to decode the multi-layer video stream, and decoding the multi-layer view The decoding capability information required by each layer in the frequency code stream, and the decoding capability information required to decode different output layer combinations in the multi-layer video code stream.
  • the multi-layer video data request sent by the second device includes the decoding capability information of the second device, and the decoding capability information of the second device includes: capability information of the decoded multi-layer video code stream that the second device has, and / or capability information of the decoded base layer video code stream possessed by the second device.
  • the response sending unit 20 is configured to: send the response information to the second device by placing the response information in one or more of the following locations:
  • the decoding capability information is a decoder processing capability indicator defined by the Profile, Tier, and Level parameters in the multi-layer video coding standard.
  • the decoder processing capability indicators include: the image contains the maximum number of luma sampling points, the maximum capacity of the encoded image buffer (CPB), the maximum number of slice segments that can be divided by the image, and the tile that the image can be divided (Tile The maximum number of rows, the maximum number of columns of tiles that can be divided by the image, the maximum sampling rate of the luminance component, the maximum code rate, and the maximum compression ratio.
  • the embodiment of the present invention further provides a first device of the device for negotiating device capability according to the foregoing embodiment 3.
  • the request receiving unit 10 and the response sending unit 20 may be combined with a central processing unit (CPU) by a communication function chip of the first device.
  • CPU central processing unit
  • MPU Microprocessor
  • DSP Digital Signal Processor
  • FPGA Field-Programmable Gate Array
  • the fourth embodiment of the present invention provides a second device.
  • the device includes:
  • the request sending unit 30 is configured to send a multi-layer video data request to the first device, where the request includes: capability information of the decoded multi-layer video code stream possessed by the second device, and/or a decoding base layer of the second device Capability information of the video stream;
  • the response receiving unit 40 is configured to receive the response information sent by the first device.
  • the response information may include at least one of the following: decoding capability information required to decode the multi-layer video stream, decoding capability information required to decode each layer in the multi-layer video stream, and different decoding of the multi-layer video stream.
  • the output layer combines the required decoding capability information.
  • the response information sent by the first device may be located at one or more of the following: a data payload of the session negotiation protocol; a parameter set (Parameter Set); an SEI unit; a multi-layer video code stream; a system layer information unit; and a custom data.
  • Information unit external information unit.
  • the embodiment of the present invention further provides a second device, which is the device for negotiating device capability according to the fourth embodiment, wherein the request sending unit 30 and the response receiving unit 40 can be combined with the CPU, the MPU, and the DSP by the communication function chip of the second device. Or FPGA implementation.
  • the device capability negotiation method in the embodiment of the present invention may be applied to the terminal and the server, that is, the device capability negotiation method in the embodiment of the present invention may refer to that one terminal requests multiple layers of video from the server.
  • the device capability negotiation method in the embodiment of the present invention may be applied to between two terminals. That is, the device capability negotiation method in the embodiment of the present invention may refer to that one terminal requests a multi-layer video stream from another terminal. In the process, a method for performing capability negotiation between the two terminals; the first device is the one terminal, and the second device is the other terminal.
  • the second device in the embodiment of the present invention may request the multi-layer video encoded code stream and/or the transport stream and/or the media file from the first device by using the method for requesting data by the second device in the foregoing embodiment.
  • the second device in the embodiment of the present invention may be a code stream receiving and playing device in a video communication application, such as a mobile phone, a computer, a server, a set top box, a portable mobile terminal, a digital camera, and a television. Broadcast system equipment, etc.
  • the first device in the embodiment of the present invention may use the method for sending the response information by the first device in the foregoing embodiment, and send the decoding capability information required to decode the multi-layer video code stream to the second device, and/or decode multiple layers.
  • the first device in the embodiment of the present invention may be a related code stream generating and/or transmitting device in a video communication application, such as a mobile phone, a computer, a server, a set top box, a portable mobile terminal, a digital video camera, a television broadcasting system device, or the like.
  • the fifth embodiment provides a method for a terminal to perform capability negotiation with a server during a process of requesting a multi-layer video code stream from a server.
  • This embodiment uses the MV-HEVC code stream as an example for description.
  • the SHVC code stream processing method is the same as this.
  • MV-HEVC code streams On the server, there are 3 layers of MV-HEVC code streams, which are L0, L1 and L2, respectively.
  • the server can perform code stream extraction and editing operations on the 4-layer MV-HEVC code stream according to requirements, and the decoding capability required to decode the code stream corresponding to the combination of different output layers is as shown in Table 1 below:
  • Code stream 1 Output L0, L1 and L2 (MV-HEVC Main, 0, 6.0)
  • Code stream 2 Output L0 (HEVC Main, 0, 3.0)
  • Code stream 3 Output L1 (HEVC Main, 0, 4.1)
  • Code stream 4 Output L2 (HEVC Main, 0, 5.2)
  • Code stream 5 output L1, L2 (MV-HEVC Main, 0, 6.0)
  • the terminal After receiving the operation instruction of the user, the terminal sends the request information of the MV-HEVC program code stream to the server, where the request information includes the decoding capability information of the decoder used on the terminal.
  • the MV-HEVC decoder on terminal A uses two existing H.265/HEVC Version 1 Decoder implementation.
  • the decoding capability of the decoder when decoding the code stream (base layer code stream) and MV-HEVC code stream of H.265/HEVC Version 1 is as shown in Table 2 below:
  • MV-HEVC decoder (Profile, Tier, Level) MV-HEVC decoder (MV-HEVC Main, 0, 6.0) H.265/HEVC Version 1 Decoder 1 (HEVC Main, 0, 6.0) H.265/HEVC Version 1 Decoder 2 (HEVC Main, 0, 3.0)
  • the data request sent by the terminal A includes the following decoder decoding capability information: MV-HEVC Decoder (MV-HEVC Main, 0, 6.0), HEVC Version 1 Decoder (HEVC Main, 0, 6.0).
  • the server After receiving the data request of the terminal A, the server determines the response information sent to the terminal A. First, the server searches for an output layer combination suitable for the decoding capability of the terminal A from the output layer combination list of Table 1 based on the decoder capability information provided by the terminal A in the data request information. Since the MV-HEVC decoder of the terminal A has the decoding capability of (MV-HEVC Main, 0, 6.0), the H.265/HEVC Version 1 decoder of the terminal A has the decoding capability of (HEVC Main, 0, 6.0). Therefore, the server chooses to send the code stream 1 containing the three output layers L0, L1, and L2 to the terminal A.
  • the server sends a response message to the terminal A.
  • the server feeds back the output layer combination of the code stream 1 in Table 1 and the decoding capability information required by the combination to the terminal A.
  • the server may further provide decoding capability information required for separately outputting the code stream 2, the code stream 3, and the code stream 4 of L0, L1, and L2 in Table 1, for reference by the terminal A.
  • the server may further provide decoding capability information required to output the code stream 5 of L1 and L2 in Table 1 in the response information, so that the terminal A may directly directly determine that the code stream 1 is unsuitable for receiving and playing.
  • the code stream 5 is requested to be sent to the server, thereby improving the capability negotiation efficiency between the terminal A and the server.
  • the server can send the above response information to the terminal A in a plurality of ways.
  • the server may use the data payload of the capability negotiation protocol to send decoding capability information required for the code stream including different output combinations;
  • the server may write the decoding capability information and other video code stream related information required by the above code stream into a Video Parameter Set (VPS) of the MV-HEVC, and send the VPS to the terminal A;
  • VPS Video Parameter Set
  • the server may write the decoding capability information and other video code stream related information required by the foregoing code stream into an SEI of the MV-HEVC, and send the SEI to the terminal A;
  • the server can directly transmit the MV-HEVC video code stream containing the decoding capability information required by the code stream 1 to the terminal A.
  • the terminal A determines that the code stream 1 is suitable for receiving, decoding, and playing, the terminal A can directly receive the MV-HEVC code stream and send it to the decoder used by it for decoding. If the terminal A finds that the code stream 1 cannot be decoded, the terminal A may perform a code stream extraction operation on the code stream, obtain the code stream 5, perform decoding and playback, and feed back to the server, and the server stops sending the code stream 1 to the terminal A and switches to the code. The stream 5 is sent; in this case, the terminal A may not receive the code stream, and send a feedback message to the server to notify the server to replace the code stream 5;
  • the server may write the decoding capability information required by the above code stream to the system layer information unit, and send the system layer information unit as the response information to the terminal A;
  • the server may write the decoding capability information required by the above code stream to a custom data information unit (such as a custom SEI information unit), and send the customized data information unit as response information to the terminal A;
  • a custom data information unit such as a custom SEI information unit
  • the server may write the decoding capability information required by the foregoing code stream into the information unit used by the external path through an external path of the capability negotiation process (such as a dedicated feedback channel), and send the information element as the response information to the Terminal A.
  • an external path of the capability negotiation process such as a dedicated feedback channel
  • the MV-HEVC decoder on terminal B is implemented using four existing H.265/HEVC Version 1 decoders.
  • the decoder decodes H.265/HEVC Version 1
  • the decoding capabilities of the code stream (base layer code stream) and MV-HEVC code stream are shown in Table 3 below:
  • MV-HEVC decoder (Profile, Tier, Level) MV-HEVC decoder (MV-HEVC Main, 0, 4.0) H.265/HEVC Version 1 Decoder 1 (HEVC Main, 0, 3.0) H.265/HEVC Version 1 Decoder 2 (HEVC Main, 0, 3.0) H.265/HEVC Version 1 Decoder 3 (HEVC Main, 0, 3.0) H.265/HEVC Version 1 Decoder 4 (HEVC Main, 0, 3.0)
  • the data request sent by the terminal B includes the following decoder decoding capability information: MV-HEVC Decoder (MV-HEVC Main, 0, 4.0), HEVC Version 1 Decoder (HEVC Main, 0, 3.0).
  • the server After receiving the data request of the terminal B, the server compares the decoder processing capability information of the terminal B with the output capability information corresponding to each output layer combination in Table 1, and determines that the terminal B can only decode the code stream 2.
  • the server sends a response message to the terminal B.
  • the server feeds back the output layer combination of the code stream 2 in Table 1 and the decoding capability information required by the combination to the terminal B.
  • the method in which the server sends the response information to the terminal B is the same as the process of transmitting the response information to the terminal A, and will not be described again.
  • the MV-HEVC decoder in terminal C is not extended by the existing HEVC Version 1 decoder, but is designed and manufactured directly according to the MV-HEVC standard.
  • the MV-HEVC decoder can be backward compatible to decode the HEVC Version 1 code stream, and the decoding capability of the decoder when decoding the code stream (base layer code stream) and the MV-HEVC code stream of H.265/HEVC Version 1 is as shown in Table 4 below. Show:
  • MV-HEVC decoder (MV-HEVC Main, 0, 6.0) H.265/HEVC Version 1 decoder (HEVC Main, 0, 6.0)
  • the data request sent by the terminal C includes the following decoder decoding capability information: MV-HEVC Decoder (MV-HEVC Main, 0, 6.0), HEVC Version 1 Decoder (HEVC Main, 0, 6.0).
  • the server After receiving the data request of the terminal C, the server compares the decoder processing capability of the terminal C with the output capability corresponding to each output layer combination in Table 1, and determines that the terminal C can decode the code stream 5.
  • the server sends a response message to the terminal C.
  • the server feeds back the output layer combination of the code stream 5 in Table 1 and the decoding capability information required by the combination to the terminal C.
  • the response information sent by the server may also include the decoding capability information required for the combination of the output layer of the code stream 3 and the code stream 4, and is selected by the terminal C.
  • the method of the server transmitting the response information to the terminal C is the same as the foregoing process of transmitting the response information to the terminal A, and details are not described herein again.
  • the above implementation method can also be used in the case where the server actively pushes the multi-layer video data to the terminal.
  • the server first sends the push information of the multi-layer video data to the terminal.
  • the terminal After the terminal determines the video data pushed by the terminal, the terminal sends the confirmation information for receiving the push data to the server by using the foregoing implementation method, that is, the request for requesting the multi-layer video data from the server. information.
  • the request information includes the decoding capability information of the terminal.
  • the server transmits the decoding capability information required for decoding the multi-layer video stream to the terminal according to the received terminal data request information and the decoding capability information included therein.
  • This embodiment provides a method for two terminals to perform capability negotiation in a process in which one terminal (terminal A) requests a multi-layer video code stream from another terminal (terminal B).
  • This embodiment uses MV-HEVC
  • the code stream is taken as an example for description, and the SHVC code stream processing method is the same as this.
  • the terminal B stores or receives multi-layer video data, and the terminal A requests the multi-layer video data, such as a data sharing application, from the terminal B.
  • the implementation of this application may use the method described in embodiment 5.
  • the terminal A uses the terminal implementation method described in the fifth embodiment
  • the terminal B uses the server implementation method described in the fifth embodiment.
  • the terminal B stores or receives multi-layer video data. At the same time, the terminal B pushes the multi-layer video data to other terminals (such as the terminal A), such as a multi-screen sharing application.
  • the implementation of this application can use the method described in Embodiment 5.
  • the terminal A sends the terminal B the confirmation information for receiving the push data, that is, the request information for requesting the multi-layer video data from the terminal B.
  • the request information includes the decoding capability information of the terminal A.
  • the terminal B transmits the decoding capability information required for decoding the multi-layer video stream to the terminal A according to the received data request information of the terminal A and the decoding capability information included therein.
  • Terminal B can use the camera to capture and encode the captured video using a multi-layer video coding standard.
  • a communication link is established between the terminal A and the terminal B, and the terminal A requests the terminal B to multi-layer video data that is encoded in real time, such as real-time video sharing.
  • the terminal A includes the terminal decoding capability information in the terminal request information in the fifth embodiment in the multi-view video data request information sent to the terminal B.
  • an implementation manner of the terminal B is: the terminal B buffers the multi-layer video encoded code stream generated by the encoder, and generates the decoding of the multi-layer video encoded stream as shown in Table 1 in the fifth embodiment.
  • the output layer combines the decoding capability information required for the corresponding code stream, and accordingly, according to the implementation method of the server described in Embodiment 5, the decoding capability information required for decoding the multi-layer video stream is transmitted to the terminal A.
  • the implementation method can be used for the point-to-point communication mode between the terminal A and the terminal B, and can also be used for the point-to-multipoint communication of the multi-layer video data request of the terminal B simultaneously received by the terminal B. the way.
  • the terminal B uses the server implementation described in the fifth embodiment to separately transmit decoding capability information required for decoding the multi-layer video stream to a plurality of terminals to which the multi-layer video data is requested.
  • another implementation manner of the terminal B is: for the point-to-point communication mode of the terminal A and the terminal B, the terminal B saves the information of the multi-layer video coded stream that can be generated and has different decoding capability requirements, such as the following.
  • Table 5 shows:
  • Code stream that terminal B can generate (Profile, Tier, Level) Code stream 1: contains three layers of L0, L1 and L2 (MV-HEVC Main, 0, 6.0) Code stream 2: only contains the L0 layer (HEVC Main, 0, 3.0) Code stream 3: only contains the L1 layer (HEVC Main, 0, 4.1) Code stream 4: only contains the L2 layer (HEVC Main, 0, 5.2) Code stream 5: only contains L1 and L2 layers (MV-HEVC Main, 0, 6.0)
  • the terminal B generates a code stream according to the decoding capability information in the terminal A request information and the multi-layer video code stream in Table 5 that can generate different decoding capabilities (for example, the terminal A requests "HEVC Main, 0, 4.0"
  • the code stream at this time, the terminal B generates the code stream 2 of "HEVC Main, 0, 3.0" to ensure that the terminal A can correctly decode), and the terminal B uses the implementation of the server described in the fifth embodiment, except that the use is Table 5 replaces the table 1 described in Embodiment 5, and transmits the decoding capability information required for decoding the multi-layer video stream to the terminal A.
  • another implementation manner of the terminal B is: for the point-to-multipoint communication mode of the terminal B and the multiple terminals, the terminal B saves the multi-layer video coded stream that can be generated and has different decoding capability requirements.
  • Information is shown in Table 5.
  • the terminal B generates a code stream according to the lowest decoding capability of the plurality of terminal request information and the multi-layer video code stream in Table 5 that can generate different decoding capabilities (for example, the lowest decoding capability request "HEVC Main, 0" , 4.0" code stream, at this time terminal B generates "HEVC Main, 0, 3.0" code stream 2 to ensure that multiple terminals can be positive Indeed decoding), terminal B uses the embodiment of the server described in the fifth embodiment, except that Table 5 is used instead of Table 1 described in Embodiment 5, and multiple layers of video codes are transmitted to all terminals requesting data.
  • Terminal A and Terminal B can be photographed using a camera, and the captured video is encoded using a multi-layer video coding standard.
  • a communication link is established between the terminal A and the terminal B, and the terminal A requests the terminal B to request real-time encoded multi-layer video data, such as a videophone application.
  • the terminal A initiates a video call to the terminal B as an example for description.
  • the terminal A initiates a call request to the terminal B, requesting the multi-layer video data of the terminal B.
  • the terminal A is implemented by using the implementation method of the terminal A in the application scenario 3 in the embodiment, and the terminal B is implemented by using the implementation method of the terminal B in the application scenario 3 in the embodiment.
  • Terminal B agrees with the call request of terminal A, and sends call confirmation information to terminal A, that is, terminal B requests multi-layer video data from terminal A.
  • the terminal A is implemented by using the implementation method of the terminal B in the application scenario 3 in the embodiment, and the terminal B is implemented by using the implementation method of the terminal A in the application scenario 3 in the embodiment.
  • An embodiment of the present invention further provides a computer storage medium, the storage medium comprising a set of computer executable instructions, the instructions being used to perform a method for device capability negotiation of the first device.
  • An embodiment of the present invention further provides a computer storage medium, the storage medium comprising a set of computer executable instructions, the instructions being used to perform a method for device capability negotiation of the second device.
  • embodiments of the present invention can be provided as a method, system, or computer program product. Accordingly, the present invention can take the form of a hardware embodiment, a software embodiment, or a combination of software and hardware. Moreover, the present invention may employ computer-usable storage media (including but not limited to disks) in one or more of the computer-usable program code embodied therein. A form of computer program product embodied on a memory and optical storage, etc.).
  • the computer program instructions can also be stored in a computer readable memory that can direct a computer or other programmable data processing device to operate in a particular manner, such that the instructions stored in the computer readable memory produce an article of manufacture comprising the instruction device.
  • the apparatus implements the functions specified in one or more blocks of a flow or a flow and/or block diagram of the flowchart.
  • These computer program instructions can also be loaded onto a computer or other programmable data processing device such that a series of operational steps are performed on a computer or other programmable device to produce computer-implemented processing for execution on a computer or other programmable device.
  • the instructions provide steps for implementing the functions specified in one or more of the flow or in a block or blocks of a flow diagram.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Computer Security & Cryptography (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

一种设备能力协商的方法,包括:第一设备接收第二设备发送的多层视频数据请求(101);第一设备向第二设备发送应答信息(102),应答信息包括以下信息的至少之一:解码多层视频码流所需的解码能力信息、解码多层视频码流中各层所需的解码能力信息、解码多层视频码流中不同输出层组合所需的解码能力信息。还包括一种设备能力协商的装置和计算机存储介质。

Description

一种设备能力协商的方法、装置和计算机存储介质 技术领域
本发明涉及视频编解码技术领域,尤其涉及一种设备能力协商的方法、装置和计算机存储介质。
背景技术
正在制定的基于高性能视频编码(H.265/High Efficiency Video Coding,HEVC)标准兼容的三维视频(Three-Dimensional Video,3DV)编码标准MV-HEVC(HEVC Multi-view video coding extension framework)和可伸缩视频编码(Scalable high-efficiency video coding,SHVC)中,采用了统一的高层结构设计。这个统一的设计结构基于“多层视频编码”的概念,将MV-HEVC的不同视点和可伸缩编码的不同可伸缩层均抽象为“层(Layer)”,并使用层标识序号(LayerId)来标识不同的视点和可伸缩层。
在多层视频编码标准草案中,使用档次(Profile)、等级(Tier)、级别(Level)来表示解码码流所需的解码器资源需求,即(P,T,L)。同时,在会话建立过程中,解码器可以使用(P,T,L)来说明其自身的最大解码能力。当解码器的最大解码能力满足解码码流所需的解码器资源需求时,解码器可正确解码码流。在目前的多层视频编码标准草案中,声明解码整个多层视频编码码流所需的累积资源对应的(P,T,L)要求,声明解码各层所需资源对应的(P,T,L)要求。其中,累积资源需求对应的(P,T,L)的标识索引值可以与H.265/HEVC基本层(即不包含扩展标准的H.265/HEVC Version 1标准)的(P,T,L)标识索引值相同,但对应的具体参数值不同。例如,在H.265/HEVC Version 1标准中,Main档次下Level标识序号为3.1时对应的MaxLumaPs参数值是983040;在SHVC标准扩展 中,Scalable Main档次下累积资源需求的Level标识序号为3.1时对应的MaxLumaPs参数值是2×983040=1966080。而解码各层所需资源对应的(P,T,L)的标识索引值对应的参数值与H.265/HEVC Version 1标准相同。
在JCT-VC标准组织的技术提案JCTVC-R0043中,指出了使用上述方法在码流中标识解码能力的缺陷,并提出在上述信息之外,对多层视频编码码流的可能或需要使用的不同输出层的组合(即Partition),进一步分别标识其对解码器解码资源的需求。这样,两种不同的解码器(使用现有H.265/HEVC Version 1解码器实现的多层视频编码解码器和直接实现的多层视频编码解码器)均可判断是否能够正确解码码流。
现有方法虽然能够充分地标识解码多层视频编码码流所需的解码器能力,但在通信会话建立之初,终端解码器没有任何关于其即将接收的多层视频编码码流信息,因此,无法直接提供所需的输出层信息供服务器端判断和选择所发送的码流。如果在此之前服务器将其码流的详细信息发送给终端、再由终端进行选择,则其会话建立过程中需要多次往复应答,效率较低,时间延迟较大。并且,在网络通信状况发生变化时,如网络传输速率由低变高,为充分利用网络资源获得最佳用户体验质量,服务器需要重新发送码流的详细信息,终端选择当前条件下最佳的输出层组合并反馈服务器请求码流。这样往复多次的会话过程增加了网络负担,占用终端处理资源。
发明内容
为解决现有存在的技术问题,本发明实施例期望提供一种设备能力协商的方法、装置和计算机存储介质。
本发明实施例提供了一种设备能力协商的方法,所述方法包括:
第一设备接收第二设备发送的多层视频数据请求;
所述第一设备向第二设备发送应答信息,所述应答信息包括以下信息 的至少之一:解码多层视频码流所需的解码能力信息、解码多层视频码流中各层所需的解码能力信息、解码多层视频码流中不同输出层组合所需的解码能力信息。
其中,所述第二设备发送的多层视频数据请求中包括所述第二设备的解码能力信息;
所述第二设备的解码能力信息包括:所述第二设备所具有的解码多层视频码流的能力信息、和/或所述第二设备所具有的解码基本层视频码流的能力信息。
其中,所述应答信息放置在以下一个或多个位置上发送给所述第二设备:
会话协商协议的数据载荷;参数集Parameter Set;补充增强信息SEI;多层视频码流;系统层信息单元;自定义数据信息单元;外部信息单元。
其中,所述解码能力信息为多层视频编码标准中档次Profile、等级Tier、级别Level参数限定的解码器处理能力指标。
其中,所述解码器处理能力指标包括:图像包含亮度采样点的最大数量,编码图像缓冲区CPB的最大容量,图像可划分的分片划分Slice segment的最大数量,图像可划分的瓦片Tile的最大行数,图像可划分的瓦片的最大列数,亮度分量的最大采样率,最大码率,最大压缩比。
本发明实施例还提供了一种设备能力协商的方法,所述方法包括:
第二设备向第一设备发送多层视频数据请求,所述请求中包括:所述第二设备所具有的解码多层视频码流的能力信息、和/或所述第二设备所具有的解码基本层视频码流的能力信息;
所述第二设备接收所述第一设备发送的应答信息。
其中,所述应答信息包括以下信息的至少之一:解码多层视频码流所需的解码能力信息、解码多层视频码流中各层所需的解码能力信息、解码多层视频码流中不同输出层组合所需的解码能力信息。
其中,应答信息位于以下一个或多个位置上发送给所述第二设备:
会话协商协议的数据载荷;参数集Parameter Set;补充增强信息SEI;多层视频码流;系统层信息单元;自定义数据信息单元;外部信息单元。
本发明实施例还提供了一种设备能力协商的装置,所述装置应用于第一设备,所述装置包括:
请求接收单元,配置为接收第二设备发送的多层视频数据请求;
应答发送单元,配置为向第二设备发送应答信息,所述应答信息包括以下信息的至少之一:解码多层视频码流所需的解码能力信息、解码多层视频码流中各层所需的解码能力信息、解码多层视频码流中不同输出层组合所需的解码能力信息。
其中,所述第二设备发送的多层视频数据请求中包括所述第二设备的解码能力信息;所述第二设备的解码能力信息包括:所述第二设备所具有的解码多层视频码流的能力信息、和/或所述第二设备所具有的解码基本层视频码流的能力信息。
其中,应答发送单元配置为,将所述应答信息放置在以下一个或多个位置上发送给所述第二设备:
会话协商协议的数据载荷;参数集Parameter Set;补充增强信息SEI;多层视频码流;系统层信息单元;自定义数据信息单元;外部信息单元。
其中,所述解码能力信息为多层视频编码标准中档次Profile、等级Tier、级别Level参数限定的解码器处理能力指标。
其中,所述解码器处理能力指标包括:图像包含亮度采样点的最大数量,编码图像缓冲区CPB的最大容量,图像可划分的分片划分Slice segment的最大数量,图像可划分的瓦片Tile的最大行数,图像可划分的瓦片的最大列数,亮度分量的最大采样率,最大码率,最大压缩比。
本发明实施例还提供了一种设备能力协商的装置,所述装置应用于第二设备,所述装置包括:
请求发送单元,配置为向第一设备发送多层视频数据请求,所述请求中包括:所述第二设备所具有的解码多层视频码流的能力信息、和/或所述第二设备所具有的解码基本层视频码流的能力信息;
应答接收单元,配置为接收所述第一设备发送的应答信息。
其中,所述应答信息包括以下信息的至少之一:解码多层视频码流所需的解码能力信息、解码多层视频码流中各层所需的解码能力信息、解码多层视频码流中不同输出层组合所需的解码能力信息。
其中,所述第一设备发送的应答信息位于以下一个或多个位置上:
会话协商协议的数据载荷;参数集Parameter Set;补充增强信息SEI;多层视频码流;系统层信息单元;自定义数据信息单元;外部信息单元。
本发明实施例还提供了一种计算机存储介质,所述存储介质包括一组计算机可执行指令,所述指令用于执行所述第一设备的设备能力协商的方法。
本发明实施例还提供了一种计算机存储介质,所述存储介质包括一组计算机可执行指令,所述指令用于执行所述第二设备的设备能力协商的方法。
本发明实施例所提供的一种设备能力协商的方法、装置和计算机存储介质,在第二设备向第一设备发送多层视频数据请求时,将所述第二设备的解码能力信息告知第一设备;第一设备向第二设备发送应答时,解码多层视频码流所需的解码能力信息、和/或解码多层视频码流中各层所需的解码能力信息、和/或解码多层视频码流中不同输出层组合所需的解码能力信息告知第二设备。通过以上方式完成设备能力的协商,使得会话建立过程中无需多次往复应答,提高了能力协商的效率,降低了时延,减轻了网络负担,优化了设备处理资源。
附图说明
图1为本发明实施例一的设备能力协商的方法流程图;
图2为本发明实施例二的设备能力协商的方法流程图;
图3为本发明实施例三的设备能力协商的装置结构示意图;
图4为本发明实施例四的设备能力协商的装置结构示意图。
具体实施方式
下面结合附图和具体实施例对本发明的技术方案进一步详细阐述。
实施例一
本发明实施例一提供的一种设备能力协商的方法,如图1所示,该方法包括:
步骤101,第一设备接收第二设备发送的多层视频数据请求。
较佳的,所述第二设备发送的多层视频数据请求中包括第二设备的解码能力信息;第二设备的解码能力信息包括:第二设备所具有的解码多层视频码流的能力信息、和/或第二设备所具有的解码基本层视频码流的能力信息。其中,所述基本层码流是指HEVC Version 1的码流。
步骤102,第一设备向第二设备发送应答信息,所述应答信息包括以下信息的至少之一:解码多层视频码流所需的解码能力信息、解码多层视频码流中各层所需的解码能力信息、解码多层视频码流中不同输出层组合所需的解码能力信息。
较佳的,应答信息放置在以下一个或多个位置上发送给所述第二设备:
会话协商协议的数据载荷;参数集(Parameter Set);补充增强信息(Supplemental Enhancement Information,SEI);多层视频码流;系统层信息单元;自定义数据信息单元;外部信息单元。
也就是说,第一设备可以将解码多层视频码流所需的解码能力信息、和/或解码多层视频码流中各层所需的解码能力信息、和/或解码多层视频码 流中不同输出层组合所需的解码能力信息,放置在会话协商协议的数据载荷中发送给第二设备;
第一设备也可以将解码多层视频码流所需的解码能力信息、和/或解码多层视频码流中各层所需的解码能力信息、和/或解码多层视频码流中不同输出层组合所需的解码能力信息,放置在多层视频编码标准的参数集(Parameter Set)单元中发送给第二设备;
第一设备也可以将解码多层视频码流所需的解码能力信息、和/或解码多层视频码流中各层所需的解码能力信息、和/或解码多层视频码流中不同输出层组合所需的解码能力信息,放置在多层视频编码标准的补充增强信息(SEI)单元中发送给第二设备;
第一设备还可以将包含解码多层视频码流所需的解码能力信息、和/或解码多层视频码流中各层所需的解码能力信息、和/或解码多层视频码流中不同输出层组合所需的解码能力信息的多层视频码流发送给第二设备;
第一设备还可以将解码多层视频码流所需的解码能力信息、和/或解码多层视频码流中各层所需的解码能力信息、和/或解码多层视频码流中不同输出层组合所需的解码能力信息,放置在对应的系统层信息单元、和/或自定义数据信息单元、和/或外部信息单元中发送给第二设备。
当然,第一设备也可以采用以上任意组合的方式,来向第二设备发送解码多层视频码流所需的解码能力信息、和/或解码多层视频码流中各层所需的解码能力信息、和/或解码多层视频码流中不同输出层组合所需的解码能力信息。
具体的,解码能力信息可以为多层视频编码标准中档次(Profile)、等级(Tier)、级别(Level)参数限定的解码器处理能力指标。
解码器处理能力指标包括但不限于:图像包含亮度采样点的最大数量,编码图像缓冲区(CPB)的最大容量,图像可划分的分片划分(Slice segment)的最大数量,图像可划分的瓦片(Tile)的最大行数,图像可划分的瓦片的 最大列数,亮度分量的最大采样率,最大码率,最大压缩比。
以上实施例一所述的设备能力协商的方法,应用于第一设备,所述第一设备可以向第二设备发送多层视频编码码流。
实施例二
本发明实施例二提供的一种设备能力协商的方法,如图2所示,该方法包括:
步骤201,第二设备向第一设备发送多层视频数据请求,所述请求中包括:第二设备所具有的解码多层视频码流的能力信息、和/或第二设备所具有的解码基本层视频码流的能力信息。
步骤202,第二设备接收第一设备发送的应答信息。
具体的,应答信息包括以下信息的至少之一:解码多层视频码流所需的解码能力信息、解码多层视频码流中各层所需的解码能力信息、解码多层视频码流中不同输出层组合所需的解码能力信息。
较佳的,应答信息可以位于以下一个或多个位置上发送给第二设备:
会话协商协议的数据载荷;参数集(Parameter Set);SEI单元;多层视频码流;系统层信息单元;自定义数据信息单元;外部信息单元。
以上实施例二所述的设备能力协商的方法,应用于第二设备,所述第二设备可以接收多层视频编码码流。
实施例三
对应实施例一所述设备能力协商的方法,本发明的实施例三提供了一种第一设备,如图3所示,该设备包括:
请求接收单元10,配置为接收第二设备发送的多层视频数据请求;
应答发送单元20,配置为向第二设备发送应答信息,应答信息包括以下信息的至少之一:解码多层视频码流所需的解码能力信息、解码多层视 频码流中各层所需的解码能力信息、解码多层视频码流中不同输出层组合所需的解码能力信息。
其中,第二设备发送的多层视频数据请求中包括第二设备的解码能力信息;所述第二设备的解码能力信息包括:第二设备所具有的解码多层视频码流的能力信息、和/或第二设备所具有的解码基本层视频码流的能力信息。
其中,应答发送单元20配置为,将应答信息放置在以下一个或多个位置上发送给第二设备:
会话协商协议的数据载荷;参数集(Parameter Set);SEI单元;多层视频码流;系统层信息单元;自定义数据信息单元;外部信息单元。
其中,解码能力信息为多层视频编码标准中Profile、Tier、Level参数限定的解码器处理能力指标。
解码器处理能力指标包括:图像包含亮度采样点的最大数量,编码图像缓冲区(CPB)的最大容量,图像可划分的分片划分(Slice segment)的最大数量,图像可划分的瓦片(Tile)的最大行数,图像可划分的瓦片的最大列数,亮度分量的最大采样率,最大码率,最大压缩比。
本发明实施例还提供了一种包含上述实施例三所述设备能力协商的装置的第一设备,上述请求接收单元10和应答发送单元20可由第一设备的通信功能芯片结合中央处理器(CPU,Central Processing Unit)、微处理器(MPU,Micro Processing Unit)、数字信号处理器(DSP,Digital Signal Processor)或可编程逻辑阵列(FPGA,Field-Programmable Gate Array)共同实现。
实施例四
对应实施例二所述设备能力协商的方法,本发明的实施例四提供了一种第二设备,如图4所示,该设备包括:
请求发送单元30,配置为向第一设备发送多层视频数据请求,请求中包括:第二设备所具有的解码多层视频码流的能力信息、和/或第二设备所具有的解码基本层视频码流的能力信息;
应答接收单元40,配置为接收第一设备发送的应答信息。
其中,应答信息可以包括以下信息的至少之一:解码多层视频码流所需的解码能力信息、解码多层视频码流中各层所需的解码能力信息、解码多层视频码流中不同输出层组合所需的解码能力信息。
其中,第一设备发送的应答信息可以位于以下一个或多个位置上:会话协商协议的数据载荷;参数集(Parameter Set);SEI单元;多层视频码流;系统层信息单元;自定义数据信息单元;外部信息单元。
本发明实施例还提供了一种包含上述实施例四所述设备能力协商的装置的第二设备,上述请求发送单元30和应答接收单元40可由第二设备的通信功能芯片结合CPU、MPU、DSP或FPGA共同实现。
需要说明的是,本发明实施例的设备能力协商的方法可以适用于终端与服务器之间,也就是说,本发明实施例的设备能力协商方法可以是指,一台终端向服务器请求多层视频码流过程中,该终端与服务器之间进行能力协商的方法;所述第一设备为服务器,第二设备为终端。
本发明实施例的设备能力协商的方法可以适用于两台终端之间,也就是说,本发明实施例的设备能力协商方法可以是指,一台终端向另一台终端请求多层视频码流过程中,这两台终端之间进行能力协商的方法;所述第一设备为所述的一台终端,第二设备为所述的另一台终端。
本发明实施例中的第二设备可使用前述实施例中第二设备请求数据的方法,向第一设备请求多层视频编码码流和/或传输流和/或媒体文件。本发明实施例中的第二设备可以是视频通信应用中的码流接收播放设备,例如:手机、计算机、服务器、机顶盒、便携式移动终端、数字摄像机、电视广 播系统设备等。
本发明实施例中的第一设备可使用前述实施例中第一设备发送应答信息的方法,向第二设备发送包含有解码多层视频码流所需的解码能力信息、和/或解码多层视频码流中各层所需的解码能力信息、和/或解码多层视频码流中不同输出层组合所需的解码能力信息的应答。本发明实施例中的第一设备可以是视频通信应用中的相关码流生成和/或发送设备,例如:手机、计算机、服务器、机顶盒、便携式移动终端、数字摄像机、电视广播系统设备等。
实施例五
本实施例五提供一个终端向服务器请求多层视频码流过程中,与服务器进行能力协商的方法。本实施例以MV-HEVC码流为例进行说明,SHVC码流处理方式与此类同。
在服务器上,存在有3层的MV-HEVC码流,分别是L0、L1和L2。服务器可以根据需求对4层MV-HEVC码流进行码流提取和编辑操作,解码不同输出层的组合对应的码流所需解码能力如下表1所示:
输出层组合对应的码流 (Profile,Tier,Level)
码流1:输出L0、L1和L2 (MV-HEVC Main,0,6.0)
码流2:输出L0 (HEVC Main,0,3.0)
码流3:输出L1 (HEVC Main,0,4.1)
码流4:输出L2 (HEVC Main,0,5.2)
码流5:输出L1、L2 (MV-HEVC Main,0,6.0)
表1
终端在接收到用户的操作指令后,向服务器发送MV-HEVC节目码流的请求信息,在该请求信息中,包含终端上所使用解码器的解码能力信息。
终端A上的MV-HEVC解码器使用两个已有的H.265/HEVC Version 1 解码器实现。该解码器解码H.265/HEVC Version 1的码流(基本层码流)和MV-HEVC码流时的解码能力如下表2所示:
解码器 (Profile,Tier,Level)
MV-HEVC解码器 (MV-HEVC Main,0,6.0)
H.265/HEVC Version 1解码器1 (HEVC Main,0,6.0)
H.265/HEVC Version 1解码器2 (HEVC Main,0,3.0)
表2
终端A向服务器发送的数据请求,请求中包含两个或更多输出层的MV-HEVC码流。在终端A发送的数据请求中,包含如下解码器解码能力信息:MV-HEVC Decoder(MV-HEVC Main,0,6.0),HEVC Version 1Decoder(HEVC Main,0,6.0)。
服务器收到终端A的数据请求后,确定向终端A发送的应答信息。首先,服务器根据终端A在数据请求信息中提供的解码器能力信息,从表1的输出层组合列表中,寻找适合终端A解码能力的输出层组合。由于终端A的MV-HEVC解码器具有(MV-HEVC Main,0,6.0)的解码能力,同时,终端A的H.265/HEVC Version 1解码器具有(HEVC Main,0,6.0)的解码能力,因此,服务器选择向终端A发送包含有L0、L1和L2三个输出层的码流1。
服务器向终端A发送应答信息,在应答信息中,服务器将表1中码流1的输出层组合及该组合下所需的解码能力信息反馈给终端A。可选地,在应答信息中,服务器还可以提供表1中单独输出L0、L1和L2的码流2、码流3和码流4所需的解码能力信息,供终端A参考。可选地,服务器还可以在应答信息中提供表1中输出L1和L2的码流5所需的解码能力信息,这样,终端A在判断码流1不适合接收和播放的情况下,可以直接向服务器请求发送码流5,从而提高终端A与服务器之间的能力协商效率。
服务器可以使用多种途径向终端A发送上述应答信息。
其中,服务器可以使用能力协商协议的数据载荷发送包含不同输出组合的码流所需的解码能力信息;
服务器可以将上述码流所需的解码能力信息和其他视频码流相关信息写入到一个MV-HEVC的视频参数集(Video Parameter Set,VPS)中,并将该VPS发送给终端A;
服务器可以将上述码流所需的解码能力信息和其他视频码流相关信息写入到一个MV-HEVC的SEI中,将该SEI发送给终端A;
服务器可以直接将包含有码流1所需的解码能力信息的MV-HEVC视频码流发送给终端A。终端A在确定码流1适合接收、解码和播放的情况下,可以直接接收该MV-HEVC码流,将其送交其所使用的解码器进行解码。如果终端A发现不能解码码流1,终端A可以对该码流进行码流抽取操作,获得码流5后进行解码播放,并反馈给服务器,服务器停止给终端A发送码流1并切换至码流5进行发送;这种情况下,终端A也可以不接收码流,给服务器发送反馈信息通知服务器更换为码流5;
服务器可以将上述码流所需的解码能力信息写入到系统层信息单元,将该系统层信息单元作为应答信息发送给终端A;
服务器可以将上述码流所需的解码能力信息写入到自定义的数据信息单元(如自定义的SEI信息单元),将该自定义的数据信息单元作为应答信息发送给终端A;
服务器可以通过该能力协商过程的外部途径(如专用的反馈信道),将上述码流所需的解码能力信息写入到该外部途径所使用的信息单元中,将该信息单元作为应答信息发送给终端A。
与终端A类似,终端B上的MV-HEVC解码器使用四个已有的H.265/HEVC Version 1解码器实现。该解码器解码H.265/HEVC Version 1的 码流(基本层码流)和MV-HEVC码流时的解码能力如下表3所示:
解码器 (Profile,Tier,Level)
MV-HEVC解码器 (MV-HEVC Main,0,4.0)
H.265/HEVC Version 1解码器1 (HEVC Main,0,3.0)
H.265/HEVC Version 1解码器2 (HEVC Main,0,3.0)
H.265/HEVC Version 1解码器3 (HEVC Main,0,3.0)
H.265/HEVC Version 1解码器4 (HEVC Main,0,3.0)
表3
终端B向服务器发送的数据请求,请求中包含一个或更多输出层的MV-HEVC码流。在终端B发送的数据请求中,包含如下解码器解码能力信息:MV-HEVC Decoder(MV-HEVC Main,0,4.0),HEVC Version 1Decoder(HEVC Main,0,3.0)。
服务器收到终端B的数据请求后,将终端B的解码器处理能力信息与表1中各输出层组合对应的输出能力信息进行比对,确定终端B仅能够解码码流2。服务器向终端B发送应答信息,在应答信息中,服务器将表1中码流2的输出层组合及该组合下所需的解码能力信息反馈给终端B。这里,服务器向终端B发送应答信息的方法与前述向终端A发送应答信息的过程相同,不再赘述。
与终端A和终端B不同,终端C中的MV-HEVC解码器没有已有的HEVC Version 1解码器扩展而成,而是直接根据MV-HEVC标准设计和制造。MV-HEVC解码器可以后向兼容解码HEVC Version 1码流,该解码器解码H.265/HEVC Version 1的码流(基本层码流)和MV-HEVC码流时的解码能力如下表4所示:
解码器 (Profile,Tier,Level)
MV-HEVC解码器 (MV-HEVC Main,0,6.0)
H.265/HEVC Version 1解码器 (HEVC Main,0,6.0)
表4
终端C向服务器发送的数据请求,请求中包含最多两个输出层的MV-HEVC码流。在终端C发送的数据请求中,包含如下解码器解码能力信息:MV-HEVC Decoder(MV-HEVC Main,0,6.0),HEVC Version 1Decoder(HEVC Main,0,6.0)。
服务器收到终端C的数据请求后,将终端C的解码器处理能力与表1中各输出层组合对应的输出能力进行比对,确定终端C可以解码码流5。服务器向终端C发送应答信息,在应答信息中,服务器将表1中码流5的输出层组合及该组合下所需的解码能力信息反馈给终端C。可选地,在服务器发送的应答信息中,也可以包含码流3、码流4的输出层组合所需的解码能力信息,供终端C选择。这里,服务器向终端C发送应答信息的方法与前述向终端A发送应答信息的过程相同,不再赘述。
需要说明的是,上述实施方法也可用于服务器主动向终端推送多层视频数据的情况。服务器首先向终端发送多层视频数据的推送信息,在终端确定接收服务器推送的视频数据后,终端使用上述实施方法向服务器发送接收推送数据的确认信息,也即向服务器请求多层视频数据的请求信息。在该请求信息里,包含终端的解码能力信息。服务器根据所接收的终端数据请求信息及其包含的解码能力信息,使用上述实施方法向终端发送多层视频码流解码所需的解码能力信息。
实施例六
本实施例提供一台终端(终端A)向另一台终端(终端B)请求多层视频码流过程中,两台终端设备进行能力协商的方法。本实施例以MV-HEVC 码流为例进行说明,SHVC码流处理方式与此类同。
应用场景1:
终端B存储或接收多层视频数据,终端A向终端B请求该多层视频数据,如数据共享应用。本应用的实施可以使用实施例五所述的方法。具体地,终端A使用实施例五中所述的终端实施方法,终端B使用实施例五中所述的服务器实施方法。
应用场景2:
终端B存储或接收多层视频数据。同时,终端B向其他终端(如终端A)推送该多层视频数据,如多屏分享应用。本应用的实施可以使用实施例五所述方法。具体地,终端A在确定接收终端B推送的多层视频数据后,向终端B发出接收推送数据的确认信息,即向终端B请求多层视频数据的请求信息。在该请求信息里,包含终端A的解码能力信息。终端B根据所接收的终端A的数据请求信息及其包含的解码能力信息,使用实施例五所述服务器实施方法向终端A发送多层视频码流解码所需的解码能力信息。
应用场景3:
终端B可以使用摄像头进行拍摄,对拍摄得到的视频使用多层视频编码标准进行编码。终端A与终端B之间建立了通信链接,终端A向终端B请求实时编码的多层视频数据,如实时视频共享。
终端A在发送给终端B的多视点视频数据请求信息中,包含实施例五中所述终端请求信息中的终端解码能力信息。
可选的,终端B的一种实施方法是:终端B缓存了其编码器产生的多层视频编码码流,并对该多层视频编码流生成如实施例五中表1所示的解码不同输出层组合对应码流所需的解码能力信息,并据此使用实施例五中所述服务器的实施方法,向终端A发送多层视频码流解码所需的解码能力信息。这种实施方法可用于终端A与终端B之间的点对点通信方式,也可以用于终端B同时收到多个终端的多层视频数据请求的一点对多点的通信 方式。在一点对多点的通信方式中,终端B使用实施例五所述的服务器实施方式,分别向多个向其请求多层视频数据的终端发送多层视频码流解码所需的解码能力信息。
可选的,终端B的另一种实施方式是:对于终端A与终端B的点对点通信方式,终端B保存其可生成的、对解码能力需求不同的多层视频编码码流的信息,如以下表5所示:
终端B可生成的码流 (Profile,Tier,Level)
码流1:包含L0、L1和L2三层 (MV-HEVC Main,0,6.0)
码流2:仅包含L0层 (HEVC Main,0,3.0)
码流3:仅包含L1层 (HEVC Main,0,4.1)
码流4:仅包含L2层 (HEVC Main,0,5.2)
码流5:仅包含L1和L2层 (MV-HEVC Main,0,6.0)
表5
终端B根据终端A请求信息中的解码能力信息及表5中其所能产生需要不同解码能力的多层视频码流的情况,产生码流(例如,终端A请求“HEVC Main,0,4.0”的码流,此时终端B产生“HEVC Main,0,3.0”的码流2以保证终端A可以正确解码),终端B使用实施例五中所述的服务器的实施方式,不同之处在于使用表5代替实施例五中所述的表1,向终端A发送多层视频码流解码所需的解码能力信息。
可选的,终端B的再一种实施方式是:对于终端B与多个终端的一点对多点的通信方式,终端B保存其可生成的、对解码能力需求不同的多层视频编码码流的信息,如表5所示。终端B根据多个终端请求信息中的解码能力最低者及表5中其所能产生需要不同解码能力的多层视频码流的情况,产生码流(例如,最低解码能力请求“HEVC Main,0,4.0”的码流,此时终端B产生“HEVC Main,0,3.0”的码流2以保证多个终端均可正 确解码),终端B使用实施例五中所述的服务器的实施方式,不同之处在于使用表5代替实施例五中所述的表1,向全部请求数据的多个终端发送多层视频码流解码所需的解码能力信息。
应用场景4:
终端A和终端B可以使用摄像头进行拍摄,对拍摄得到的视频使用多层视频编码标准进行编码。终端A与终端B之间建立了通信链接,终端A向终端B请求实时编码的多层视频数据,如可视电话应用。
这里,以终端A向终端B发起可视电话为例进行说明。
终端A向终端B发起呼叫请求,请求终端B的多层视频数据。终端A使用本实施例所述应用场景3中所述终端A实施方法进行实施,终端B使用本实施例所述应用场景3中所述终端B的实施方法进行实施。
终端B同意终端A的呼叫请求,向终端A发送呼叫确认信息,即终端B向终端A请求多层视频数据。终端A使用本实施例所述应用场景3中所述终端B的实施方法进行实施,终端B使用本实施例所述应用场景3中所述终端A的实施方法进行实施。
本发明实施例还提供了一种计算机存储介质,所述存储介质包括一组计算机可执行指令,所述指令用于执行所述第一设备的设备能力协商的方法。
本发明实施例还提供了一种计算机存储介质,所述存储介质包括一组计算机可执行指令,所述指令用于执行所述第二设备的设备能力协商的方法。
本领域内的技术人员应明白,本发明的实施例可提供为方法、系统、或计算机程序产品。因此,本发明可采用硬件实施例、软件实施例、或结合软件和硬件方面的实施例的形式。而且,本发明可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘 存储器和光学存储器等)上实施的计算机程序产品的形式。
本发明是参照根据本发明实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。
这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。
这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。
以上所述,仅为本发明的较佳实施例而已,并非用于限定本发明的保护范围。

Claims (18)

  1. 一种设备能力协商的方法,所述方法包括:
    第一设备接收第二设备发送的多层视频数据请求;
    所述第一设备向第二设备发送应答信息,所述应答信息包括以下信息的至少之一:解码多层视频码流所需的解码能力信息、解码多层视频码流中各层所需的解码能力信息、解码多层视频码流中不同输出层组合所需的解码能力信息。
  2. 根据权利要求1所述设备能力协商的方法,其中,所述第二设备发送的多层视频数据请求中包括所述第二设备的解码能力信息;
    所述第二设备的解码能力信息包括:所述第二设备所具有的解码多层视频码流的能力信息、和/或所述第二设备所具有的解码基本层视频码流的能力信息。
  3. 根据权利要求1所述设备能力协商的方法,其中,所述应答信息放置在以下一个或多个位置上发送给所述第二设备:
    会话协商协议的数据载荷;参数集Parameter Set;补充增强信息SEI;多层视频码流;系统层信息单元;自定义数据信息单元;外部信息单元。
  4. 根据权利要求1、2或3所述设备能力协商的方法,其中,所述解码能力信息为多层视频编码标准中档次Profile、等级Tier、级别Level参数限定的解码器处理能力指标。
  5. 根据权利要求4所述设备能力协商的方法,其中,所述解码器处理能力指标包括:图像包含亮度采样点的最大数量,编码图像缓冲区CPB的最大容量,图像可划分的分片划分Slice segment的最大数量,图像可划分的瓦片Tile的最大行数,图像可划分的瓦片的最大列数,亮度分量的最大采样率,最大码率,最大压缩比。
  6. 一种设备能力协商的方法,所述方法包括:
    第二设备向第一设备发送多层视频数据请求,所述请求中包括:所述第二设备所具有的解码多层视频码流的能力信息、和/或所述第二设备所具有的解码基本层视频码流的能力信息;
    所述第二设备接收所述第一设备发送的应答信息。
  7. 根据权利要求6所述设备能力协商的方法,其中,所述应答信息包括以下信息的至少之一:解码多层视频码流所需的解码能力信息、解码多层视频码流中各层所需的解码能力信息、解码多层视频码流中不同输出层组合所需的解码能力信息。
  8. 根据权利要求7所述设备能力协商的方法,其中,应答信息位于以下一个或多个位置上发送给所述第二设备:
    会话协商协议的数据载荷;参数集Parameter Set;补充增强信息SEI;多层视频码流;系统层信息单元;自定义数据信息单元;外部信息单元。
  9. 一种设备能力协商的装置,所述装置应用于第一设备,所述装置包括:
    请求接收单元,配置为接收第二设备发送的多层视频数据请求;
    应答发送单元,配置为向第二设备发送应答信息,所述应答信息包括以下信息的至少之一:解码多层视频码流所需的解码能力信息、解码多层视频码流中各层所需的解码能力信息、解码多层视频码流中不同输出层组合所需的解码能力信息。
  10. 根据权利要求9所述设备能力协商的装置,其中,所述第二设备发送的多层视频数据请求中包括所述第二设备的解码能力信息;所述第二设备的解码能力信息包括:所述第二设备所具有的解码多层视频码流的能力信息、和/或所述第二设备所具有的解码基本层视频码流的能力信息。
  11. 根据权利要求9所述设备能力协商的装置,其中,应答发送单元配置为,将所述应答信息放置在以下一个或多个位置上发送给所述第二 设备:
    会话协商协议的数据载荷;参数集Parameter Set;补充增强信息SEI;多层视频码流;系统层信息单元;自定义数据信息单元;外部信息单元。
  12. 根据权利要求9、10或11所述设备能力协商的装置,其中,所述解码能力信息为多层视频编码标准中档次Profile、等级Tier、级别Level参数限定的解码器处理能力指标。
  13. 根据权利要求12所述设备能力协商的装置,其中,所述解码器处理能力指标包括:图像包含亮度采样点的最大数量,编码图像缓冲区CPB的最大容量,图像可划分的分片划分Slice segment的最大数量,图像可划分的瓦片Tile的最大行数,图像可划分的瓦片的最大列数,亮度分量的最大采样率,最大码率,最大压缩比。
  14. 一种设备能力协商的装置,所述装置应用于第二设备,所述装置包括:
    请求发送单元,配置为向第一设备发送多层视频数据请求,所述请求中包括:所述第二设备所具有的解码多层视频码流的能力信息、和/或所述第二设备所具有的解码基本层视频码流的能力信息;
    应答接收单元,配置为接收所述第一设备发送的应答信息。
  15. 根据权利要求14所述设备能力协商的装置,其中,所述应答信息包括以下信息的至少之一:解码多层视频码流所需的解码能力信息、解码多层视频码流中各层所需的解码能力信息、解码多层视频码流中不同输出层组合所需的解码能力信息。
  16. 根据权利要求15所述设备能力协商的装置,其中,所述第一设备发送的应答信息位于以下一个或多个位置上:
    会话协商协议的数据载荷;参数集Parameter Set;补充增强信息SEI;多层视频码流;系统层信息单元;自定义数据信息单元;外部信息单元。
  17. 一种计算机存储介质,所述存储介质包括一组计算机可执行指 令,所述指令用于执行权利要求1-5任一项所述设备能力协商的方法。
  18. 一种计算机存储介质,所述存储介质包括一组计算机可执行指令,所述指令用于执行权利要求6-8任一项所述设备能力协商的方法。
PCT/CN2014/091679 2014-06-25 2014-11-19 一种设备能力协商的方法、装置和计算机存储介质 WO2015196710A1 (zh)

Priority Applications (3)

Application Number Priority Date Filing Date Title
EP14895886.1A EP3163878A4 (en) 2014-06-25 2014-11-19 Device capability negotiation method and apparatus, and computer storage medium
US15/322,143 US10375408B2 (en) 2014-06-25 2014-11-19 Device capability negotiation method and apparatus, and computer storage medium
EP20171393.0A EP3793198A1 (en) 2014-06-25 2014-11-19 Device capability negotiation method and apparatus, and computer storage medium

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201410291451.5 2014-06-25
CN201410291451.5A CN104093028B (zh) 2014-06-25 2014-06-25 一种设备能力协商的方法和装置

Related Child Applications (1)

Application Number Title Priority Date Filing Date
EP20171393.0A Previously-Filed-Application EP3793198A1 (en) 2014-06-25 2014-11-19 Device capability negotiation method and apparatus, and computer storage medium

Publications (1)

Publication Number Publication Date
WO2015196710A1 true WO2015196710A1 (zh) 2015-12-30

Family

ID=51640703

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2014/091679 WO2015196710A1 (zh) 2014-06-25 2014-11-19 一种设备能力协商的方法、装置和计算机存储介质

Country Status (4)

Country Link
US (1) US10375408B2 (zh)
EP (2) EP3793198A1 (zh)
CN (1) CN104093028B (zh)
WO (1) WO2015196710A1 (zh)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104093028B (zh) * 2014-06-25 2019-02-01 中兴通讯股份有限公司 一种设备能力协商的方法和装置
US20180213216A1 (en) * 2015-06-16 2018-07-26 Lg Electronics Inc. Media data transmission device, media data reception device, media data transmission method, and media data rececption method
CN107431669B (zh) * 2015-10-26 2020-12-01 华为技术有限公司 协商对象的选择方法、响应发现消息的方法、相关装置
CN111132130B (zh) * 2018-10-30 2022-08-02 中兴通讯股份有限公司 一种数据通讯方法、基站、终端、系统及存储介质
CN114845134B (zh) * 2020-10-16 2023-01-24 腾讯科技(深圳)有限公司 文件封装方法、文件传输方法、文件解码方法及相关设备
CN113259690A (zh) * 2021-07-05 2021-08-13 人民法院信息技术服务中心 一种跨网系的音视频实时在线协同系统及方法
CN114205337B (zh) * 2021-11-05 2024-03-15 海能达通信股份有限公司 媒体能力协商方法、装置、设备及计算机可读存储介质
WO2023163632A1 (en) * 2022-02-25 2023-08-31 Telefonaktiebolaget Lm Ericsson (Publ) Neural network complexity metric for image processing

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100202453A1 (en) * 2009-02-12 2010-08-12 Qualcomm Incorporated Association procedure to enable multiple multicast streams
CN102123299A (zh) * 2011-01-11 2011-07-13 中国联合网络通信集团有限公司 可伸缩视频的播放方法和播放装置
CN102223201A (zh) * 2010-04-15 2011-10-19 中兴通讯股份有限公司 一种编解码器能力协商方法及终端
CN104093028A (zh) * 2014-06-25 2014-10-08 中兴通讯股份有限公司 一种设备能力协商的方法和装置

Family Cites Families (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6584093B1 (en) * 1998-08-25 2003-06-24 Cisco Technology, Inc. Method and apparatus for automatic inter-domain routing of calls
CN100550908C (zh) * 2005-12-13 2009-10-14 华为技术有限公司 一种进行会话能力信息操作的方法及网络实体
CN100433725C (zh) * 2006-03-30 2008-11-12 华为技术有限公司 一种优选编解码的解决方法
EP2051525A1 (en) * 2007-10-15 2009-04-22 Mitsubishi Electric Information Technology Centre Europe B.V. Bandwidth and content dependent transmission of scalable video layers
EP2406954A1 (en) * 2009-03-13 2012-01-18 Telefonaktiebolaget L M Ericsson (PUBL) Technique for bringing encoded data items into conformity with a scalable coding protocol
CN102215210A (zh) * 2010-04-02 2011-10-12 中兴通讯股份有限公司 在ip多媒体子系统中建立会话的方法和装置
WO2013009441A2 (en) * 2011-07-12 2013-01-17 Vidyo, Inc. Scalable video coding using multiple coding technologies
US9591318B2 (en) * 2011-09-16 2017-03-07 Microsoft Technology Licensing, Llc Multi-layer encoding and decoding
US11089343B2 (en) * 2012-01-11 2021-08-10 Microsoft Technology Licensing, Llc Capability advertisement, configuration and control for video coding and decoding
EP2642755B1 (en) * 2012-03-20 2018-01-03 Dolby Laboratories Licensing Corporation Complexity scalable multilayer video coding
US9451256B2 (en) * 2012-07-20 2016-09-20 Qualcomm Incorporated Reusing parameter sets for video coding
CN103634697B (zh) * 2012-08-24 2017-09-26 中兴通讯股份有限公司 网真技术的实现方法和网真设备
US9161039B2 (en) * 2012-09-24 2015-10-13 Qualcomm Incorporated Bitstream properties in video coding
US9473771B2 (en) * 2013-04-08 2016-10-18 Qualcomm Incorporated Coding video data for an output layer set
EP3011747A1 (en) * 2013-06-18 2016-04-27 VID SCALE, Inc. Inter-layer parameter set for hevc extensions
US9578342B2 (en) * 2013-06-24 2017-02-21 Dialogic Corporation Rate control algorithm for scalable video encoding with disposable P-frames
CN103702061A (zh) * 2013-12-23 2014-04-02 上海网达软件股份有限公司 一种智能终端视频通话分辨率自动调节方法
KR20160104678A (ko) * 2014-01-02 2016-09-05 브이아이디 스케일, 인크. Hevc 확장 규격을 위한 서브 비트스트림 추출 프로세스
KR20150081236A (ko) * 2014-01-03 2015-07-13 삼성전자주식회사 효율적인 파라미터 전달을 사용하는 비디오 부호화 방법 및 그 장치, 비디오 복호화 방법 및 그 장치
WO2015174893A1 (en) * 2014-05-13 2015-11-19 Telefonaktiebolaget L M Ericsson (Publ) Methods, decoder and encoder for selection of reference pictures to be used during encoding

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100202453A1 (en) * 2009-02-12 2010-08-12 Qualcomm Incorporated Association procedure to enable multiple multicast streams
CN102223201A (zh) * 2010-04-15 2011-10-19 中兴通讯股份有限公司 一种编解码器能力协商方法及终端
CN102123299A (zh) * 2011-01-11 2011-07-13 中国联合网络通信集团有限公司 可伸缩视频的播放方法和播放装置
CN104093028A (zh) * 2014-06-25 2014-10-08 中兴通讯股份有限公司 一种设备能力协商的方法和装置

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP3163878A4 *

Also Published As

Publication number Publication date
CN104093028B (zh) 2019-02-01
EP3793198A1 (en) 2021-03-17
US20170223371A1 (en) 2017-08-03
US10375408B2 (en) 2019-08-06
CN104093028A (zh) 2014-10-08
EP3163878A4 (en) 2017-06-07
EP3163878A1 (en) 2017-05-03

Similar Documents

Publication Publication Date Title
WO2015196710A1 (zh) 一种设备能力协商的方法、装置和计算机存储介质
TWI829058B (zh) 用於在多媒體通訊中使用壓縮並行轉碼器的方法和裝置
JP2016167841A (ja) 移動体メディアサーバ
KR101882596B1 (ko) 비트스트림의 생성과 처리 방법, 장치 및 시스템
CN104685886A (zh) 用于视频处理的装置及其方法
JP2009117896A (ja) 合成映像配信装置ならびにその方法およびプログラム
CN109803056A (zh) 一种视频彩铃播放的方法、装置及系统
EP2974321A2 (en) Resilience in the presence of missing media segments in dynamic adaptive streaming over http
US11044278B2 (en) Transcoding capability configuration method and device and computer storage medium
CN111131755A (zh) 加入音视频会议的方法、流媒体协转服务器
WO2012095032A1 (zh) 指示和获取用户流数的方法及通信装置
US20160261643A1 (en) Signalling assymetric media capabilities
CN104333765A (zh) 一种视频直播流的处理方法及处理装置
WO2021093882A1 (zh) 一种视频会议方法、会议终端、服务器及存储介质
CN104702970A (zh) 一种同步视频数据的方法、设备及系统
CN114143567A (zh) 直播方法、装置、电子设备及计算机可读存储介质
CN117836815A (zh) 点云数据发送设备、点云数据发送方法、点云数据接收设备以及点云数据接收方法
CN114157919A (zh) 数据处理的方法和系统、云终端、服务器和计算设备
CN114500914A (zh) 音视频转发方法、装置、终端与系统
CN105227527A (zh) 播放多媒体的方法及系统、服务器

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14895886

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

REEP Request for entry into the european phase

Ref document number: 2014895886

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2014895886

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 15322143

Country of ref document: US