WO2024104307A1 - 直播视频流渲染方法、装置、设备、存储介质及产品 - Google Patents

直播视频流渲染方法、装置、设备、存储介质及产品 Download PDF

Info

Publication number
WO2024104307A1
WO2024104307A1 PCT/CN2023/131376 CN2023131376W WO2024104307A1 WO 2024104307 A1 WO2024104307 A1 WO 2024104307A1 CN 2023131376 W CN2023131376 W CN 2023131376W WO 2024104307 A1 WO2024104307 A1 WO 2024104307A1
Authority
WO
WIPO (PCT)
Prior art keywords
video stream
target
rendering
special effect
live
Prior art date
Application number
PCT/CN2023/131376
Other languages
English (en)
French (fr)
Inventor
王璨
游东
李嘉维
Original Assignee
北京字跳网络技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京字跳网络技术有限公司 filed Critical 北京字跳网络技术有限公司
Publication of WO2024104307A1 publication Critical patent/WO2024104307A1/zh

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs

Definitions

  • the embodiments of the present disclosure relate to the field of image processing technology, and in particular to a live video stream rendering method, device, equipment, storage medium and product.
  • VR virtual reality
  • the existing live broadcasting scenarios are generally aimed at ordinary live broadcasting on mobile phones. Since the resolution of live broadcasting on mobile phones is low, algorithms for special effects rendering can be installed on mobile phones to implement algorithm recognition and rendering operations on mobile phones.
  • the live broadcast screen is generally an ultra-high resolution screen.
  • the embodiments of the present disclosure provide a live video stream rendering method, apparatus, device, storage medium and product, which are used to solve the technical problem that the existing rendering method cannot guarantee the real-time performance of the rendering operation and the clarity of the rendering result.
  • an embodiment of the present disclosure provides a live video stream rendering method, which is applied to a cloud rendering system, wherein the cloud rendering system includes a cloud server and a virtual reality device, and the method includes:
  • the cloud server obtains a live video stream and special effect identification information corresponding to the virtual reality live broadcast, where the special effect identification information is identification information of a target special effect used to render the live video stream;
  • the cloud server determines a rendering mode corresponding to the target special effect according to the special effect identification information, wherein the rendering mode includes a first rendering mode and/or a second rendering mode;
  • the cloud server When the rendering mode is the first rendering mode, the cloud server performs special effect algorithm processing according to the live video stream and the special effect identification information, and performs special effect rendering processing according to the recognition result corresponding to the special effect algorithm processing, obtains the target video stream, and sends the target video stream to the virtual reality device, and the virtual reality device obtains the target video stream sent by the cloud server and plays it;
  • the cloud server When the rendering mode is the second rendering mode, the cloud server performs special effect algorithm processing according to the live video stream and the special effect identification information, and sends the live video stream and the recognition result to the virtual reality device.
  • the virtual reality device performs special effect rendering processing according to the live video stream and the algorithm result, obtains the processed target video stream, and plays the processed target video stream.
  • an embodiment of the present disclosure provides a live video stream rendering method, which is applied to a cloud server, comprising:
  • the special effect identification information is identification information of a target special effect used to render the live video stream
  • the target video stream is sent to a preset virtual reality device for playback.
  • an embodiment of the present disclosure provides a live video stream rendering method, which is applied to a virtual reality device, including:
  • Target video stream sent by the cloud server, wherein the target video stream is obtained after the cloud server performs rendering processing according to the live video stream corresponding to the virtual reality live broadcast and the special effect identification information, using a rendering method corresponding to the target special effect;
  • an embodiment of the present disclosure provides a live video stream rendering device, which is applied to a cloud server, wherein a first acquisition module is used to acquire a live video stream and special effect identification information corresponding to a virtual reality live broadcast, wherein the special effect identification information is identification information of a target special effect used to render the live video stream;
  • a determination module used to determine a rendering mode corresponding to the target special effect according to the special effect identification information
  • a rendering module used to render the live video stream in a rendering mode corresponding to the target special effect to obtain a processed target video stream
  • the sending module is used to send the target video stream to a preset virtual reality device for playback.
  • an embodiment of the present disclosure provides a live video stream rendering device, which is applied to a virtual reality device, including:
  • a second acquisition module is used to acquire a target video stream sent by the cloud server, wherein the target video stream is obtained after the cloud server performs rendering processing according to the live video stream corresponding to the virtual reality live broadcast and the special effect identification information, using a rendering method corresponding to the target special effect;
  • a processing module used for processing the target video stream by adopting a rendering method corresponding to the target special effect to obtain a processed target video stream
  • the playing module is used to play the processed target video stream.
  • an embodiment of the present disclosure provides a cloud rendering system, including a cloud server and a virtual reality device, wherein the cloud server is communicatively connected with the virtual reality device;
  • the cloud server is used to obtain a live video stream and special effect identification information corresponding to the virtual reality live broadcast, and determine a rendering method corresponding to the target special effect according to the special effect identification information, wherein the special effect identification information is identification information of the target special effect used to render the live video stream;
  • the cloud server is further configured to perform special effect algorithm processing and special effect rendering processing according to the live video stream and the special effect identification information to obtain a target video stream when the rendering mode is the first rendering mode;
  • the cloud server is further configured to perform special effect algorithm processing according to the live video stream and the special effect identification information when the rendering mode is the second rendering mode, and to The identification result is sent to the virtual reality device;
  • the virtual reality device is used to obtain the live video stream sent by the cloud server and the recognition result for special effects rendering.
  • an embodiment of the present disclosure provides an electronic device, including: a processor and a memory;
  • the memory stores computer-executable instructions
  • the processor executes the computer-executable instructions stored in the memory, so that the at least one processor executes the live video stream rendering method described in the first aspect and various possible designs of the first aspect or the second aspect and various possible designs of the second aspect or the third aspect and various possible designs of the third aspect.
  • an embodiment of the present disclosure provides a computer-readable storage medium, wherein the computer-readable storage medium stores computer execution instructions.
  • the live video stream rendering method described in the first aspect and various possible designs of the first aspect or the second aspect and various possible designs of the second aspect or the third aspect and various possible designs of the third aspect is implemented.
  • an embodiment of the present disclosure provides a computer program product, including a computer program, which, when executed by a processor, implements the live video stream rendering method described in the first aspect and various possible designs of the first aspect, or the second aspect and various possible designs of the second aspect, or the third aspect and various possible designs of the third aspect.
  • the live video stream rendering method, device, equipment, storage medium and product provided in this embodiment can avoid putting pressure on the virtual reality device at the viewing end by implementing algorithm recognition processing and/or rendering processing of the live video stream corresponding to the virtual reality live broadcast in the cloud server after acquiring the live video stream and special effect identification information corresponding to the virtual reality live broadcast.
  • the target special effects can be rendered in a targeted manner, thereby improving the accuracy and real-time performance of special effect rendering.
  • FIG1 is a diagram of the system architecture on which the present disclosure is based
  • FIG2 is a schematic diagram of a flow chart of a live video stream rendering method provided by an embodiment of the present disclosure
  • FIG3 is a flow chart of a live video stream rendering method provided by another embodiment of the present disclosure.
  • FIG4 is a schematic diagram of a flow chart of a live video stream rendering method provided by another embodiment of the present disclosure.
  • FIG5 is a schematic diagram of a flow chart of a live video stream rendering method provided by another embodiment of the present disclosure.
  • FIG6 is a schematic diagram of a flow chart of a live video stream rendering method provided by an embodiment of the present disclosure
  • FIG7 is a schematic diagram of the structure of a live video stream rendering device provided by an embodiment of the present disclosure.
  • FIG8 is a schematic diagram of the structure of a live video stream rendering device provided by an embodiment of the present disclosure.
  • FIG. 9 is a schematic diagram of the structure of an electronic device provided in an embodiment of the present disclosure.
  • the present disclosure provides a live video stream rendering method, device, equipment, storage medium and product.
  • live video stream rendering method, device, equipment, storage medium and product provided by the present disclosure can be applied to any virtual reality live rendering scenario to ensure the real-time performance of virtual reality live rendering and the clarity of the live content.
  • the inventors found through research that in order to solve the problem of high-quality VR live broadcast images, the calculation of special effect algorithms cannot be done on the Native side because the computing power of the Native side is limited. Therefore, the calculation of special effect algorithms can be performed on the cloud server, which can fully utilize the computing power of the cloud server and save the computing power originally on the Native side. It can ensure the clarity and real-time performance of the rendering results.
  • FIG1 is a diagram of the system architecture on which the present disclosure is based.
  • the system architecture on which the present disclosure is based includes at least a server 11, a live broadcast companion server 12, and a virtual reality device 13.
  • the server 11 is provided with a live broadcast video stream rendering device, which can be written in languages such as C/C++, Java, Shell, or Python;
  • the virtual reality device 13 can be a VR device such as VR glasses that can realize the viewing of live broadcast content.
  • the server 11 can obtain the live video stream and special effect identification information corresponding to the virtual reality live broadcast sent by the live companion server 12. Based on the live video stream and the special effect identification information, the live video stream is rendered using a rendering method corresponding to the target special effect to obtain the processed target video stream, and the target video stream is sent to the preset virtual reality device 13 for playback.
  • the present disclosure also provides a live video stream rendering method, which is applied to a cloud rendering system, the cloud rendering system includes a cloud server and a virtual reality device, and the method includes:
  • the cloud server obtains a live video stream and special effect identification information corresponding to the virtual reality live broadcast, where the special effect identification information is identification information of a target special effect used to render the live video stream;
  • the cloud server determines a rendering mode corresponding to the target special effect according to the special effect identification information, wherein the rendering mode includes a first rendering mode and/or a second rendering mode;
  • the cloud server When the rendering mode is the first rendering mode, the cloud server performs special effect algorithm processing according to the live video stream and the special effect identification information, and performs special effect rendering processing according to the recognition result corresponding to the special effect algorithm processing, obtains the target video stream, and sends the target video stream to the virtual reality device, and the virtual reality device obtains the target video stream sent by the cloud server and plays it;
  • the cloud server When the rendering mode is the second rendering mode, the cloud server performs special effect algorithm processing according to the live video stream and the special effect identification information, and sends the live video stream and the recognition result to the virtual reality device.
  • the virtual reality device performs special effect rendering processing according to the live video stream and the algorithm result, obtains the processed target video stream, and plays the processed target video stream.
  • This embodiment provides a cloud rendering system, which includes at least a cloud server and a virtual reality device.
  • the cloud server can obtain the live broadcast corresponding to the virtual reality live broadcast.
  • the cloud server can determine the rendering method corresponding to the target special effect according to the special effect identification information.
  • the rendering method includes a first rendering method and/or a second rendering method.
  • the special effect algorithm and special effect rendering can be implemented in the cloud server.
  • the rendered target video stream is obtained, and the target video stream is sent to the virtual reality device. Therefore, after the virtual reality device obtains the target video stream, it does not need to perform further rendering processing and can directly play the target video stream.
  • the special effect algorithm can be performed on the cloud server, and the recognition result corresponding to the special effect algorithm can be sent to the virtual reality device together with the live video stream. Accordingly, after obtaining the recognition result and the live video stream, the virtual reality device can perform special effect rendering processing based on the recognition result and the live video stream. In this way, the real-time nature of the target special effect can be guaranteed, and the host can be supported to interact with the target special effect.
  • the special effect algorithm processing and special effect rendering processing can be implemented on the cloud server for all target special effects, thereby effectively saving the computing power of the virtual reality device and ensuring the effect of the virtual reality live broadcast.
  • the cloud server can process the special effects algorithm for all target special effects, and send the recognition result corresponding to the special effects algorithm processing to the virtual reality device.
  • the virtual reality device can perform special effects rendering processing on the live video stream according to the recognition result. In this way, the real-time special effects rendering can be guaranteed, and the anchor can interact with the target special effects, further optimizing the live broadcast effect.
  • the rendering mode includes the first rendering mode and the second rendering mode
  • different rendering modes can be selected for different target special effects to perform rendering operations, so that special effect rendering operations can be implemented in a targeted manner.
  • the real-time performance of the special effect rendering operation is guaranteed.
  • the cloud server performs special effect algorithm processing according to the live video stream and the special effect identification information, including:
  • the cloud server acquires target data for rendering the target special effect in a preset configuration platform according to the special effect identification information
  • the cloud server determines a target rendering area corresponding to the target special effect according to the target data
  • the cloud server performs an identification operation on the live video stream to obtain an identification result corresponding to the live video stream, wherein the identification result includes key point information corresponding to the target rendering area in a live image frame corresponding to the live video stream and depth data corresponding to the live image frame;
  • a configuration platform may be pre-set, and the cloud server may also be connected to the configuration platform in communication.
  • the configuration platform may store target data corresponding to a plurality of special effects, wherein the target data includes but is not limited to the rendering mode corresponding to the special effect, the rendering time corresponding to the special effect, the rendering position corresponding to the special effect, decorations, etc.
  • the cloud server can obtain the target data corresponding to the target special effect in the preset configuration platform according to the special effect identification information.
  • the target rendering area and depth information corresponding to the target special effect can be determined according to the target data. Furthermore, the live video stream can be identified to obtain the identification result corresponding to the live video stream, wherein the identification result includes the key point information corresponding to the target rendering area in the live image frame corresponding to the live video stream and the depth data corresponding to the live image frame.
  • the special effects rendering process is performed on the recognition result corresponding to the special effects algorithm to obtain the target video stream, including:
  • the cloud server performs a rendering operation on the live video stream according to the recognition result and the target data to obtain a rendered video stream;
  • the cloud server performs encoding operation on the rendered video stream to obtain the target video stream.
  • the decoration in the target data can be accurately rendered according to the key point information and depth data in the recognition result to obtain a rendered video stream.
  • the rendered video stream can be encoded to obtain the target video stream.
  • the target video stream can be obtained.
  • the virtual reality device does not need to perform special effects algorithms and special effects rendering processing locally, which effectively saves the computing power of the virtual reality device.
  • the cloud server determines the rendering mode corresponding to the target special effect according to the special effect identification information, it also includes:
  • the cloud server acquires target data for rendering the target special effect in a preset configuration platform according to the special effect identification information
  • the cloud server determines a target rendering area corresponding to the target special effect according to the target data
  • the cloud server performs an identification operation on the live video stream to obtain an identification result corresponding to the live video stream, wherein the identification result includes key point information corresponding to the target rendering area in a live image frame corresponding to the live video stream and depth data corresponding to the live image frame;
  • the cloud server performs encoding operation on the live video stream, the recognition result and the target data to obtain the target video stream, and sends the target video stream to the virtual reality device;
  • the virtual reality device acquires the target video stream, performs a decoding operation on the target video stream, and obtains a live video stream corresponding to the virtual reality live broadcast, a recognition result corresponding to the live video stream, and target data, wherein the recognition result includes key point information corresponding to the target rendering area in a live image frame corresponding to the live video stream and depth data corresponding to the live image frame;
  • the live video stream is rendered according to the recognition result and the target data to obtain a processed target video stream.
  • the rendering mode corresponding to the target special effect is the second rendering mode, that is, the special effect algorithm operation is executed on the cloud server, and the special effect rendering processing is performed on the virtual reality device.
  • the cloud server can determine the target rendering area and depth corresponding to the target special effect according to the target data. Furthermore, the live video stream can be identified to obtain the identification result corresponding to the live video stream, wherein the identification result includes the key point information corresponding to the target rendering area in the live image frame corresponding to the live video stream and the depth data corresponding to the live image frame.
  • the cloud server can encode the live video stream, the recognition result and the target data to obtain the target video stream. Sent to a virtual reality device.
  • the method further includes:
  • the virtual reality device can obtain the target video stream sent by the cloud server. After obtaining the target video stream, the virtual reality device can decode the target video stream to obtain the live video stream corresponding to the virtual reality live broadcast, the recognition result corresponding to the live video stream, the target data and the rendering time, wherein the recognition result includes the key point information corresponding to the target rendering area in the live image frame corresponding to the live video stream and the depth data corresponding to the live image frame.
  • the live video stream is rendered according to the recognition result and the target data to obtain a processed target video stream, so that the subsequent virtual reality device can play the target video stream.
  • the special effect algorithm is calculated in the cloud server and the special effect rendering is performed in the virtual reality device, thereby saving the computing power of the virtual reality device and ensuring the real-time nature of the special effect. This allows the user to interact with the special effect.
  • FIG2 is a flow chart of a live video stream rendering method provided by an embodiment of the present disclosure, which is applied to a cloud server. As shown in FIG2 , the method includes:
  • Step 201 Acquire a live video stream and special effect identification information corresponding to the virtual reality live broadcast, where the special effect identification information is identification information of a target special effect used to render the live video stream.
  • the execution subject of this embodiment is a live video stream rendering device, which can be coupled to a cloud server.
  • the cloud server is respectively connected to the virtual reality device and the virtual reality live server.
  • the anchor can initiate a virtual reality live broadcast through a preset live broadcast application.
  • the live broadcast application can be installed in the virtual reality live broadcast server.
  • the live video stream can be collected through a preset binocular image acquisition device.
  • the anchor can select a target special effect for display.
  • the audience of the virtual reality live broadcast can also trigger a virtual resource transfer operation according to actual needs, and the virtual resource transfer operation can correspond to the target special effect one by one.
  • the virtual reality live broadcast server can obtain the live video stream and the special effect identification information respectively, and send the live video stream and the special effect identification information to the cloud server.
  • the cloud server can obtain the live video stream and the special effect identification information.
  • Step 202 Determine a rendering method corresponding to the target special effect according to the special effect identification information.
  • different target special effects may correspond to different display effects, rendering areas, etc.
  • the target special effect may be a special effect of a virtual glasses accessory, which acts on the host's face and can interact with the host in real time.
  • a pair of glasses that can perform corresponding interactive operations based on the host's movements can be rendered on the host's face.
  • the target special effect has strict requirements on timeliness.
  • the target special effect can be an underwater world special effect, which is applied to the entire live broadcast room.
  • the target special effect is triggered, the background where the anchor is located can be rendered.
  • the target special effect does not need to interact with the anchor, so there is no requirement for timeliness.
  • Different rendering methods can be used for the above-mentioned different target special effects.
  • special effect calculation can be implemented on the cloud server to obtain recognition results for the live video stream. Rendering operations are performed in the virtual reality device to ensure real-time interaction with the user.
  • special effect calculation and special effect rendering can be implemented on the cloud server, and the virtual reality device can directly play the target video stream sent by the cloud server.
  • the rendering method corresponding to the target special effect can be determined according to the special effect identification information.
  • Step 203 Render the live video stream using a rendering method corresponding to the target special effect to obtain a processed target video stream.
  • the cloud server can use the rendering method corresponding to the target special effect to render the live video stream to obtain the processed target video stream.
  • the rendering process includes at least one of special effect algorithm processing and special effect rendering processing.
  • Step 204 Send the target video stream to a preset virtual reality device for playback.
  • the target video stream after obtaining the target video stream, can be sent to a preset virtual reality device for playback.
  • the computing power of the virtual reality device can be effectively saved, reducing the pressure on the virtual reality device.
  • the live video stream rendering method provided in this embodiment realizes the virtual reality live broadcast in the cloud server after acquiring the live video stream and special effect identification information corresponding to the virtual reality live broadcast.
  • the corresponding algorithm recognition processing and/or rendering processing of the live video stream can avoid putting pressure on the virtual reality device at the viewing end.
  • the target special effects can be rendered in a targeted manner, thereby improving the accuracy and real-time performance of special effect rendering.
  • step 202 includes:
  • target data corresponding to the target special effect is obtained in a preset configuration platform.
  • a rendering mode corresponding to the target special effect is determined in the target data.
  • the rendering mode includes a first rendering mode and/or a second rendering mode.
  • a configuration platform may be pre-set, and target data corresponding to multiple special effects may be stored in the configuration platform, wherein the target data includes but is not limited to the rendering mode corresponding to the special effect, the rendering time corresponding to the special effect, the rendering position corresponding to the special effect, and decorations.
  • the target data corresponding to the target special effect can be obtained in the preset configuration platform according to the special effect identification information, and the rendering method corresponding to the target special effect can be determined in the target data.
  • the rendering method includes a first rendering method and/or a second rendering method.
  • the first rendering method may be a cloud rendering method, that is, the special effect algorithm and the special effect rendering are both performed in the cloud server.
  • the second rendering method may be a local rendering method, that is, the special effect algorithm is performed in the cloud server, and the special effect rendering processing is performed in the virtual reality device.
  • the live video stream rendering method provided in this embodiment can determine the rendering mode corresponding to the target special effect by obtaining the target data corresponding to the target special effect in the preset configuration platform according to the special effect identification information. Then, the rendering operation of the target special effect can be accurately implemented according to the rendering mode corresponding to the target special effect, thereby improving the accuracy of the rendering operation while saving the computing power of the virtual reality device.
  • the method further includes:
  • the rendering time corresponding to the target special effect is determined in a preset configuration platform according to the special effect identification information.
  • the rendering time is encoded into the target video stream, so that the virtual reality device performs a rendering operation on the live video stream within the rendering time.
  • different target special effects have different corresponding rendering times, which can be determined by the developer of the target special effect and stored in the configuration platform.
  • A needs to be rendered in the third second after the user triggers the virtual resource transfer operation, while the target special effect B needs to be rendered immediately after the user triggers the virtual resource transfer operation.
  • the rendering time corresponding to the target special effect can be determined in the preset configuration platform according to the special effect identification information.
  • the rendering time is encoded into the target video stream so that the virtual reality device can render the live video stream within the rendering time.
  • the live video stream rendering method provided in this embodiment encodes the rendering time into the target video stream, so that the virtual reality device can accurately render the target special effect according to the rendering time, thereby ensuring the rendering effect of the target special effect.
  • FIG3 is a flow chart of a live video stream rendering method provided by another embodiment of the present disclosure. Based on any of the above embodiments, as shown in FIG3 , step 201 includes:
  • Step 301 Obtain the data to be rendered sent by the virtual reality live broadcast server.
  • Step 302 Decode the data to be rendered to obtain the live video stream and preset special effect rendering instructions.
  • Step 303 parse the special effect rendering instruction to obtain the special effect identification information.
  • the virtual reality live broadcast server can obtain the live video stream and special effect identification information respectively, encode the special effect identification information together with the live video stream in the form of assembly instructions, obtain the data to be rendered, and send the data to be rendered to the cloud server.
  • the corresponding information for controlling the special effects rendering can be carried during the transmission of the live video stream, making the special effects rendering operation more in line with the user's personalized needs and improving the user experience.
  • the cloud server can decode the data to be rendered to obtain the live video stream and the preset special effect rendering instructions, and further parse the special effect rendering instructions to obtain special effect identification information.
  • the live video stream rendering method provided in this embodiment can obtain the live video stream by decoding the data to be rendered sent by the virtual reality live broadcast server.
  • the rendering operation can be accurately performed based on the target special effect triggered by the user.
  • FIG4 is a flowchart of a live video stream rendering method provided by another embodiment of the present disclosure. Based on any of the above embodiments, the rendering method includes a first rendering method. As shown in FIG4 , step 203 includes:
  • Step 401 Determine a target rendering area corresponding to the target special effect according to the target data.
  • Step 402 perform a recognition operation on the live video stream to obtain a recognition result corresponding to the live video stream, wherein the recognition result includes key point information corresponding to the target rendering area in a live image frame corresponding to the live video stream and depth data corresponding to the live image frame.
  • Step 403 Render the live video stream according to the recognition result and the target data to obtain a rendered video stream.
  • Step 404 Encode the rendered video stream to obtain the target video stream.
  • the rendering mode when the rendering mode is the first rendering mode, it is necessary to perform special effect algorithms and special effect rendering processing in the cloud server.
  • the target rendering area and depth and other information corresponding to the target special effect can be determined according to the target data. Furthermore, the live video stream can be identified to obtain the identification result corresponding to the live video stream, wherein the identification result includes the key point information corresponding to the target rendering area in the live image frame corresponding to the live video stream and the depth data corresponding to the live image frame.
  • the decoration in the target data can be accurately rendered according to the key point information and depth data in the recognition result to obtain a rendered video stream. Furthermore, in order to facilitate data transmission, the rendered video stream can be encoded to obtain the target video stream.
  • step 403 includes:
  • a rendering operation is performed on a plurality of live image frames corresponding to the live video stream according to the recognition result and the target data to obtain a plurality of rendered target image frames.
  • the rendered video stream is generated according to the multiple rendered target image frames.
  • the recognition result corresponding to each live image frame can be determined, and the live image frame can be rendered according to the recognition result and the target data.
  • the above operation is repeated to obtain multiple rendered target image frames.
  • a rendered video stream can be generated according to the multiple rendered target image frames.
  • the live video stream rendering method provided in this embodiment can obtain the target video stream by implementing the algorithm recognition operation and rendering operation on the live video stream in the cloud server when the rendering mode is the first rendering mode.
  • the virtual reality device After sending the target video stream to the virtual reality device, the virtual reality device does not need to perform special effect algorithm and special effect rendering processing locally, which effectively saves the computing power of the virtual reality device. And performing special effect algorithms and special effect rendering processing on the cloud server can ensure the rendering effect of the target video stream after special effect rendering and improve the user experience.
  • FIG5 is a flowchart of a live video stream rendering method provided by another embodiment of the present disclosure. Based on any of the above embodiments, the rendering method includes a second rendering method. As shown in FIG5 , step 203 includes:
  • Step 501 Determine a target rendering area corresponding to the target special effect according to the target data.
  • Step 502 perform a recognition operation on the live video stream to obtain a recognition result corresponding to the live video stream, wherein the recognition result includes key point information corresponding to the target rendering area in a live image frame corresponding to the live video stream and depth data corresponding to the live image frame.
  • Step 503 Encode the live video stream, the recognition result, and the target data to obtain the target video stream.
  • the rendering mode corresponding to the target special effect is the second rendering mode, that is, the special effect algorithm operation is executed on the cloud server, and the special effect rendering processing is performed on the virtual reality device.
  • the target rendering area and depth and other information corresponding to the target special effect can be determined according to the target data. Furthermore, the live video stream can be identified to obtain the identification result corresponding to the live video stream, wherein the identification result includes the key point information corresponding to the target rendering area in the live image frame corresponding to the live video stream and the depth data corresponding to the live image frame.
  • the live video stream, the recognition result and the target data are encoded to obtain the target video stream.
  • the virtual reality device can render the live video stream based on the recognition result and the target data.
  • the performing an identification operation on the live video stream to obtain an identification result corresponding to the live video stream includes:
  • the key point information corresponding to the target rendering area in the live image frame corresponding to the live video stream is identified through a preset key point recognition model.
  • the depth data corresponding to the live image frame is identified by a preset depth recognition model.
  • the key point information and depth data corresponding to each live image frame are determined as the recognition result corresponding to the live video stream.
  • a key point recognition model and a depth recognition model can be pre-set. Therefore, the key point information corresponding to the target rendering area in the live image frame corresponding to the live video stream can be recognized by the preset key point recognition model.
  • the depth data corresponding to the live image frame is recognized by the preset depth recognition model.
  • the key point information and depth data corresponding to each live image frame are determined as the recognition result corresponding to the live video stream.
  • the key point recognition model and the depth recognition model can accurately realize the recognition operation of the key point information and the depth data, thereby improving the rendering effect of the special effect rendering.
  • the live video stream rendering method provided in this embodiment calculates the special effect algorithm in the cloud server and performs special effect rendering processing in the virtual reality device when the rendering mode is the second rendering mode, thereby ensuring the real-time nature of the special effect while saving the computing power of the virtual reality device, so that the user can interact with the special effect.
  • FIG6 is a flow chart of a live video stream rendering method provided by an embodiment of the present disclosure, which is applied to a virtual reality device. As shown in FIG6 , the method includes:
  • Step 601 Acquire a target video stream sent by the cloud server, wherein the target video stream is obtained after the cloud server performs rendering processing according to the live video stream corresponding to the virtual reality live broadcast and special effect identification information, using a rendering method corresponding to the target special effect.
  • Step 602 Process the target video stream using a rendering method corresponding to the target special effect to obtain a processed target video stream.
  • Step 603 Play the processed target video stream.
  • the execution subject of this embodiment is a live video stream rendering device, which can be coupled to a virtual reality device.
  • the virtual reality device can be connected to a cloud server for communication.
  • the cloud server can obtain the live video stream and special effect identification information, and determine the rendering method corresponding to the target special effect according to the special effect identification information.
  • the live video stream is rendered using the rendering method corresponding to the target special effect to obtain a processed target video stream.
  • the target video stream is sent to a preset virtual reality device for playback.
  • the virtual reality device can obtain the target video stream, and process the target video stream using a rendering method corresponding to the target special effect to obtain a processed target video stream, and play the processed target video stream.
  • the live video stream rendering method provided in this embodiment obtains the video stream sent by the cloud server.
  • the target video stream is obtained after rendering processing using the rendering method corresponding to the target special effect, so that the virtual reality device only needs to perform simple processing on the target video stream to realize the display of virtual reality live broadcast, which effectively saves the computing power of the virtual reality device and can ensure the live broadcast effect of virtual reality live broadcast.
  • Step 602 includes:
  • a decoding operation is performed on the target video stream to obtain the processed target video stream.
  • the processed video stream is a video stream rendered by a cloud server.
  • the target video stream when the rendering mode is the first rendering mode, that is, when the special effects algorithm and special effects rendering processing are performed in the cloud server, can be a video stream rendered by the cloud server. Therefore, after obtaining the target video stream, the virtual reality device can decode the target video stream to obtain the processed target video stream. Therefore, the processed target video stream can be played directly later without additional data processing of the processed target video stream, which effectively reduces the computing power of the virtual reality device.
  • the live video stream rendering method provided in this embodiment can directly play the processed target video stream by decoding the target video stream when the rendering mode is the first rendering mode, without the need for subsequent special effect calculation and special effect rendering processing on the processed target video stream, which effectively saves the computing power of the virtual reality device.
  • Step 602 includes:
  • the target video stream is decoded to obtain a live video stream corresponding to the virtual reality live broadcast, a recognition result corresponding to the live video stream, target data, and rendering time, wherein the recognition result includes key point information corresponding to the target rendering area in a live image frame corresponding to the live video stream and depth data corresponding to the live image frame.
  • the live video stream is rendered according to the recognition result and the target data to obtain a processed target video stream.
  • the rendering mode corresponding to the target special effect is the second rendering mode, that is, the special effect algorithm operation is executed on the cloud server
  • the virtual reality device performs special effect rendering processing
  • the target video stream can be decoded to obtain the live video stream corresponding to the virtual reality live broadcast, the recognition result corresponding to the live video stream, the target data and the rendering time.
  • the recognition result includes key point information corresponding to the target rendering area in the live image frame corresponding to the live video stream and depth data corresponding to the live image frame.
  • the live video stream is rendered according to the recognition result and the target data to obtain a processed target video stream, so that the subsequent virtual reality device can play the target video stream.
  • the virtual reality device can perform special effects rendering processing according to the live video stream, the recognition result corresponding to the live video stream, and the target data within the rendering time.
  • the live video stream rendering method provided in this embodiment calculates the special effect algorithm in the cloud server and performs special effect rendering processing in the virtual reality device when the rendering mode is the second rendering mode, thereby ensuring the real-time nature of the special effect while saving the computing power of the virtual reality device, so that the user can interact with the special effect.
  • the method further includes:
  • the live video stream is rendered according to the recognition result and the target data within the rendering time to obtain a processed target video stream.
  • different target special effects correspond to different rendering times, which are acquired by the cloud server from the configuration platform and encoded in the target video stream in the form of assembly instructions.
  • the rendering time corresponding to the target special effect can be determined.
  • the live video stream is rendered according to the recognition result and the target data within the rendering time to obtain the processed target video stream.
  • the live video stream rendering method provided in this embodiment encodes the rendering time into the target video stream, so that the virtual reality device can accurately render the target special effect according to the rendering time, thereby ensuring the rendering effect of the target special effect.
  • the rendering mode corresponding to the target special effect includes a second rendering mode.
  • the method further includes:
  • the target video stream is decoded to obtain a live video stream corresponding to the virtual reality live broadcast, an identification result corresponding to the live video stream, and target data, wherein the identification result includes key point information corresponding to the target rendering area in the live image frame corresponding to the live video stream and depth data corresponding to the live image frame.
  • the rendering mode corresponding to the target special effect is the second rendering mode, that is, the special effect algorithm operation is executed on the cloud server, and the special effect rendering processing is performed on the virtual reality device, users watching the virtual reality live broadcast can determine whether to render the target special effect according to actual needs.
  • the target video stream can be decoded to obtain the live video stream corresponding to the virtual reality live broadcast, the recognition result corresponding to the live video stream, and the target data, wherein the recognition result includes the key point information corresponding to the target rendering area in the live image frame corresponding to the live video stream and the depth data corresponding to the live image frame.
  • the virtual reality device does not need to render the live video stream according to the recognition result and the target data, and can directly display the live video stream.
  • the live video stream rendering method provided in this embodiment can further save the computing power of the virtual reality device by not rendering the target special effects of the second rendering mode in response to the special effects closing instruction triggered by the user when the rendering mode is the second rendering mode.
  • the display effect of the virtual reality live broadcast can be more in line with the personalized needs of the user, thereby improving the user experience.
  • FIG7 is a schematic diagram of the structure of a live video stream rendering device provided in an embodiment of the present disclosure, which is applied to a cloud server.
  • the device includes: a first acquisition module 71, a determination module 72, a rendering module 73, and a sending module 74.
  • the first acquisition module 71 is used to obtain the live video stream and special effect identification information corresponding to the virtual reality live broadcast
  • the special effect identification information is the identification information of the target special effect used to render the live video stream.
  • the determination module 72 is used to determine the rendering method corresponding to the target special effect according to the special effect identification information.
  • the rendering module 73 is used to render the live video stream using the rendering method corresponding to the target special effect to obtain the processed target video stream.
  • the sending module 74 is used to send the target video stream to a preset virtual reality device for playback.
  • the first acquisition module is used to: acquire the data to be rendered sent by the virtual reality live broadcast server, decode the data to be rendered, and obtain the live video stream and the preset special effect rendering instruction, and parse the special effect rendering instruction to obtain the special effect identification information.
  • the determination module is used to: obtain target data for rendering the target special effect in a preset configuration platform according to the special effect identification information. Determine the rendering method corresponding to the target special effect in the target data. Includes a first rendering mode and/or a second rendering mode.
  • the rendering method includes a first rendering method.
  • the rendering module is used to: determine the target rendering area corresponding to the target special effect according to the target data.
  • the rendering module is used to: perform a rendering operation on multiple live image frames corresponding to the live video stream according to the recognition result and the target data to obtain multiple rendered target image frames, and generate the rendered video stream according to the multiple rendered target image frames.
  • the rendering method includes a second rendering method.
  • the rendering module is used to: determine the target rendering area corresponding to the target special effect according to the target data. Perform a recognition operation on the live video stream to obtain a recognition result corresponding to the live video stream, wherein the recognition result includes key point information corresponding to the target rendering area in the live image frame corresponding to the live video stream and depth data corresponding to the live image frame. Perform an encoding operation on the live video stream, the recognition result and the target data to obtain the target video stream.
  • the rendering module is used to: identify key point information corresponding to the target rendering area in the live image frame corresponding to the live video stream through a preset key point recognition model. Identify depth data corresponding to the live image frame through a preset depth recognition model. Determine the key point information and depth data corresponding to each live image frame as the recognition result corresponding to the live video stream.
  • the device further includes: a determination module, configured to determine the rendering time corresponding to the target special effect in a preset configuration platform according to the special effect identification information.
  • An encoding module configured to encode the rendering time into the target video stream, so that the virtual reality device performs a rendering operation on the live video stream within the rendering time.
  • FIG8 is a schematic diagram of the structure of a live video stream rendering device provided by an embodiment of the present disclosure, which is applied to a virtual reality device.
  • the device includes: a second acquisition module 81, a processing module 82, and and a playback module 83.
  • the second acquisition module 81 is used to acquire the target video stream sent by the cloud server, and the target video stream is obtained by the cloud server rendering the live video stream corresponding to the virtual reality live broadcast and the special effect identification information using the rendering method corresponding to the target special effect.
  • the processing module 82 is used to process the target video stream using the rendering method corresponding to the target special effect to obtain the processed target video stream.
  • the playback module 83 is used to play the processed target video stream.
  • the rendering mode corresponding to the target special effect includes a first rendering mode.
  • the processing module is used to: perform a decoding operation on the target video stream to obtain the processed target video stream.
  • the processed video stream is a video stream rendered by a cloud server.
  • the rendering method corresponding to the target special effect includes a second rendering method.
  • the processing module is used to: decode the target video stream to obtain a live video stream corresponding to the virtual reality live broadcast, a recognition result corresponding to the live video stream, target data and rendering time, wherein the recognition result includes key point information corresponding to the target rendering area in the live image frame corresponding to the live video stream and depth data corresponding to the live image frame.
  • the live video stream is rendered according to the recognition result and the target data to obtain a processed target video stream.
  • the processing module is used to: determine a rendering time corresponding to the target special effect, and perform a rendering operation on the live video stream according to the recognition result and the target data within the rendering time to obtain a processed target video stream.
  • the rendering method corresponding to the target special effect includes a second rendering method.
  • the device also includes: a decoding module, which is used to decode the target video stream if a special effect closing instruction triggered by a user is obtained, and obtain a live video stream corresponding to the virtual reality live broadcast, a recognition result corresponding to the live video stream, and target data, wherein the recognition result includes key point information corresponding to the target rendering area in the live image frame corresponding to the live video stream and depth data corresponding to the live image frame.
  • a playback module is used to play the live video stream corresponding to the virtual reality live broadcast.
  • the device provided in this embodiment can be used to execute the technical solution of the above method embodiment. Its implementation principle and technical effect are similar, and this embodiment will not be repeated here.
  • the present disclosure also provides a cloud rendering system, including a cloud A terminal server and a virtual reality device, wherein the cloud server is in communication connection with the virtual reality device;
  • the cloud server is used to obtain a live video stream and special effect identification information corresponding to the virtual reality live broadcast, and determine a rendering method corresponding to the target special effect according to the special effect identification information, wherein the special effect identification information is identification information of the target special effect used to render the live video stream;
  • the cloud server is further configured to perform special effect algorithm processing and special effect rendering processing according to the live video stream and the special effect identification information to obtain a target video stream when the rendering mode is the first rendering mode;
  • the cloud server is further configured to perform special effect algorithm processing according to the live video stream and the special effect identification information when the rendering mode is the second rendering mode, and send the live video stream and the recognition result to the virtual reality device;
  • the virtual reality device is used to obtain the live video stream sent by the cloud server and the recognition result for special effects rendering.
  • the embodiment of the present disclosure further provides an electronic device, including: a processor and a memory.
  • the memory stores computer executable instructions.
  • the processor executes the computer-executable instructions stored in the memory, so that the processor executes the live video stream rendering method as described in any of the above embodiments.
  • FIG9 is a schematic diagram of the structure of an electronic device provided by an embodiment of the present disclosure.
  • the electronic device 900 may be a terminal device or a server.
  • the terminal device may include but is not limited to mobile terminals such as mobile phones, laptop computers, digital broadcast receivers, personal digital assistants (PDAs), tablet computers (Portable Android Devices, PADs), portable multimedia players (PMPs), vehicle terminals (such as vehicle navigation terminals), etc., and fixed terminals such as digital TVs, desktop computers, etc.
  • PDAs personal digital assistants
  • PADs Portable Android Devices
  • PMPs portable multimedia players
  • vehicle terminals such as vehicle navigation terminals
  • fixed terminals such as digital TVs, desktop computers, etc.
  • the electronic device shown in FIG9 is only an example and should not impose any limitations on the functions and scope of use of the embodiments of the present disclosure.
  • the electronic device 900 may include a processing device (e.g., a central processing unit, a graphics processing unit, etc.) 901, which can perform various appropriate actions and processes according to a program stored in a read-only memory (ROM) 902 or a program loaded from a storage device 908 to a random access memory (RAM) 903.
  • a processing device e.g., a central processing unit, a graphics processing unit, etc.
  • RAM random access memory
  • Various programs and data required for the operation of the electronic device 900 are also stored in the RAM 903.
  • the memory 901, the ROM 902, and the RAM 903 are connected to one another via a bus 904.
  • An input/output (I/O) interface 905 is also connected to the bus 904.
  • the following devices may be connected to the I/O interface 905: input devices 906 including, for example, a touch screen, a touchpad, a keyboard, a mouse, a camera, a microphone, an accelerometer, a gyroscope, etc.; output devices 907 including, for example, a liquid crystal display (LCD), a speaker, a vibrator, etc.; storage devices 908 including, for example, a magnetic tape, a hard disk, etc.; and communication devices 909.
  • the communication device 909 may allow the electronic device 900 to communicate with other devices wirelessly or by wire to exchange data.
  • FIG. 9 shows an electronic device 900 with various devices, it should be understood that it is not required to implement or have all of the devices shown. More or fewer devices may be implemented or have alternatively.
  • an embodiment of the present disclosure includes a computer program product, which includes a computer program carried on a computer-readable medium, and the computer program contains program code for executing the method shown in the flowchart.
  • the computer program can be downloaded and installed from a network through a communication device 909, or installed from a storage device 908, or installed from a ROM 902.
  • the processing device 901 the above-mentioned functions defined in the method of the embodiment of the present disclosure are executed.
  • the computer-readable medium disclosed above may be a computer-readable signal medium or a computer-readable storage medium or any combination of the above two.
  • the computer-readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device or device, or any combination of the above.
  • Computer-readable storage media may include, but are not limited to: an electrical connection with one or more wires, a portable computer disk, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disk read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the above.
  • a computer-readable storage medium may be any tangible medium containing or storing a program that may be used by or in combination with an instruction execution system, device or device.
  • a computer-readable signal medium may include a data signal propagated in a baseband or as part of a carrier wave, in which a computer-readable program code is carried.
  • This propagated data signal may take a variety of forms, including but not limited to an electromagnetic signal, an optical signal, or any suitable combination of the above.
  • the computer readable signal medium may also be any computer readable medium other than a computer readable storage medium.
  • the medium can send, propagate or transmit a program for use by or in conjunction with an instruction execution system, apparatus or device.
  • the program code contained on the computer-readable medium can be transmitted using any suitable medium, including but not limited to: wires, optical cables, RF (radio frequency), etc., or any suitable combination of the above.
  • the embodiments of the present disclosure further provide a computer-readable storage medium, wherein the computer-readable storage medium stores computer-executable instructions.
  • the live video stream rendering method as described in any of the above embodiments is implemented.
  • the embodiments of the present disclosure also provide a computer program product, including a computer program, which, when executed by a processor, implements the method for virtual reality live broadcast rendering as described in any of the above embodiments.
  • the computer-readable medium may be included in the electronic device, or may exist independently without being incorporated into the electronic device.
  • the computer-readable medium carries one or more programs.
  • the electronic device executes the method shown in the above embodiment.
  • Computer program code for performing the operations of the present disclosure may be written in one or more programming languages or a combination thereof, including object-oriented programming languages such as Java, Smalltalk, C++, and conventional procedural programming languages such as "C" or similar programming languages.
  • the program code may be executed entirely on the user's computer, partially on the user's computer, as a separate software package, partially on the user's computer and partially on a remote computer, or entirely on a remote computer or server.
  • the remote computer may be connected to the user's computer through any type of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or may be connected to an external computer (e.g., via the Internet using an Internet service provider).
  • LAN Local Area Network
  • WAN Wide Area Network
  • each box in the flowchart or block diagram may represent a module, a program segment, or a portion of code, which contains one or more executable instructions for implementing the specified logical functions.
  • the functions marked in the boxes may also occur in an order different from that marked in the accompanying drawings. For example, two boxes represented in succession may actually be The above steps are executed substantially in parallel, and they may sometimes be executed in reverse order, depending on the functions involved.
  • each block in the block diagram and/or flow chart, and the combination of blocks in the block diagram and/or flow chart can be implemented by a dedicated hardware-based system that performs the specified functions or operations, or can be implemented by a combination of dedicated hardware and computer instructions.
  • the units involved in the embodiments described in the present disclosure may be implemented by software or hardware.
  • the name of a unit does not limit the unit itself in some cases.
  • the first acquisition unit may also be described as a "unit for acquiring at least two Internet Protocol addresses".
  • exemplary types of hardware logic components include: field programmable gate arrays (FPGAs), application specific integrated circuits (ASICs), application specific standard products (ASSPs), systems on chip (SOCs), complex programmable logic devices (CPLDs), and the like.
  • FPGAs field programmable gate arrays
  • ASICs application specific integrated circuits
  • ASSPs application specific standard products
  • SOCs systems on chip
  • CPLDs complex programmable logic devices
  • a machine-readable medium may be a tangible medium that may contain or store a program for use by or in conjunction with an instruction execution system, device, or equipment.
  • a machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium.
  • a machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or equipment, or any suitable combination of the foregoing.
  • a more specific example of a machine-readable storage medium may include an electrical connection based on one or more lines, a portable computer disk, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disk read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
  • RAM random access memory
  • ROM read-only memory
  • EPROM or flash memory erasable programmable read-only memory
  • CD-ROM portable compact disk read-only memory
  • CD-ROM compact disk read-only memory
  • magnetic storage device or any suitable combination of the foregoing.
  • a live video stream rendering method is provided, which is applied to a cloud rendering system, wherein the cloud rendering system includes a cloud server and a virtual reality device, and the method includes:
  • the cloud server obtains a live video stream and special effect identification information corresponding to the virtual reality live broadcast, where the special effect identification information is identification information of a target special effect used to render the live video stream;
  • the cloud server determines a rendering mode corresponding to the target special effect according to the special effect identification information, wherein the rendering mode includes a first rendering mode and/or a second rendering mode;
  • the cloud server When the rendering mode is the first rendering mode, the cloud server performs special effect algorithm processing according to the live video stream and the special effect identification information, and processes the corresponding The recognition result is processed by special effects rendering to obtain a target video stream, and the target video stream is sent to the virtual reality device, and the virtual reality device obtains the target video stream sent by the cloud server and plays it;
  • the cloud server When the rendering mode is the second rendering mode, the cloud server performs special effect algorithm processing according to the live video stream and the special effect identification information, and sends the live video stream and the recognition result to the virtual reality device.
  • the virtual reality device performs special effect rendering processing according to the live video stream and the algorithm result, obtains the processed target video stream, and plays the processed target video stream.
  • the cloud server performs special effect algorithm processing according to the live video stream and special effect identification information, including:
  • the cloud server acquires target data for rendering the target special effect in a preset configuration platform according to the special effect identification information
  • the cloud server determines a target rendering area corresponding to the target special effect according to the target data
  • the cloud server performs an identification operation on the live video stream to obtain an identification result corresponding to the live video stream, wherein the identification result includes key point information corresponding to the target rendering area in a live image frame corresponding to the live video stream and depth data corresponding to the live image frame.
  • performing special effects rendering processing on the recognition result processed by the special effects algorithm to obtain a target video stream includes:
  • the cloud server performs a rendering operation on the live video stream according to the recognition result and the target data to obtain a rendered video stream;
  • the cloud server performs encoding operation on the rendered video stream to obtain the target video stream.
  • the cloud server determines the rendering mode corresponding to the target special effect according to the special effect identification information, it further includes:
  • the cloud server acquires target data for rendering the target special effect in a preset configuration platform according to the special effect identification information
  • the cloud server determines a target rendering area corresponding to the target special effect according to the target data
  • the cloud server performs an identification operation on the live video stream to obtain an identification result corresponding to the live video stream, wherein the identification result includes key point information corresponding to the target rendering area in a live image frame corresponding to the live video stream and depth data corresponding to the live image frame;
  • the cloud server encodes the live video stream, the recognition result and the target data to obtain the target video stream, and sends the target video stream to the virtual reality device.
  • the method after sending the target video stream to the virtual reality device, the method further includes:
  • the virtual reality device acquires the target video stream, performs a decoding operation on the target video stream, and obtains a live video stream corresponding to the virtual reality live broadcast, a recognition result corresponding to the live video stream, and target data, wherein the recognition result includes key point information corresponding to the target rendering area in a live image frame corresponding to the live video stream and depth data corresponding to the live image frame;
  • the live video stream is rendered according to the recognition result and the target data to obtain a processed target video stream.
  • a live video stream rendering method is provided, which is applied to a cloud server, comprising:
  • the special effect identification information is identification information of a target special effect used to render the live video stream
  • the target video stream is sent to a preset virtual reality device for playback.
  • obtaining the live video stream and special effect identification information corresponding to the virtual reality live broadcast includes:
  • the special effect rendering instruction is parsed to obtain the special effect identification information.
  • the determining of the special effect identification information according to the special effect identification information include:
  • the rendering mode includes a first rendering mode and/or a second rendering mode.
  • the rendering mode includes a first rendering mode
  • An encoding operation is performed on the rendered video stream to obtain the target video stream.
  • the rendering operation is performed on the live video stream according to the recognition result and the target data to obtain the target video stream, including:
  • the rendered video stream is generated according to the multiple rendered target image frames.
  • the rendering mode includes a second rendering mode
  • An encoding operation is performed on the live video stream, the recognition result, and the target data to obtain the target video stream.
  • the live video stream is identified. Obtaining the recognition result corresponding to the live video stream includes:
  • the key point information and depth data corresponding to each live image frame are determined as the recognition result corresponding to the live video stream.
  • the method after rendering the live video stream in a rendering manner corresponding to the target special effect to obtain the processed target video stream, the method further includes:
  • the rendering time is encoded into the target video stream, so that the virtual reality device performs a rendering operation on the live video stream within the rendering time.
  • a live video stream rendering method is provided, which is applied to a virtual reality device, including:
  • Target video stream sent by the cloud server, wherein the target video stream is obtained after the cloud server performs rendering processing according to the live video stream corresponding to the virtual reality live broadcast and the special effect identification information, using a rendering method corresponding to the target special effect;
  • the rendering mode corresponding to the target special effect includes a first rendering mode
  • the processed video stream is a video stream rendered by a cloud server.
  • the rendering mode corresponding to the target special effect includes a second rendering mode
  • the target video stream is processed by using a rendering method corresponding to the target special effect, including include:
  • the live video stream is rendered according to the recognition result and the target data to obtain a processed target video stream.
  • the method further includes:
  • the live video stream is rendered according to the recognition result and the target data within the rendering time to obtain a processed target video stream.
  • the rendering mode corresponding to the target special effect includes a second rendering mode
  • the method further includes:
  • a decoding operation is performed on the target video stream to obtain a live video stream corresponding to the virtual reality live broadcast, a recognition result corresponding to the live video stream, and target data, wherein the recognition result includes key point information corresponding to the target rendering area in a live image frame corresponding to the live video stream and depth data corresponding to the live image frame;
  • a live video stream rendering device is provided, which is applied to a cloud server, including:
  • a first acquisition module is used to acquire a live video stream and special effect identification information corresponding to the virtual reality live broadcast, wherein the special effect identification information is identification information of a target special effect used to render the live video stream;
  • a determination module used to determine a rendering mode corresponding to the target special effect according to the special effect identification information
  • a rendering module used to render the live video stream in a rendering mode corresponding to the target special effect to obtain a processed target video stream
  • the sending module is used to send the target video stream to a preset virtual reality device for playback.
  • the first acquisition module is used to:
  • the special effect rendering instruction is parsed to obtain the special effect identification information.
  • the determining module is used to:
  • the rendering mode includes a first rendering mode and/or a second rendering mode.
  • the rendering mode includes a first rendering mode
  • the rendering module is used to:
  • An encoding operation is performed on the rendered video stream to obtain the target video stream.
  • the rendering module is used to:
  • the rendered video stream is generated according to the multiple rendered target image frames.
  • the rendering mode includes a second rendering mode
  • the rendering module is used to:
  • Encoding the live video stream, the recognition result and the target data is performed to obtain Get the target video stream.
  • the rendering module is used to:
  • the key point information and depth data corresponding to each live image frame are determined as the recognition result corresponding to the live video stream.
  • the device further includes:
  • a determination module used to determine the rendering time corresponding to the target special effect in a preset configuration platform according to the special effect identification information
  • the encoding module is used to encode the rendering time into the target video stream so that the virtual reality device can render the live video stream within the rendering time.
  • a live video stream rendering device is provided, which is applied to a virtual reality device, including:
  • a second acquisition module is used to acquire a target video stream sent by the cloud server, wherein the target video stream is obtained after the cloud server performs rendering processing according to the live video stream corresponding to the virtual reality live broadcast and the special effect identification information, using a rendering method corresponding to the target special effect;
  • a processing module used for processing the target video stream by adopting a rendering method corresponding to the target special effect to obtain a processed target video stream
  • the playing module is used to play the processed target video stream.
  • the rendering mode corresponding to the target special effect includes a first rendering mode
  • the processing module is used to:
  • the processed video stream is a video stream rendered by a cloud server.
  • the rendering mode corresponding to the target special effect includes a second rendering mode
  • the processing module is used to:
  • the target video stream is decoded to obtain a live video stream corresponding to the virtual reality live broadcast, a recognition result corresponding to the live video stream, target data, and rendering time, wherein the recognition result
  • the result includes key point information corresponding to the target rendering area in the live image frame corresponding to the live video stream and depth data corresponding to the live image frame;
  • the live video stream is rendered according to the recognition result and the target data to obtain a processed target video stream.
  • the processing module is used to:
  • the live video stream is rendered according to the recognition result and the target data within the rendering time to obtain a processed target video stream.
  • the rendering mode corresponding to the target special effect includes a second rendering mode
  • the device also includes:
  • a decoding module for decoding the target video stream if a special effect closing instruction triggered by a user is obtained, to obtain a live video stream corresponding to the virtual reality live broadcast, a recognition result corresponding to the live video stream, and target data, wherein the recognition result includes key point information corresponding to the target rendering area in a live image frame corresponding to the live video stream and depth data corresponding to the live image frame;
  • the playback module is used to play the live video stream corresponding to the virtual reality live broadcast.
  • a cloud rendering system comprising a cloud server and a virtual reality device, wherein the cloud server is communicatively connected to the virtual reality device;
  • the cloud server is used to obtain a live video stream and special effect identification information corresponding to the virtual reality live broadcast, and determine a rendering method corresponding to the target special effect according to the special effect identification information, wherein the special effect identification information is identification information of the target special effect used to render the live video stream;
  • the cloud server is further configured to perform special effect algorithm processing and special effect rendering processing according to the live video stream and the special effect identification information to obtain a target video stream when the rendering mode is the first rendering mode;
  • the cloud server is further configured to perform special effect algorithm processing according to the live video stream and the special effect identification information when the rendering mode is the second rendering mode, and send the live video stream and the recognition result to the virtual reality device;
  • the virtual reality device is used to obtain the live video stream sent by the cloud server and identify The result is special effects rendering.
  • an electronic device comprising: at least one processor and a memory;
  • the memory stores computer-executable instructions
  • the at least one processor executes the computer-executable instructions stored in the memory, so that the at least one processor executes the live video stream rendering method described in the first aspect and various possible designs of the first aspect or the second aspect and various possible designs of the second aspect.
  • a computer-readable storage medium stores computer execution instructions.
  • the live video stream rendering method described in the first aspect and various possible designs of the first aspect or the second aspect and various possible designs of the second aspect is implemented.
  • a computer program product including a computer program, which, when executed by a processor, implements the live video stream rendering method described in the first aspect and various possible designs of the first aspect or the second aspect and various possible designs of the second aspect.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Processing Or Creating Images (AREA)

Abstract

本公开实施例提供一种直播视频流渲染方法、装置、设备、存储介质及产品,该方法包括:云端服务器获取虚拟现实直播对应的直播视频流以及特效标识信息,并根据特效标识信息确定目标特效对应的渲染方式;当渲染方式为第一渲染方式时,云端服务器根据直播视频流以及特效标识信息进行特效算法处理以及特效渲染处理,获得目标视频流;当渲染方式为第二渲染方式时,云端服务器进行特效算法处理,虚拟现实设备根据直播视频流以及算法结果进行特效渲染处理,获得处理后的目标视频流,播放渲染后的视频流。从而能够节约虚拟现实设备的算力。通过针对不同类型的目标特效,采用不同的渲染处理方式,从而能够保证特效渲染的准确性以及实时性。

Description

直播视频流渲染方法、装置、设备、存储介质及产品
本申请要求2022年11月17日递交的、标题为“直播视频流渲染方法、装置、设备、存储介质及产品”、申请号为202211441161.5的中国发明专利申请的优先权,该申请的全部内容通过引用结合在本申请中。
技术领域
本公开实施例涉及图像处理技术领域,尤其涉及一种直播视频流渲染方法、装置、设备、存储介质及产品。
背景技术
随着科技的发展,虚拟现实(Virtual Reality,简称VR)技术逐渐走进用户的生活中。用户可以通过VR技术实现3D的VR直播。在VR直播过程中,通常会存在特效渲染的需求。
现有的直播场景一般都是针对手机端的普通直播。由于手机端直播的分辨率较低,因此,可以在手机端安装用于特效渲染的算法,在手机端实现算法识别以及渲染操作。
但是,对于VR直播,由于需要临场感,因此对于清晰度的要求非常高,直播画面一般为超高分辨率画面,采用上述方式进行特效渲染则会影响硬件性能,且无法保证直播的实时性。
发明内容
本公开实施例提供一种直播视频流渲染方法、装置、设备、存储介质及产品,用于解决现有的渲染方法无法保证渲染操作的实时性以及渲染结果的清晰度的技术问题。
第一方面,本公开实施例提供一种直播视频流渲染方法,应用于云端渲染系统,所述云端渲染系统包括云端服务器以及虚拟现实设备,所述方法包括:
云端服务器获取虚拟现实直播对应的直播视频流以及特效标识信息,所述特效标识信息为用于渲染所述直播视频流的目标特效的标识信息;
云端服务器根据所述特效标识信息确定所述目标特效对应的渲染方式,其中,所述渲染方式包括第一渲染方式和/或第二渲染方式;
当所述渲染方式为第一渲染方式时,所述云端服务器根据所述直播视频流以及特效标识信息进行特效算法处理,以及根据所述特效算法处理对应的识别结果进行特效渲染处理,获得目标视频流,并将所述目标视频流发送至所述虚拟现实设备,所述虚拟现实设备获取所述云端服务器发送的目标视频流并播放;
当所述渲染方式为第二渲染方式时,所述云端服务器根据所述直播视频流以及特效标识信息进行特效算法处理,并将所述直播视频流以及识别结果发送至所述虚拟现实设备,所述虚拟现实设备根据所述直播视频流以及算法结果进行特效渲染处理,获得处理后的目标视频流,播放所述处理后的目标视频流。
第二方面,本公开实施例提供一种直播视频流渲染方法,应用于云端服务器,包括:
获取虚拟现实直播对应的直播视频流以及特效标识信息,所述特效标识信息为用于渲染所述直播视频流的目标特效的标识信息;
根据所述特效标识信息确定所述目标特效对应的渲染方式;
采用与所述目标特效对应的渲染方式对所述直播视频流进行渲染处理,获得处理后的目标视频流;
将所述目标视频流发送至预设的虚拟现实设备进行播放。
第三方面,本公开实施例提供一种直播视频流渲染方法,应用于虚拟现实设备,包括:
获取所述云端服务器发送的目标视频流,所述目标视频流为所述云端服务器根据虚拟现实直播对应的直播视频流以及特效标识信息,采用与所述目标特效对应的渲染方式进行渲染处理后获得的;
采用所述目标特效对应的渲染方式对所述目标视频流进行处理,获得处理后的目标视频流;
播放所述处理后的目标视频流。
第四方面,本公开实施例提供一种直播视频流渲染装置,应用于云端服务器,第一获取模块,用于获取虚拟现实直播对应的直播视频流以及特效标识信息,所述特效标识信息为用于渲染所述直播视频流的目标特效的标识信息;
确定模块,用于根据所述特效标识信息确定所述目标特效对应的渲染方式;
渲染模块,用于采用与所述目标特效对应的渲染方式对所述直播视频流进行渲染处理,获得处理后的目标视频流;
发送模块,用于将所述目标视频流发送至预设的虚拟现实设备进行播放。
第五方面,本公开实施例提供一种直播视频流渲染装置,应用于虚拟现实设备,包括:
第二获取模块,用于获取所述云端服务器发送的目标视频流,所述目标视频流为所述云端服务器根据虚拟现实直播对应的直播视频流以及特效标识信息,采用与所述目标特效对应的渲染方式进行渲染处理后获得的;
处理模块,用于采用所述目标特效对应的渲染方式对所述目标视频流进行处理,获得处理后的目标视频流;
播放模块,用于播放所述处理后的目标视频流。
第六方面,本公开实施例提供一种云渲染系统,包括云端服务器以及虚拟现实设备,所述云端服务器与所述虚拟现实设备通信连接;
其中,所述云端服务器用于获取虚拟现实直播对应的直播视频流以及特效标识信息,并根据所述特效标识信息确定所述目标特效对应的渲染方式,所述特效标识信息为用于渲染所述直播视频流的目标特效的标识信息;
所述云端服务器还用于在所述渲染方式为第一渲染方式时,根据所述直播视频流以及特效标识信息进行特效算法处理以及特效渲染处理,获得目标视频流;
所述云端服务器还用于在所述渲染方式为第二渲染方式时,根据所述直播视频流以及特效标识信息进行特效算法处理,并将所述直播视频流以及识 别结果发送至所述虚拟现实设备;
所述虚拟现实设备用于获取所述云端服务器发送的直播视频流以及识别结果进行特效渲染处理。
第七方面,本公开实施例提供一种电子设备,包括:处理器和存储器;
所述存储器存储计算机执行指令;
所述处理器执行所述存储器存储的计算机执行指令,使得所述至少一个处理器执行如上第一方面以及第一方面各种可能的设计或第二方面以及第二方面各种可能的设计或第三方面以及第三方面各种可能的设计所述的直播视频流渲染方法。
第八方面,本公开实施例提供一种计算机可读存储介质,所述计算机可读存储介质中存储有计算机执行指令,当处理器执行所述计算机执行指令时,实现如上第一方面以及第一方面各种可能的设计或第二方面以及第二方面各种可能的设计或第三方面以及第三方面各种可能的设计所述的直播视频流渲染方法。
第九方面,本公开实施例提供一种计算机程序产品,包括计算机程序,所述计算机程序被处理器执行时实现如上第一方面以及第一方面各种可能的设计或第二方面以及第二方面各种可能的设计或第三方面以及第三方面各种可能的设计所述的直播视频流渲染方法。
本实施例提供的直播视频流渲染方法、装置、设备、存储介质及产品,通过在获取到虚拟现实直播对应的直播视频流以及特效标识信息之后,在云端服务器中实现对虚拟现实直播对应的直播视频流的算法识别处理和/或渲染处理,从而能够避免对看播端虚拟现实设备造成压力。此外,通过针对不同类型的目标特效,采用不同的渲染处理方式,从而能够有针对性地对目标特效进行渲染操作,提高了特效渲染的准确性以及实时性。
附图说明
为了更清楚地说明本公开实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作一简单地介绍,显而易见地,下面描述中的附图是本公开的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。
图1为本公开所基于的系统架构图;
图2为本公开实施例提供的直播视频流渲染方法的流程示意图;
图3为本公开又一实施例提供的直播视频流渲染方法的流程示意图;
图4为本公开又一实施例提供的直播视频流渲染方法的流程示意图;
图5为本公开又一实施例提供的直播视频流渲染方法的流程示意图;
图6为本公开实施例提供的直播视频流渲染方法的流程示意图;
图7为本公开实施例提供的直播视频流渲染装置的结构示意图;
图8为本公开实施例提供的直播视频流渲染装置的结构示意图;
图9为本公开实施例提供的电子设备的结构示意图。
具体实施方式
为使本公开实施例的目的、技术方案和优点更加清楚,下面将结合本公开实施例中的附图,对本公开实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本公开一部分实施例,而不是全部的实施例。基于本公开中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本公开保护的范围。
为了解决现有的渲染方法无法保证渲染操作的实时性以及渲染结果的清晰度的技术问题,本公开提供了一种直播视频流渲染方法、装置、设备、存储介质及产品。
需要说明的是,本公开提供的直播视频流渲染方法、装置、设备、存储介质及产品可应用在任意一种虚拟现实直播渲染的场景中,以保证虚拟现实直播渲染的实时性以及直播内容的清晰度。
对于普通直播,如运行在手机端(后续称为Native端)的直播,特效处理操作往往不需要太高分辨率的视频画面输入,在Native端运行识别算法以及渲染操作即可以达到性能预期。但是对于VR直播,由于需要临场感,因此对于清晰度的要求非常高,这对于算法处理方面提出了新的挑战。
在解决上述技术问题的过程中,发明人通过研究发现,为了解决VR直播画面高画质的问题,因为Native端的算力有限,特效算法的计算不能放在Native端来做。因此可以将特效算法的计算放在云端服务器执行,即可以充分利用云端服务器的算力,也可以省掉原本这一块在Native端的算力,且能 够保证渲染结果的清晰度以及实时性。
图1为本公开所基于的系统架构图,如图1所示,本公开所基于的系统架构至少包括服务器11、直播伴侣服务器12、虚拟现实设备13。其中,服务器11中设置有直播视频流渲染装置,该直播视频流渲染装置可采用C/C++、Java、Shell或Python等语言编写;虚拟现实设备13则可以为VR眼镜等能够实现直播内容观看的VR设备。
基于上述系统架构,服务器11可以获取直播伴侣服务器12发送的虚拟现实直播对应的直播视频流以及特效标识信息。基于该直播视频流以及特效标识信息采用与目标特效对应的渲染方式对所述直播视频流进行渲染处理,获得处理后的目标视频流,将所述目标视频流发送至预设的虚拟现实设备13进行播放。
为了能够在降低虚拟现实设备端算力的基础上,实现对直播视频流的渲染操作,本公开还提供了一种直播视频流渲染方法,应用于云端渲染系统,所述云端渲染系统包括云端服务器以及虚拟现实设备,所述方法包括:
云端服务器获取虚拟现实直播对应的直播视频流以及特效标识信息,所述特效标识信息为用于渲染所述直播视频流的目标特效的标识信息;
云端服务器根据所述特效标识信息确定所述目标特效对应的渲染方式,其中,所述渲染方式包括第一渲染方式和/或第二渲染方式;
当所述渲染方式为第一渲染方式时,所述云端服务器根据所述直播视频流以及特效标识信息进行特效算法处理,以及根据所述特效算法处理对应的识别结果进行特效渲染处理,获得目标视频流,并将所述目标视频流发送至所述虚拟现实设备,所述虚拟现实设备获取所述云端服务器发送的目标视频流并播放;
当所述渲染方式为第二渲染方式时,所述云端服务器根据所述直播视频流以及特效标识信息进行特效算法处理,并将所述直播视频流以及识别结果发送至所述虚拟现实设备,所述虚拟现实设备根据所述直播视频流以及算法结果进行特效渲染处理,获得处理后的目标视频流,播放所述处理后的目标视频流。
本实施例提供了一种云渲染系统,在该云渲染系统中至少包括云端服务器以及虚拟现实设备。该云端服务器能够分别获取虚拟现实直播对应的直播 视频流以及用于渲染该直播视频流的目标特效的特效标识信息。
实际应用中,由于不同的目标特效可能对应有不同的显示效果、渲染区域以及对实时度的不同要求等,为了保证直播视频流的渲染效果,针对不同的目标特效,可以预先设置不同的渲染方式。因此,云端服务器在获取到直播视频流以及特效标识信息之后,可以根据该特效标识信息确定该目标特效对应的渲染方式。其中,渲染方式包括第一渲染方式和/或第二渲染方式。
当渲染方式为第一渲染方式时,可以在云端服务器中实现特效算法以及特效渲染。得到渲染后的目标视频流,并将目标视频流发送至虚拟现实设备。因此虚拟现实设备在获取到该目标视频流之后,无需进行进一步地渲染处理,可以直接播放该目标视频流。
当渲染方式为第二渲染方式时,可以在云端服务器进行特效算法,将特效算法对应的识别结果与直播视频流一同发送至虚拟现实设备中。相应地,虚拟现实设备在获取到该识别结果与直播视频流之后,可以基于识别结果与直播视频流进行特效渲染处理。从而能够保证目标特效的实时性,进而能够支持主播与目标特效进行交互。
需要说明的是,当渲染方式仅包括第一渲染方式时,则针对全部的目标特效,均可以在云端服务器实现特效算法处理以及特效渲染处理。从而能够有效地节约虚拟现实设备的算力,且能够保证虚拟现实直播的效果。
当渲染方式仅包括第二渲染方式时,则针对全部的目标特效,可以在云端服务器进行特效算法处理,将特效算法处理对应的识别结果发送至虚拟现实设备。虚拟现实设备可以根据该识别结果对直播视频流进行特效渲染处理。从而能够保证特效渲染的实时性,进而主播能够与该目标特效进行交互操作,进一步地优化了直播效果。
而当渲染方式包括第一渲染方式以及第二渲染方式时,可以针对不同的目标特效选择不同的渲染方式进行渲染操作,能够有针对性地实现特效渲染操作。在提高了渲染效果的基础上,保证了特效渲染操作的实时性。
进一步地,在上述任一实施例的基础上,所述云端服务器根据所述直播视频流以及特效标识信息进行特效算法处理,包括:
所述云端服务器根据所述特效标识信息在预设的配置平台中获取用于渲染所述目标特效的目标数据;
所述云端服务器根据所述目标数据确定所述目标特效对应的目标渲染区域;
所述云端服务器对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
在本实施例中,可以预先设置有配置平台,云端服务器还可以与该配置平台通信连接。该配置平台中可以存储有多个特效对应的目标数据,其中,该目标数据包括但不限于特效对应的渲染方式、特效对应的渲染时间、特效对应的渲染位置、装饰物等。
因此,云端服务器在获取到虚拟现实直播对应的直播视频流以及特效标识信息之后,可以根据该特效标识信息在预设的配置平台中获取目标特效对应的目标数据。
在获取到目标数据之后,可以根据该目标数据确定该目标特效对应的目标渲染区域以及深度等信息。进一步地,可以对直播视频流进行识别操作,获得直播视频流对应的识别结果,其中,识别结果中包括直播视频流对应的直播图像帧中与目标渲染区域对应的关键点信息以及直播图像帧对应的深度数据。
进一步地,在上述任一实施例的基础上,所述根据所述特效算法处理对应的识别结果进行特效渲染处理,获得目标视频流,包括:
所述云端服务器根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得渲染后的视频流;
所述云端服务器对所述渲染后的视频流进行编码操作,获得所述目标视频流。
在本实施例中,在确定直播视频流对应的识别结果之后,即可以根据该识别结果中的关键点信息以及深度数据准确地对目标数据中的装饰物进行渲染操作,得到渲染后的视频流。进一步地,为了便于数据的传输,可以对该渲染后的视频流进行编码操作,得到该目标视频流。
通过在渲染方式为第一渲染方式时,在云端服务器中实现对直播视频流的算法识别操作以及渲染操作,能够获得目标视频流。将目标视频流发送至 虚拟现实设备之后,虚拟现实设备无需在本地进行特效算法以及特效渲染处理,有效地节约了虚拟现实设备的算力。
进一步地,在上述任一实施例的基础上,所述云端服务器根据所述特效标识信息确定所述目标特效对应的渲染方式之后,还包括:
所述云端服务器根据所述特效标识信息在预设的配置平台中获取用于渲染所述目标特效的目标数据;
所述云端服务器根据所述目标数据确定所述目标特效对应的目标渲染区域;
所述云端服务器对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
所述云端服务器对所述直播视频流、所述识别结果以及所述目标数据进行编码操作,获得所述目标视频流,将所述目标视频流发送至所述虚拟现实设备;
所述虚拟现实设备获取所述目标视频流,对所述目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、所述直播视频流对应的识别结果、目标数据,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得处理后的目标视频流。
在本实施例中,当目标特效对应的渲染方式为第二渲染方式时,也即在云端服务器执行特效算法操作,在虚拟现实设备进行特效渲染处理。
云端服务器在获取到目标数据之后,可以根据该目标数据确定该目标特效对应的目标渲染区域以及深度等信息。进一步地,可以对直播视频流进行识别操作,获得直播视频流对应的识别结果,其中,识别结果中包括直播视频流对应的直播图像帧中与目标渲染区域对应的关键点信息以及直播图像帧对应的深度数据。
云端服务器在确定直播视频流对应的识别结果之后,可以对直播视频流、识别结果以及目标数据进行编码操作,获得目标视频流。将该目标视频流发 送至虚拟现实设备。
进一步地,在上述任一实施例的基础上,所述将所述目标视频流发送至所述虚拟现实设备之后,还包括:
相应地,虚拟现实设备可以获取云端服务器发送的目标视频流。虚拟现实设备在获取到目标视频流之后,可以对目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、直播视频流对应的识别结果、目标数据以及渲染时间,其中,识别结果中包括直播视频流对应的直播图像帧中与目标渲染区域对应的关键点信息以及直播图像帧对应的深度数据。
根据该识别结果以及目标数据对直播视频流进行渲染操作,获得处理后的目标视频流。从而后续虚拟现实设备可以播放该目标视频流。
通过在渲染方式为第二渲染方式时,在云端服务器进行特效算法的计算,在虚拟现实设备中进行特效渲染处理,从而能够在节约虚拟现实设备算力的基础上,保证了特效的实时性。使得用户能够与该特效进行交互操作。
图2为本公开实施例提供的直播视频流渲染方法的流程示意图,应用于云端服务器,如图2所示,该方法包括:
步骤201、获取虚拟现实直播对应的直播视频流以及特效标识信息,所述特效标识信息为用于渲染所述直播视频流的目标特效的标识信息。
本实施例的执行主体为直播视频流渲染装置,该直播视频流渲染装置可耦合于云端服务器中。云端服务器分别与虚拟现实设备以及虚拟现实直播服务器通信连接。
在本实施方式中,主播可以通过预设的直播应用发起虚拟现实直播。该直播应用可安装与该虚拟现实直播服务器中。当主播发起虚拟现实直播之后,可以通过预设的双目图像采集装置实现直播视频流的采集。可选地,主播在直播过程中,为了优化直播效果,可以选择目标特效进行显示。或者,该虚拟现实直播的观众也可以根据实际需求触发虚拟资源转移操作,该虚拟资源转移操作可以与目标特效一一对应。
因此,该虚拟现实直播服务器可以分别获取直播视频流以及特效标识信息,并将直播视频流以及特效标识信息发送给云端服务器。相应地,云端服务器可以获取该直播视频流以及特效标识信息。
步骤202、根据所述特效标识信息确定所述目标特效对应的渲染方式。
在本实施方式中,不同的目标特效可能对应有不同的显示效果、渲染区域等。举例来说,目标特效可以为虚拟眼镜配件的特效,该特效作用于主播的脸上,可以与主播进行实时交互。当触发该目标特效之后,可以在主播的脸上渲染出一副能够基于主播的移动进行相应交互操作的眼镜。该目标特效对时效的要求较为严格。
仍旧举例来说,目标特效可以为海底世界的特效,该特效作用于整个直播间。当触发该目标特效之后,可以对主播所在的背景进行渲染操作。该目标特效则无需与主播进行交互,因此不存在对时效的要求。
针对上述不同的目标特效,可以对应有不同的渲染方式。例如,针对对时效的要求较为严格的目标特效,则可以在云端服务器实现特效计算,得到针对直播视频流的识别结果。在虚拟现实设备中进行渲染操作,以保证与用户的实时交互。而针对不存在对时效的要求的目标特效,则可以在云端服务器实现特效计算以及特效渲染,虚拟现实设备可以直接播放云端服务器发送的目标视频流。
因此,为了能够准确地实现对目标特效的渲染操作,在获取到直播视频流以及特效标识信息之后,可以根据该特效标识信息确定所述目标特效对应的渲染方式。
步骤203、采用与所述目标特效对应的渲染方式对所述直播视频流进行渲染处理,获得处理后的目标视频流。
在本实施方式中,在确定目标特效对应的渲染方式之后,云端服务器即可以采用与该目标特效对应的渲染方式对直播视频流进行渲染处理,得到处理后的目标视频流。其中,该渲染处理包括特效算法处理以及特效渲染处理中的至少一项。
步骤204、将所述目标视频流发送至预设的虚拟现实设备进行播放。
在本实施方式中,在得到目标视频流之后,可以将目标视频流发送至预设的虚拟现实设备进行播放。通过在云端服务器进行特效算法处理和/或特效渲染处理,从而能够有效地节约虚拟现实设备的算力,减少了虚拟现实设备的压力。
本实施例提供的直播视频流渲染方法,通过在获取到虚拟现实直播对应的直播视频流以及特效标识信息之后,在云端服务器中实现对虚拟现实直播 对应的直播视频流的算法识别处理和/或渲染处理,从而能够避免对看播端虚拟现实设备造成压力。此外,通过针对不同类型的目标特效,采用不同的渲染处理方式,从而能够有针对性地对目标特效进行渲染操作,提高了特效渲染的准确性以及实时性。
进一步地,在上述任一实施例的基础上,步骤202包括:
根据所述特效标识信息在预设的配置平台中获取所述目标特效对应的目标数据。
在所述目标数据中确定所述目标特效对应的渲染方式。
其中,所述渲染方式包括第一渲染方式和/或第二渲染方式。
在本实施例中,可以预先设置有配置平台。该配置平台中可以存储有多个特效对应的目标数据,其中,该目标数据包括但不限于特效对应的渲染方式、特效对应的渲染时间、特效对应的渲染位置、装饰物等。
因此,在获取到虚拟现实直播对应的直播视频流以及特效标识信息之后,可以根据该特效标识信息在预设的配置平台中获取目标特效对应的目标数据。在该目标数据中确定目标特效对应的渲染方式。
其中,渲染方式包括第一渲染方式和/或第二渲染方式。该第一渲染方式可以为云端渲染方式,也即特效算法以及特效渲染均在云端服务器中进行。第二渲染方式则可以为本地渲染方式,也即特效算法在云端服务器中进行,特效渲染处理在虚拟现实设备中进行。
本实施例提供的直播视频流渲染方法,通过根据特效标识信息在预设的配置平台中获取目标特效对应的目标数据,从而能够确定目标特效对应的渲染方式。进而能够准确地根据该目标特效对应的渲染方式实现对目标特效的渲染操作,在节约虚拟现实设备算力的基础上,提高了渲染操作的准确性。
进一步地,在上述任一实施例的基础上,步骤203之后,还包括:
根据所述特效标识信息在预设的配置平台中确定所述目标特效对应的渲染时间。
将所述渲染时间编码至所述目标视频流中,以使虚拟现实设备在所述渲染时间内对所述直播视频流进行渲染操作。
在本实施例中,不同的目标特效对应的渲染时间有所不同,该渲染时间可以为目标特效的开发者确定并存储在配置平台中的。举例来说,目标特效 A需要在用户触发虚拟资源转移操作后的第三秒进行渲染,而目标特效B则需要在用户触发虚拟资源转移操作后立刻进行渲染操作。
因此,为了实现目标特效的准确渲染操作,可以根据特效标识信息在预设的配置平台中确定目标特效对应的渲染时间。将渲染时间编码至目标视频流中,以使虚拟现实设备在渲染时间内对直播视频流进行渲染操作。
本实施例提供的直播视频流渲染方法,通过将渲染时间编码至所述目标视频流中,从而虚拟现实设备能够根据该渲染时间精准地对目标特效进行渲染操作,保证了目标特效的渲染效果。
图3为本公开又一实施例提供的直播视频流渲染方法的流程示意图,在上述任一实施例的基础上,如图3所示,步骤201包括:
步骤301、获取虚拟现实直播服务器发送的待渲染数据。
步骤302、对所述待渲染数据进行解码操作,获得所述直播视频流以及预设的特效渲染指令。
步骤303、对所述特效渲染指令进行解析操作,获得所述特效标识信息。
在本实施例中,虚拟现实直播服务器可以分别获取直播视频流以及特效标识信息,将特效标识信息以汇编指令的方式与直播视频流编码在一起,得到待渲染数据,并将待渲染数据发送至云端服务器。
而通过将汇编指令与直播视频流编码在一起,从而能够在传输直播视频流的过程中携带用于控制特效渲染的相应信息,使得特效渲染操作更加贴合用户的个性化需求,提升用户体验。
相应地,云端服务器在获取到虚拟现实直播服务器发送的待渲染数据之后,可以对该待渲染数据进行解码操作,获得直播视频流以及预设的特效渲染指令。进一步地对特效渲染指令进行解析操作,获得特效标识信息。
本实施例提供的直播视频流渲染方法,通过在获取虚拟现实直播服务器发送的待渲染数据,对待渲染数据进行解码操作,从而能够顾得到直播视频流。此外,通过将特效标识信息携带在特效渲染指令中,从而能够基于用户触发的目标特效准确地进行渲染操作。
图4为本公开又一实施例提供的直播视频流渲染方法的流程示意图,在上述任一实施例的基础上,所述渲染方式包括第一渲染方式。如图4所示,步骤203包括:
步骤401、根据所述目标数据确定所述目标特效对应的目标渲染区域。
步骤402、对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据。
步骤403、根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得渲染后的视频流。
步骤404、对所述渲染后的视频流进行编码操作,获得所述目标视频流。
在本实施例中,在渲染方式为第一渲染方式时,需要在云端服务器中进行特效算法以及特效渲染处理。
在获取到目标数据之后,可以根据该目标数据确定该目标特效对应的目标渲染区域以及深度等信息。进一步地,可以对直播视频流进行识别操作,获得直播视频流对应的识别结果,其中,识别结果中包括直播视频流对应的直播图像帧中与目标渲染区域对应的关键点信息以及直播图像帧对应的深度数据。
在确定直播视频流对应的识别结果之后,即可以根据该识别结果中的关键点信息以及深度数据准确地对目标数据中的装饰物进行渲染操作,得到渲染后的视频流。进一步地,为了便于数据的传输,可以对该渲染后的视频流进行编码操作,得到该目标视频流。
进一步地,在上述任一实施例的基础上,步骤403包括:
根据所述识别结果以及所述目标数据对所述直播视频流对应的多个直播图像帧进行渲染操作,获得多张渲染后的目标图像帧。
根据所述多张渲染后的目标图像帧生成所述渲染后的视频流。
在本实施例中,可以确定每一直播图像帧对应的识别结果,根据该识别结果以及目标数据对该直播图像帧进行渲染操作。针对每一直播图像帧,重复上述操作,得到多张渲染后的目标图像帧。进而能够根据该多张渲染后的目标图像帧生成渲染后的视频流。
本实施例提供的直播视频流渲染方法,通过在渲染方式为第一渲染方式时,在云端服务器中实现对直播视频流的算法识别操作以及渲染操作,能够获得目标视频流。将目标视频流发送至虚拟现实设备之后,虚拟现实设备无需在本地进行特效算法以及特效渲染处理,有效地节约了虚拟现实设备的算 力。且在云端服务器进行特效算法以及特效渲染处理能够保证特效渲染后的目标视频流的渲染效果提升用户体验。
图5为本公开又一实施例提供的直播视频流渲染方法的流程示意图,在上述任一实施例的基础上,所述渲染方式包括第二渲染方式。如图5所示,步骤203包括:
步骤501、根据所述目标数据确定所述目标特效对应的目标渲染区域。
步骤502、对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据。
步骤503、对所述直播视频流、所述识别结果以及所述目标数据进行编码操作,获得所述目标视频流。
在本实施例中,当目标特效对应的渲染方式为第二渲染方式时,也即在云端服务器执行特效算法操作,在虚拟现实设备进行特效渲染处理。
在获取到目标数据之后,可以根据该目标数据确定该目标特效对应的目标渲染区域以及深度等信息。进一步地,可以对直播视频流进行识别操作,获得直播视频流对应的识别结果,其中,识别结果中包括直播视频流对应的直播图像帧中与目标渲染区域对应的关键点信息以及直播图像帧对应的深度数据。
在确定直播视频流对应的识别结果之后,对直播视频流、识别结果以及目标数据进行编码操作,获得目标视频流。从而后续虚拟现实设备在获取到该目标视频流之后,能够基于该识别结果以及目标数据对直播视频流进行渲染操作。
可选地,在上述任一实施例的基础上,所述对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果,包括:
通过预设的关键点识别模型识别所述直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息。
通过预设的深度识别模型识别所述直播图像帧对应的深度数据。
将各直播图像帧对应的关键点信息以及深度数据确定为所述直播视频流对应的识别结果。
在本实施例中,为了实现对直播视频流中关键点信息以及深度数据的识 别操作,可以预先设置有关键点识别模型以及深度识别模型。因此,可以通过预设的关键点识别模型识别直播视频流对应的直播图像帧中与目标渲染区域对应的关键点信息。通过预设的深度识别模型识别直播图像帧对应的深度数据。将各直播图像帧对应的关键点信息以及深度数据确定为直播视频流对应的识别结果。通过关键点识别模型以及深度识别模型能够准确地实现对关键点信息以及深度数据的识别操作,提高了特效渲染的渲染效果。
本实施例提供的直播视频流渲染方法,通过在渲染方式为第二渲染方式时,在云端服务器进行特效算法的计算,在虚拟现实设备中进行特效渲染处理,从而能够在节约虚拟现实设备算力的基础上,保证了特效的实时性。使得用户能够与该特效进行交互操作。
图6为本公开实施例提供的直播视频流渲染方法的流程示意图,应用于虚拟现实设备,如图6所示,该方法包括:
步骤601、获取所述云端服务器发送的目标视频流,所述目标视频流为所述云端服务器根据虚拟现实直播对应的直播视频流以及特效标识信息,采用与所述目标特效对应的渲染方式进行渲染处理后获得的。
步骤602、采用所述目标特效对应的渲染方式对所述目标视频流进行处理,获得处理后的目标视频流。
步骤603、播放所述处理后的目标视频流。
本实施例的执行主体为直播视频流渲染装置,该直播视频流渲染装置可耦合于虚拟现实设备中。该虚拟现实设备可以与云端服务器通信连接。
在本实施例中,云端服务器可以获取该直播视频流以及特效标识信息,并根据该特效标识信息确定目标特效对应的渲染方式。采用与该目标特效对应的渲染方式对直播视频流进行渲染处理,得到处理后的目标视频流。将目标视频流发送至预设的虚拟现实设备进行播放。
相应地,虚拟现实设备可以获取该目标视频流,并采用目标特效对应的渲染方式对目标视频流进行处理,获得处理后的目标视频流。播放该处理后的目标视频流。
从而能够将复杂的特效算法操作放在云端服务器执行,虚拟现实设备仅需进行简单的解码以及渲染操作即可以实现渲染后的目标视频流的显示。
本实施例提供的直播视频流渲染方法,通过获取云端服务器发送的采用 与所述目标特效对应的渲染方式进行渲染处理后获得的目标视频流,从而虚拟现实设备仅需要对该目标视频流进行简单的处理即可以实现虚拟现实直播的显示,有效地节约了虚拟现实设备的算力,且能够保证虚拟现实直播的直播效果。
进一步地,在上述任一实施例的基础上,所述目标特效对应的渲染方式包括第一渲染方式。步骤602包括:
对所述目标视频流进行解码操作,获得所述处理后的目标视频流。
其中,所述处理后的视频流为云端服务器渲染后的视频流。
在本实施例中,在渲染方式为第一渲染方式时,也即在云端服务器中进行特效算法以及特效渲染处理时,该目标视频流可以为云端服务器渲染完毕的视频流。因此,在获取到目标视频流之后,虚拟现实设备可以对目标视频流进行解码操作,获得处理后的目标视频流。从而后续可以直接播放该处理后的目标视频流,无需对该处理后的目标视频流进行额外的数据处理,有效地降低了虚拟现实设备的算力。
本实施例提供的直播视频流渲染方法,通过在渲染方式为第一渲染方式时,对目标视频流进行解码操作,从而能够直接播放该处理后的目标视频流,无需对该处理后的目标视频流进行后续的特效计算以及特效渲染处理。有效地节约了虚拟现实设备的算力。
进一步地,在上述任一实施例的基础上,所述目标特效对应的渲染方式包括第二渲染方式。步骤602包括:
对所述目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、所述直播视频流对应的识别结果、目标数据以及渲染时间,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据。
根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得处理后的目标视频流。
在本实施例中,当目标特效对应的渲染方式为第二渲染方式时,也即在云端服务器执行特效算法操作,在虚拟现实设备进行特效渲染处理时,在获取到目标视频流之后,可以对目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、直播视频流对应的识别结果、目标数据以及渲染时间, 其中,识别结果中包括直播视频流对应的直播图像帧中与目标渲染区域对应的关键点信息以及直播图像帧对应的深度数据。
根据该识别结果以及目标数据对直播视频流进行渲染操作,获得处理后的目标视频流。从而后续虚拟现实设备可以播放该目标视频流。
可选地,虚拟现实设备可以在该渲染时间内根据直播视频流、直播视频流对应的识别结果、目标数据进行特效渲染处理。
本实施例提供的直播视频流渲染方法,通过在渲染方式为第二渲染方式时,在云端服务器进行特效算法的计算,在虚拟现实设备中进行特效渲染处理,从而能够在节约虚拟现实设备算力的基础上,保证了特效的实时性。使得用户能够与该特效进行交互操作。
进一步地,在上述任一实施例的基础上,所述对所述目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、所述直播视频流对应的识别结果以及目标数据之后,还包括:
确定所述目标特效对应的渲染时间。
在所述渲染时间内根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得处理后的目标视频流。
在本实施例中,不同的目标特效对应的渲染时间有所不同,该渲染时间云端服务器从配置平台中获取,并以汇编指令的方式编码在目标视频流里的。
因此,为了实现目标特效的准确渲染操作,在对目标视频流进行解码操作之后,可以确定目标特效对应的渲染时间。在渲染时间内根据识别结果以及目标数据对直播视频流进行渲染操作,获得处理后的目标视频流。
本实施例提供的直播视频流渲染方法,通过将渲染时间编码至所述目标视频流中,从而虚拟现实设备能够根据该渲染时间精准地对目标特效进行渲染操作,保证了目标特效的渲染效果。
进一步地,在上述任一实施例的基础上,所述目标特效对应的渲染方式包括第二渲染方式。步骤601之后,还包括:
若获取到用户触发的关闭特效指令,则对所述目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、所述直播视频流对应的识别结果以及目标数据,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据。
播放所述虚拟现实直播对应的直播视频流。
在本实施例中,当目标特效对应的渲染方式为第二渲染方式时,也即在云端服务器执行特效算法操作,在虚拟现实设备进行特效渲染处理时,观看虚拟现实直播的用户可以根据实际需求确定是否对目标特效进行渲染。
当用户触发关闭特效指令时,在获取到目标视频流之后,可以对目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、直播视频流对应的识别结果以及目标数据,其中,识别结果中包括直播视频流对应的直播图像帧中与目标渲染区域对应的关键点信息以及直播图像帧对应的深度数据。
虚拟现实设备可以无需根据识别结果以及目标数据对直播视频流进行渲染操作。其可以直接显示该直播视频流。
本实施例提供的直播视频流渲染方法,通过在渲染方式为第二渲染方式时,响应于用户触发的关闭特效指令,不对第二渲染方式的目标特效进行渲染操作,从而能够进一步地节约虚拟现实设备的算力。此外,能够使得虚拟现实直播的显示效果更加贴合用户的个性化需求,提升用户体验。
图7为本公开实施例提供的直播视频流渲染装置的结构示意图,应用于云端服务器,如图7所示,该装置包括:第一获取模块71、确定模块72、渲染模块73以及发送模块74。其中,第一获取模块71,用于获取虚拟现实直播对应的直播视频流以及特效标识信息,所述特效标识信息为用于渲染所述直播视频流的目标特效的标识信息。确定模块72,用于根据所述特效标识信息确定所述目标特效对应的渲染方式。渲染模块73,用于采用与所述目标特效对应的渲染方式对所述直播视频流进行渲染处理,获得处理后的目标视频流。发送模块74,用于将所述目标视频流发送至预设的虚拟现实设备进行播放。
进一步地,在上述任一实施例的基础上,所述第一获取模块用于:获取虚拟现实直播服务器发送的待渲染数据。对所述待渲染数据进行解码操作,获得所述直播视频流以及预设的特效渲染指令。对所述特效渲染指令进行解析操作,获得所述特效标识信息。
进一步地,在上述任一实施例的基础上,所述确定模块,用于:根据所述特效标识信息在预设的配置平台中获取用于渲染所述目标特效的目标数据。在所述目标数据中确定所述目标特效对应的渲染方式。其中,所述渲染方式 包括第一渲染方式和/或第二渲染方式。
进一步地,在上述任一实施例的基础上,所述渲染方式包括第一渲染方式。所述渲染模块,用于:根据所述目标数据确定所述目标特效对应的目标渲染区域。对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据。根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得渲染后的视频流。对所述渲染后的视频流进行编码操作,获得所述目标视频流。
进一步地,在上述任一实施例的基础上,所述渲染模块,用于:根据所述识别结果以及所述目标数据对所述直播视频流对应的多个直播图像帧进行渲染操作,获得多张渲染后的目标图像帧。根据所述多张渲染后的目标图像帧生成所述渲染后的视频流。
进一步地,在上述任一实施例的基础上,所述渲染方式包括第二渲染方式。所述渲染模块,用于:根据所述目标数据确定所述目标特效对应的目标渲染区域。对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据。对所述直播视频流、所述识别结果以及所述目标数据进行编码操作,获得所述目标视频流。
进一步地,在上述任一实施例的基础上,所述渲染模块,用于:通过预设的关键点识别模型识别所述直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息。通过预设的深度识别模型识别所述直播图像帧对应的深度数据。将各直播图像帧对应的关键点信息以及深度数据确定为所述直播视频流对应的识别结果。
进一步地,在上述任一实施例的基础上,所述装置还包括:确定模块,用于根据所述特效标识信息在预设的配置平台中确定所述目标特效对应的渲染时间。编码模块,用于将所述渲染时间编码至所述目标视频流中,以使虚拟现实设备在所述渲染时间内对所述直播视频流进行渲染操作。
图8为本公开实施例提供的直播视频流渲染装置的结构示意图,应用于虚拟现实设备,如图8所示,该装置包括:第二获取模块81、处理模块82以 及播放模块83。其中,第二获取模块81,用于获取所述云端服务器发送的目标视频流,所述目标视频流为所述云端服务器根据虚拟现实直播对应的直播视频流以及特效标识信息,采用与所述目标特效对应的渲染方式进行渲染处理后获得的。处理模块82,用于采用所述目标特效对应的渲染方式对所述目标视频流进行处理,获得处理后的目标视频流。播放模块83,用于播放所述处理后的目标视频流。
进一步地,在上述任一实施例的基础上,所述目标特效对应的渲染方式包括第一渲染方式。所述处理模块,用于:对所述目标视频流进行解码操作,获得所述处理后的目标视频流。其中,所述处理后的视频流为云端服务器渲染后的视频流。
进一步地,在上述任一实施例的基础上,所述目标特效对应的渲染方式包括第二渲染方式。所述处理模块,用于:对所述目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、所述直播视频流对应的识别结果、目标数据以及渲染时间,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据。根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得处理后的目标视频流。
进一步地,在上述任一实施例的基础上,所述处理模块,用于:确定所述目标特效对应的渲染时间。在所述渲染时间内根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得处理后的目标视频流。
进一步地,在上述任一实施例的基础上,所述目标特效对应的渲染方式包括第二渲染方式。所述装置还包括:解码模块,用于若获取到用户触发的关闭特效指令,则对所述目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、所述直播视频流对应的识别结果以及目标数据,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据。播放模块,用于播放所述虚拟现实直播对应的直播视频流。
本实施例提供的设备,可用于执行上述方法实施例的技术方案,其实现原理和技术效果类似,本实施例此处不再赘述。
为了实现上述实施例,本公开实施例还提供了一种云渲染系统,包括云 端服务器以及虚拟现实设备,所述云端服务器与所述虚拟现实设备通信连接;
其中,所述云端服务器用于获取虚拟现实直播对应的直播视频流以及特效标识信息,并根据所述特效标识信息确定所述目标特效对应的渲染方式,所述特效标识信息为用于渲染所述直播视频流的目标特效的标识信息;
所述云端服务器还用于在所述渲染方式为第一渲染方式时,根据所述直播视频流以及特效标识信息进行特效算法处理以及特效渲染处理,获得目标视频流;
所述云端服务器还用于在所述渲染方式为第二渲染方式时,根据所述直播视频流以及特效标识信息进行特效算法处理,并将所述直播视频流以及识别结果发送至所述虚拟现实设备;
所述虚拟现实设备用于获取所述云端服务器发送的直播视频流以及识别结果进行特效渲染处理。
为了实现上述实施例,本公开实施例还提供了一种电子设备,包括:处理器和存储器。
所述存储器存储计算机执行指令。
所述处理器执行所述存储器存储的计算机执行指令,使得所述处理器执行如上述任一实施例所述的直播视频流渲染方法。
图9为本公开实施例提供的电子设备的结构示意图,如图9所示,该电子设备900可以为终端设备或服务器。其中,终端设备可以包括但不限于诸如移动电话、笔记本电脑、数字广播接收器、个人数字助理(Personal Digital Assistant,简称PDA)、平板电脑(Portable Android Device,简称PAD)、便携式多媒体播放器(Portable Media Player,简称PMP)、车载终端(例如车载导航终端)等等的移动终端以及诸如数字TV、台式计算机等等的固定终端。图9示出的电子设备仅仅是一个示例,不应对本公开实施例的功能和使用范围带来任何限制。
如图9所示,电子设备900可以包括处理装置(例如中央处理器、图形处理器等)901,其可以根据存储在只读存储器(Read Only Memory,简称ROM)902中的程序或者从存储装置908加载到随机访问存储器(Random Access Memory,简称RAM)903中的程序而执行各种适当的动作和处理。在RAM 903中,还存储有电子设备900操作所需的各种程序和数据。处理装 置901、ROM 902以及RAM 903通过总线904彼此相连。输入/输出(I/O)接口905也连接至总线904。
通常,以下装置可以连接至I/O接口905:包括例如触摸屏、触摸板、键盘、鼠标、摄像头、麦克风、加速度计、陀螺仪等的输入装置906;包括例如液晶显示器(Liquid Crystal Display,简称LCD)、扬声器、振动器等的输出装置907;包括例如磁带、硬盘等的存储装置908;以及通信装置909。通信装置909可以允许电子设备900与其他设备进行无线或有线通信以交换数据。虽然图9示出了具有各种装置的电子设备900,但是应理解的是,并不要求实施或具备所有示出的装置。可以替代地实施或具备更多或更少的装置。
特别地,根据本公开的实施例,上文参考流程图描述的过程可以被实现为计算机软件程序。例如,本公开的实施例包括一种计算机程序产品,其包括承载在计算机可读介质上的计算机程序,该计算机程序包含用于执行流程图所示的方法的程序代码。在这样的实施例中,该计算机程序可以通过通信装置909从网络上被下载和安装,或者从存储装置908被安装,或者从ROM902被安装。在该计算机程序被处理装置901执行时,执行本公开实施例的方法中限定的上述功能。
需要说明的是,本公开上述的计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质或者是上述两者的任意组合。计算机可读存储介质例如可以是——但不限于——电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。计算机可读存储介质的更具体的例子可以包括但不限于:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机访问存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、光纤、便携式紧凑磁盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。在本公开中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。而在本公开中,计算机可读信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。计算机可读信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机可读信号 介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。计算机可读介质上包含的程序代码可以用任何适当的介质传输,包括但不限于:电线、光缆、RF(射频)等等,或者上述的任意合适的组合。
本公开实施例还提供了一种计算机可读存储介质,所述计算机可读存储介质中存储有计算机执行指令,当处理器执行所述计算机执行指令时,实现如上述任一实施例所述的直播视频流渲染方法。
本公开实施例还提供了一种计算机程序产品,包括计算机程序,所述计算机程序被处理器执行时实现如上述任一实施例所述的虚拟现实直播渲染的方法。
上述计算机可读介质可以是上述电子设备中所包含的;也可以是单独存在,而未装配入该电子设备中。
上述计算机可读介质承载有一个或者多个程序,当上述一个或者多个程序被该电子设备执行时,使得该电子设备执行上述实施例所示的方法。
可以以一种或多种程序设计语言或其组合来编写用于执行本公开的操作的计算机程序代码,上述程序设计语言包括面向对象的程序设计语言—诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络——包括局域网(Local Area Network,简称LAN)或广域网(Wide Area Network,简称WAN)—连接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。
附图中的流程图和框图,图示了按照本公开各种实施例的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段、或代码的一部分,该模块、程序段、或代码的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。也应当注意,在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个接连地表示的方框实际上可以 基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。
描述于本公开实施例中所涉及到的单元可以通过软件的方式实现,也可以通过硬件的方式来实现。其中,单元的名称在某种情况下并不构成对该单元本身的限定,例如,第一获取单元还可以被描述为“获取至少两个网际协议地址的单元”。
本文中以上描述的功能可以至少部分地由一个或多个硬件逻辑部件来执行。例如,非限制性地,可以使用的示范类型的硬件逻辑部件包括:现场可编程门阵列(FPGA)、专用集成电路(ASIC)、专用标准产品(ASSP)、片上系统(SOC)、复杂可编程逻辑设备(CPLD)等等。
在本公开的上下文中,机器可读介质可以是有形的介质,其可以包含或存储以供指令执行系统、装置或设备使用或与指令执行系统、装置或设备结合地使用的程序。机器可读介质可以是机器可读信号介质或机器可读储存介质。机器可读介质可以包括但不限于电子的、磁性的、光学的、电磁的、红外的、或半导体系统、装置或设备,或者上述内容的任何合适组合。机器可读存储介质的更具体示例会包括基于一个或多个线的电气连接、便携式计算机盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦除可编程只读存储器(EPROM或快闪存储器)、光纤、便捷式紧凑盘只读存储器(CD-ROM)、光学储存设备、磁储存设备、或上述内容的任何合适组合。
第一方面,根据本公开的一个或多个实施例,提供了一种直播视频流渲染方法,应用于云端渲染系统,所述云端渲染系统包括云端服务器以及虚拟现实设备,所述方法包括:
云端服务器获取虚拟现实直播对应的直播视频流以及特效标识信息,所述特效标识信息为用于渲染所述直播视频流的目标特效的标识信息;
云端服务器根据所述特效标识信息确定所述目标特效对应的渲染方式,其中,所述渲染方式包括第一渲染方式和/或第二渲染方式;
当所述渲染方式为第一渲染方式时,所述云端服务器根据所述直播视频流以及特效标识信息进行特效算法处理,以及根据所述特效算法处理对应的 识别结果进行特效渲染处理,获得目标视频流,并将所述目标视频流发送至所述虚拟现实设备,所述虚拟现实设备获取所述云端服务器发送的目标视频流并播放;
当所述渲染方式为第二渲染方式时,所述云端服务器根据所述直播视频流以及特效标识信息进行特效算法处理,并将所述直播视频流以及识别结果发送至所述虚拟现实设备,所述虚拟现实设备根据所述直播视频流以及算法结果进行特效渲染处理,获得处理后的目标视频流,播放所述处理后的目标视频流。
根据本公开的一个或多个实施例,所述云端服务器根据所述直播视频流以及特效标识信息进行特效算法处理,包括:
所述云端服务器根据所述特效标识信息在预设的配置平台中获取用于渲染所述目标特效的目标数据;
所述云端服务器根据所述目标数据确定所述目标特效对应的目标渲染区域;
所述云端服务器对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据。
根据本公开的一个或多个实施例,所述根据所述特效算法处理的识别结果进行特效渲染处理,获得目标视频流,包括:
所述云端服务器根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得渲染后的视频流;
所述云端服务器对所述渲染后的视频流进行编码操作,获得所述目标视频流。
根据本公开的一个或多个实施例,所述云端服务器根据所述特效标识信息确定所述目标特效对应的渲染方式之后,还包括:
所述云端服务器根据所述特效标识信息在预设的配置平台中获取用于渲染所述目标特效的目标数据;
所述云端服务器根据所述目标数据确定所述目标特效对应的目标渲染区域;
所述云端服务器对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
所述云端服务器对所述直播视频流、所述识别结果以及所述目标数据进行编码操作,获得所述目标视频流,将所述目标视频流发送至所述虚拟现实设备。
根据本公开的一个或多个实施例,所述将所述目标视频流发送至所述虚拟现实设备之后,还包括:
所述虚拟现实设备获取所述目标视频流,对所述目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、所述直播视频流对应的识别结果、目标数据,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得处理后的目标视频流。
第二方面,根据本公开的一个或多个实施例,提供了一种直播视频流渲染方法,应用于云端服务器,包括:
获取虚拟现实直播对应的直播视频流以及特效标识信息,所述特效标识信息为用于渲染所述直播视频流的目标特效的标识信息;
根据所述特效标识信息确定所述目标特效对应的渲染方式;
采用与所述目标特效对应的渲染方式对所述直播视频流进行渲染处理,获得处理后的目标视频流;
将所述目标视频流发送至预设的虚拟现实设备进行播放。
根据本公开的一个或多个实施例,所述获取虚拟现实直播对应的直播视频流以及特效标识信息,包括:
获取虚拟现实直播服务器发送的待渲染数据;
对所述待渲染数据进行解码操作,获得所述直播视频流以及预设的特效渲染指令;
对所述特效渲染指令进行解析操作,获得所述特效标识信息。
根据本公开的一个或多个实施例,所述根据所述特效标识信息确定所述 目标特效对应的渲染方式,包括:
根据所述特效标识信息在预设的配置平台中获取用于渲染所述目标特效的目标数据;
在所述目标数据中确定所述目标特效对应的渲染方式;
其中,所述渲染方式包括第一渲染方式和/或第二渲染方式。
根据本公开的一个或多个实施例,所述渲染方式包括第一渲染方式;
所述采用与所述目标特效对应的渲染方式对所述直播视频流进行渲染处理,获得处理后的目标视频流,包括:
根据所述目标数据确定所述目标特效对应的目标渲染区域;
对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得渲染后的视频流;
对所述渲染后的视频流进行编码操作,获得所述目标视频流。
根据本公开的一个或多个实施例,所述根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得所述目标视频流,包括:
根据所述识别结果以及所述目标数据对所述直播视频流对应的多个直播图像帧进行渲染操作,获得多张渲染后的目标图像帧;
根据所述多张渲染后的目标图像帧生成所述渲染后的视频流。
根据本公开的一个或多个实施例,所述渲染方式包括第二渲染方式;
所述采用与所述目标特效对应的渲染方式对所述直播视频流进行渲染处理,获得处理后的目标视频流,包括:
根据所述目标数据确定所述目标特效对应的目标渲染区域;
对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
对所述直播视频流、所述识别结果以及所述目标数据进行编码操作,获得所述目标视频流。
根据本公开的一个或多个实施例,所述对所述直播视频流进行识别操作, 获得所述直播视频流对应的识别结果,包括:
通过预设的关键点识别模型识别所述直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息;
通过预设的深度识别模型识别所述直播图像帧对应的深度数据;
将各直播图像帧对应的关键点信息以及深度数据确定为所述直播视频流对应的识别结果。
根据本公开的一个或多个实施例,所述采用与所述目标特效对应的渲染方式对所述直播视频流进行渲染处理,获得处理后的目标视频流之后,还包括:
根据所述特效标识信息在预设的配置平台中确定所述目标特效对应的渲染时间;
将所述渲染时间编码至所述目标视频流中,以使虚拟现实设备在所述渲染时间内对所述直播视频流进行渲染操作。
第三方面,根据本公开的一个或多个实施例,提供了一种直播视频流渲染方法,应用于虚拟现实设备,包括:
获取所述云端服务器发送的目标视频流,所述目标视频流为所述云端服务器根据虚拟现实直播对应的直播视频流以及特效标识信息,采用与所述目标特效对应的渲染方式进行渲染处理后获得的;
采用所述目标特效对应的渲染方式对所述目标视频流进行处理,获得处理后的目标视频流;
播放所述处理后的目标视频流。
根据本公开的一个或多个实施例,所述目标特效对应的渲染方式包括第一渲染方式;
所述采用所述目标特效对应的渲染方式对所述目标视频流进行处理,获得处理后的目标视频流,包括:
对所述目标视频流进行解码操作,获得所述处理后的目标视频流;
其中,所述处理后的视频流为云端服务器渲染后的视频流。
根据本公开的一个或多个实施例,所述目标特效对应的渲染方式包括第二渲染方式;
所述采用所述目标特效对应的渲染方式对所述目标视频流进行处理,包 括:
对所述目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、所述直播视频流对应的识别结果、目标数据以及渲染时间,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得处理后的目标视频流。
根据本公开的一个或多个实施例,所述对所述目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、所述直播视频流对应的识别结果以及目标数据之后,还包括:
确定所述目标特效对应的渲染时间;
在所述渲染时间内根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得处理后的目标视频流。
根据本公开的一个或多个实施例,所述目标特效对应的渲染方式包括第二渲染方式;
所述获取所述云端服务器发送的目标视频流之后,还包括:
若获取到用户触发的关闭特效指令,则对所述目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、所述直播视频流对应的识别结果以及目标数据,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
播放所述虚拟现实直播对应的直播视频流。
第四方面,根据本公开的一个或多个实施例,提供了一种直播视频流渲染装置,应用于云端服务器,包括:
第一获取模块,用于获取虚拟现实直播对应的直播视频流以及特效标识信息,所述特效标识信息为用于渲染所述直播视频流的目标特效的标识信息;
确定模块,用于根据所述特效标识信息确定所述目标特效对应的渲染方式;
渲染模块,用于采用与所述目标特效对应的渲染方式对所述直播视频流进行渲染处理,获得处理后的目标视频流;
发送模块,用于将所述目标视频流发送至预设的虚拟现实设备进行播放。
根据本公开的一个或多个实施例,所述第一获取模块用于:
获取虚拟现实直播服务器发送的待渲染数据;
对所述待渲染数据进行解码操作,获得所述直播视频流以及预设的特效渲染指令;
对所述特效渲染指令进行解析操作,获得所述特效标识信息。
根据本公开的一个或多个实施例,所述确定模块,用于:
根据所述特效标识信息在预设的配置平台中获取用于渲染所述目标特效的目标数据;
在所述目标数据中确定所述目标特效对应的渲染方式;
其中,所述渲染方式包括第一渲染方式和/或第二渲染方式。
根据本公开的一个或多个实施例,所述渲染方式包括第一渲染方式;
所述渲染模块,用于:
根据所述目标数据确定所述目标特效对应的目标渲染区域;
对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得渲染后的视频流;
对所述渲染后的视频流进行编码操作,获得所述目标视频流。
根据本公开的一个或多个实施例,所述渲染模块,用于:
根据所述识别结果以及所述目标数据对所述直播视频流对应的多个直播图像帧进行渲染操作,获得多张渲染后的目标图像帧;
根据所述多张渲染后的目标图像帧生成所述渲染后的视频流。
根据本公开的一个或多个实施例,所述渲染方式包括第二渲染方式;
所述渲染模块,用于:
根据所述目标数据确定所述目标特效对应的目标渲染区域;
对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
对所述直播视频流、所述识别结果以及所述目标数据进行编码操作,获 得所述目标视频流。
根据本公开的一个或多个实施例,所述渲染模块,用于:
通过预设的关键点识别模型识别所述直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息;
通过预设的深度识别模型识别所述直播图像帧对应的深度数据;
将各直播图像帧对应的关键点信息以及深度数据确定为所述直播视频流对应的识别结果。
根据本公开的一个或多个实施例,所述装置还包括:
确定模块,用于根据所述特效标识信息在预设的配置平台中确定所述目标特效对应的渲染时间;
编码模块,用于将所述渲染时间编码至所述目标视频流中,以使虚拟现实设备在所述渲染时间内对所述直播视频流进行渲染操作。
第五方面,根据本公开的一个或多个实施例,提供了一种直播视频流渲染装置,应用于虚拟现实设备,包括:
第二获取模块,用于获取所述云端服务器发送的目标视频流,所述目标视频流为所述云端服务器根据虚拟现实直播对应的直播视频流以及特效标识信息,采用与所述目标特效对应的渲染方式进行渲染处理后获得的;
处理模块,用于采用所述目标特效对应的渲染方式对所述目标视频流进行处理,获得处理后的目标视频流;
播放模块,用于播放所述处理后的目标视频流。
根据本公开的一个或多个实施例,所述目标特效对应的渲染方式包括第一渲染方式;
所述处理模块,用于:
对所述目标视频流进行解码操作,获得所述处理后的目标视频流;
其中,所述处理后的视频流为云端服务器渲染后的视频流。
根据本公开的一个或多个实施例,所述目标特效对应的渲染方式包括第二渲染方式;
所述处理模块,用于:
对所述目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、所述直播视频流对应的识别结果、目标数据以及渲染时间,其中,所述识别 结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得处理后的目标视频流。
根据本公开的一个或多个实施例,所述处理模块,用于:
确定所述目标特效对应的渲染时间;
在所述渲染时间内根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得处理后的目标视频流。
根据本公开的一个或多个实施例,所述目标特效对应的渲染方式包括第二渲染方式;
所述装置还包括:
解码模块,用于若获取到用户触发的关闭特效指令,则对所述目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、所述直播视频流对应的识别结果以及目标数据,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
播放模块,用于播放所述虚拟现实直播对应的直播视频流。
第六方面,根据本公开的一个或多个实施例,提供了一种云渲染系统,包括云端服务器以及虚拟现实设备,所述云端服务器与所述虚拟现实设备通信连接;
其中,所述云端服务器用于获取虚拟现实直播对应的直播视频流以及特效标识信息,并根据所述特效标识信息确定所述目标特效对应的渲染方式,所述特效标识信息为用于渲染所述直播视频流的目标特效的标识信息;
所述云端服务器还用于在所述渲染方式为第一渲染方式时,根据所述直播视频流以及特效标识信息进行特效算法处理以及特效渲染处理,获得目标视频流;
所述云端服务器还用于在所述渲染方式为第二渲染方式时,根据所述直播视频流以及特效标识信息进行特效算法处理,并将所述直播视频流以及识别结果发送至所述虚拟现实设备;
所述虚拟现实设备用于获取所述云端服务器发送的直播视频流以及识别 结果进行特效渲染处理。
第七方面,根据本公开的一个或多个实施例,提供了一种电子设备,包括:至少一个处理器和存储器;
所述存储器存储计算机执行指令;
所述至少一个处理器执行所述存储器存储的计算机执行指令,使得所述至少一个处理器执行如上第一方面以及第一方面各种可能的设计或第二方面以及第二方面各种可能的设计所述的直播视频流渲染方法。
第八方面,根据本公开的一个或多个实施例,提供了一种计算机可读存储介质,所述计算机可读存储介质中存储有计算机执行指令,当处理器执行所述计算机执行指令时,实现如上第一方面以及第一方面各种可能的设计或第二方面以及第二方面各种可能的设计所述的直播视频流渲染方法。
第九方面,根据本公开的一个或多个实施例,提供了一种计算机程序产品,包括计算机程序,所述计算机程序被处理器执行时实现如上第一方面以及第一方面各种可能的设计或第二方面以及第二方面各种可能的设计所述的直播视频流渲染方法。
以上描述仅为本公开的较佳实施例以及对所运用技术原理的说明。本领域技术人员应当理解,本公开中所涉及的公开范围,并不限于上述技术特征的特定组合而成的技术方案,同时也应涵盖在不脱离上述公开构思的情况下,由上述技术特征或其等同特征进行任意组合而形成的其它技术方案。例如上述特征与本公开中公开的(但不限于)具有类似功能的技术特征进行互相替换而形成的技术方案。
此外,虽然采用特定次序描绘了各操作,但是这不应当理解为要求这些操作以所示出的特定次序或以顺序次序执行来执行。在一定环境下,多任务和并行处理可能是有利的。同样地,虽然在上面论述中包含了若干具体实现细节,但是这些不应当被解释为对本公开的范围的限制。在单独的实施例的上下文中描述的某些特征还可以组合地实现在单个实施例中。相反地,在单个实施例的上下文中描述的各种特征也可以单独地或以任何合适的子组合的方式实现在多个实施例中。
尽管已经采用特定于结构特征和/或方法逻辑动作的语言描述了本主题,但是应当理解所附权利要求书中所限定的主题未必局限于上面描述的特定特 征或动作。相反,上面所描述的特定特征和动作仅仅是实现权利要求书的示例形式。

Claims (20)

  1. 一种直播视频流渲染方法,应用于云端渲染系统,其特征在于,所述云端渲染系统包括云端服务器以及虚拟现实设备,所述方法包括:
    云端服务器获取虚拟现实直播对应的直播视频流以及特效标识信息,所述特效标识信息为用于渲染所述直播视频流的目标特效的标识信息;
    云端服务器根据所述特效标识信息确定所述目标特效对应的渲染方式,其中,所述渲染方式包括第一渲染方式和/或第二渲染方式;
    当所述渲染方式为第一渲染方式时,所述云端服务器根据所述直播视频流以及特效标识信息进行特效算法处理,以及根据所述特效算法处理对应的识别结果进行特效渲染处理,获得目标视频流,并将所述目标视频流发送至所述虚拟现实设备,所述虚拟现实设备获取所述云端服务器发送的目标视频流并播放;
    当所述渲染方式为第二渲染方式时,所述云端服务器根据所述直播视频流以及特效标识信息进行特效算法处理,并将所述直播视频流以及识别结果发送至所述虚拟现实设备,所述虚拟现实设备根据所述直播视频流以及算法结果进行特效渲染处理,获得处理后的目标视频流,播放所述处理后的目标视频流。
  2. 根据权利要求1所述的方法,其特征在于,所述云端服务器根据所述直播视频流以及特效标识信息进行特效算法处理,包括:
    所述云端服务器根据所述特效标识信息在预设的配置平台中获取用于渲染所述目标特效的目标数据;
    所述云端服务器根据所述目标数据确定所述目标特效对应的目标渲染区域;
    所述云端服务器对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据。
  3. 根据权利要求1所述的方法,其特征在于,所述根据所述特效算法处理的识别结果进行特效渲染处理,获得目标视频流,包括:
    所述云端服务器根据所述识别结果以及所述目标数据对所述直播视频流 进行渲染操作,获得渲染后的视频流;
    所述云端服务器对所述渲染后的视频流进行编码操作,获得所述目标视频流。
  4. 根据权利要求1所述的方法,其特征在于,所述云端服务器根据所述特效标识信息确定所述目标特效对应的渲染方式之后,还包括:
    所述云端服务器根据所述特效标识信息在预设的配置平台中获取用于渲染所述目标特效的目标数据;
    所述云端服务器根据所述目标数据确定所述目标特效对应的目标渲染区域;
    所述云端服务器对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
    所述云端服务器对所述直播视频流、所述识别结果以及所述目标数据进行编码操作,获得所述目标视频流,将所述目标视频流发送至所述虚拟现实设备。
  5. 根据权利要求4所述的方法,其特征在于,所述将所述目标视频流发送至所述虚拟现实设备之后,还包括:
    所述虚拟现实设备获取所述目标视频流,对所述目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、所述直播视频流对应的识别结果、目标数据,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
    根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得处理后的目标视频流。
  6. 一种直播视频流渲染方法,应用于云端服务器,其特征在于,包括:
    获取虚拟现实直播对应的直播视频流以及特效标识信息,所述特效标识信息为用于渲染所述直播视频流的目标特效的标识信息;
    根据所述特效标识信息确定所述目标特效对应的渲染方式;
    采用与所述目标特效对应的渲染方式对所述直播视频流进行渲染处理,获得处理后的目标视频流;
    将所述目标视频流发送至预设的虚拟现实设备进行播放。
  7. 根据权利要求6所述的方法,其特征在于,所述获取虚拟现实直播对应的直播视频流以及特效标识信息,包括:
    获取虚拟现实直播服务器发送的待渲染数据;
    对所述待渲染数据进行解码操作,获得所述直播视频流以及预设的特效渲染指令;
    对所述特效渲染指令进行解析操作,获得所述特效标识信息。
  8. 根据权利要求6所述的方法,其特征在于,所述根据所述特效标识信息确定所述目标特效对应的渲染方式,包括:
    根据所述特效标识信息在预设的配置平台中获取用于渲染所述目标特效的目标数据;
    在所述目标数据中确定所述目标特效对应的渲染方式;
    其中,所述渲染方式包括第一渲染方式和/或第二渲染方式。
  9. 根据权利要求8所述的方法,其特征在于,所述渲染方式包括第一渲染方式;
    所述采用与所述目标特效对应的渲染方式对所述直播视频流进行渲染处理,获得处理后的目标视频流,包括:
    根据所述目标数据确定所述目标特效对应的目标渲染区域;
    对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
    根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得渲染后的视频流;
    对所述渲染后的视频流进行编码操作,获得所述目标视频流。
  10. 根据权利要求8所述的方法,其特征在于,所述渲染方式包括第二渲染方式;
    所述采用与所述目标特效对应的渲染方式对所述直播视频流进行渲染处理,获得处理后的目标视频流,包括:
    根据所述目标数据确定所述目标特效对应的目标渲染区域;
    对所述直播视频流进行识别操作,获得所述直播视频流对应的识别结果, 其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
    对所述直播视频流、所述识别结果以及所述目标数据进行编码操作,获得所述目标视频流。
  11. 一种直播视频流渲染方法,应用于虚拟现实设备,其特征在于,包括:
    获取云端服务器发送的目标视频流,所述目标视频流为所述云端服务器根据虚拟现实直播对应的直播视频流以及特效标识信息,采用与所述目标特效对应的渲染方式进行渲染处理后获得的;
    采用所述目标特效对应的渲染方式对所述目标视频流进行处理,获得处理后的目标视频流;
    播放所述处理后的目标视频流。
  12. 根据权利要求11所述的方法,其特征在于,所述目标特效对应的渲染方式包括第一渲染方式;
    所述采用所述目标特效对应的渲染方式对所述目标视频流进行处理,获得处理后的目标视频流,包括:
    对所述目标视频流进行解码操作,获得所述处理后的目标视频流;
    其中,所述处理后的视频流为云端服务器渲染后的视频流。
  13. 根据权利要求11所述的方法,其特征在于,所述目标特效对应的渲染方式包括第二渲染方式;
    所述采用所述目标特效对应的渲染方式对所述目标视频流进行处理,包括:
    对所述目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、所述直播视频流对应的识别结果、目标数据,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
    根据所述识别结果以及所述目标数据对所述直播视频流进行渲染操作,获得处理后的目标视频流。
  14. 根据权利要求11所述的方法,其特征在于,所述目标特效对应的渲染方式包括第二渲染方式;
    所述获取所述云端服务器发送的目标视频流之后,还包括:
    若获取到用户触发的关闭特效指令,则对所述目标视频流进行解码操作,获得虚拟现实直播对应的直播视频流、所述直播视频流对应的识别结果以及目标数据,其中,所述识别结果中包括直播视频流对应的直播图像帧中与所述目标渲染区域对应的关键点信息以及所述直播图像帧对应的深度数据;
    播放所述虚拟现实直播对应的直播视频流。
  15. 一种直播视频流渲染装置,应用于云端服务器,其特征在于,包括:
    第一获取模块,用于获取虚拟现实直播对应的直播视频流以及特效标识信息,所述特效标识信息为用于渲染所述直播视频流的目标特效的标识信息;
    确定模块,用于根据所述特效标识信息确定所述目标特效对应的渲染方式;
    渲染模块,用于采用与所述目标特效对应的渲染方式对所述直播视频流进行渲染处理,获得处理后的目标视频流;
    发送模块,用于将所述目标视频流发送至预设的虚拟现实设备进行播放。
  16. 一种直播视频流渲染装置,应用于虚拟现实设备,其特征在于,包括:
    第二获取模块,用于获取云端服务器发送的目标视频流,所述目标视频流为所述云端服务器根据虚拟现实直播对应的直播视频流以及特效标识信息,采用与所述目标特效对应的渲染方式进行渲染处理后获得的;
    处理模块,用于采用所述目标特效对应的渲染方式对所述目标视频流进行处理,获得处理后的目标视频流;
    播放模块,用于播放所述处理后的目标视频流。
  17. 一种云渲染系统,其特征在于,包括云端服务器以及虚拟现实设备,所述云端服务器与所述虚拟现实设备通信连接;
    其中,所述云端服务器用于获取虚拟现实直播对应的直播视频流以及特效标识信息,并根据所述特效标识信息确定目标特效对应的渲染方式,所述特效标识信息为用于渲染所述直播视频流的目标特效的标识信息;
    所述云端服务器还用于在所述渲染方式为第一渲染方式时,根据所述直播视频流以及特效标识信息进行特效算法处理以及特效渲染处理,获得目标视频流;
    所述云端服务器还用于在所述渲染方式为第二渲染方式时,根据所述直播视频流以及特效标识信息进行特效算法处理,并将所述直播视频流以及识别结果发送至所述虚拟现实设备;
    所述虚拟现实设备用于获取所述云端服务器发送的直播视频流以及识别结果进行特效渲染处理。
  18. 一种电子设备,其特征在于,包括:至少一个处理器和存储器;
    所述存储器存储计算机执行指令;
    所述处理器执行所述存储器存储的计算机执行指令,使得所述处理器执行如权利要求1-14任一项所述的直播视频流渲染方法。
  19. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质中存储有计算机执行指令,当处理器执行所述计算机执行指令时,实现如权利要求1-14任一项所述的直播视频流渲染方法。
  20. 一种计算机程序产品,包括计算机程序,其特征在于,所述计算机程序被处理器执行时实现如权利要求1-14任一项所述的虚拟现实直播渲染的方法。
PCT/CN2023/131376 2022-11-17 2023-11-13 直播视频流渲染方法、装置、设备、存储介质及产品 WO2024104307A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202211441161.5A CN115767181A (zh) 2022-11-17 2022-11-17 直播视频流渲染方法、装置、设备、存储介质及产品
CN202211441161.5 2022-11-17

Publications (1)

Publication Number Publication Date
WO2024104307A1 true WO2024104307A1 (zh) 2024-05-23

Family

ID=85372761

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/131376 WO2024104307A1 (zh) 2022-11-17 2023-11-13 直播视频流渲染方法、装置、设备、存储介质及产品

Country Status (2)

Country Link
CN (1) CN115767181A (zh)
WO (1) WO2024104307A1 (zh)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115767181A (zh) * 2022-11-17 2023-03-07 北京字跳网络技术有限公司 直播视频流渲染方法、装置、设备、存储介质及产品
CN117041628B (zh) * 2023-10-09 2024-02-02 腾讯科技(深圳)有限公司 直播画面的渲染方法、系统、装置、设备及介质
CN117061791B (zh) * 2023-10-12 2024-01-26 深圳云天畅想信息科技有限公司 云视频帧自适应协作渲染方法、装置及计算机设备

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111163330A (zh) * 2020-01-13 2020-05-15 广州虎牙科技有限公司 直播视频的渲染方法、装置、系统、设备及存储介质
CN111464828A (zh) * 2020-05-14 2020-07-28 广州酷狗计算机科技有限公司 虚拟特效显示方法、装置、终端及存储介质
CN112235585A (zh) * 2020-08-31 2021-01-15 江苏视博云信息技术有限公司 一种虚拟场景的直播方法、装置及系统
CN113554737A (zh) * 2020-12-04 2021-10-26 腾讯科技(深圳)有限公司 目标对象的动作驱动方法、装置、设备及存储介质
WO2021249414A1 (zh) * 2020-06-10 2021-12-16 阿里巴巴集团控股有限公司 数据处理方法、系统、相关设备和存储介质
CN113965812A (zh) * 2021-12-21 2022-01-21 广州虎牙信息科技有限公司 直播方法、系统及直播设备
CN114900678A (zh) * 2022-07-15 2022-08-12 北京蔚领时代科技有限公司 一种vr端端云结合的虚拟演唱会渲染方法及系统
CN115767181A (zh) * 2022-11-17 2023-03-07 北京字跳网络技术有限公司 直播视频流渲染方法、装置、设备、存储介质及产品

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111163330A (zh) * 2020-01-13 2020-05-15 广州虎牙科技有限公司 直播视频的渲染方法、装置、系统、设备及存储介质
CN111464828A (zh) * 2020-05-14 2020-07-28 广州酷狗计算机科技有限公司 虚拟特效显示方法、装置、终端及存储介质
WO2021249414A1 (zh) * 2020-06-10 2021-12-16 阿里巴巴集团控股有限公司 数据处理方法、系统、相关设备和存储介质
CN112235585A (zh) * 2020-08-31 2021-01-15 江苏视博云信息技术有限公司 一种虚拟场景的直播方法、装置及系统
CN113554737A (zh) * 2020-12-04 2021-10-26 腾讯科技(深圳)有限公司 目标对象的动作驱动方法、装置、设备及存储介质
CN113965812A (zh) * 2021-12-21 2022-01-21 广州虎牙信息科技有限公司 直播方法、系统及直播设备
CN114900678A (zh) * 2022-07-15 2022-08-12 北京蔚领时代科技有限公司 一种vr端端云结合的虚拟演唱会渲染方法及系统
CN115767181A (zh) * 2022-11-17 2023-03-07 北京字跳网络技术有限公司 直播视频流渲染方法、装置、设备、存储介质及产品

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
CHEN, XUELIANG ET AL.: "Research on Virtual Reality Optimization Scheme Based on 5G Network", GUANGDONG COMMUNICATION TECHNOLOGY, GUANGDONG SHENG DIANXIN YOUXIAN GONGSI YANJIUYUAN, CN, no. 10, 15 October 2018 (2018-10-15), CN , pages 2 - 6, XP009555455, ISSN: 1006-6403, DOI: 10.3969/j.issn.1006-6403.2018.10.001 *

Also Published As

Publication number Publication date
CN115767181A (zh) 2023-03-07

Similar Documents

Publication Publication Date Title
WO2024104307A1 (zh) 直播视频流渲染方法、装置、设备、存储介质及产品
US20240305860A1 (en) Method and apparatus for playing back video at multiple-speed, electronic device and storage medium
WO2021008223A1 (zh) 信息的确定方法、装置及电子设备
CN110898429B (zh) 一种游戏剧情展现方法、装置、电子设备及存储介质
WO2020077914A1 (zh) 图像处理方法、装置、硬件装置
US11785195B2 (en) Method and apparatus for processing three-dimensional video, readable storage medium and electronic device
WO2022100735A1 (zh) 视频处理方法、装置、电子设备及存储介质
US20170195617A1 (en) Image processing method and electronic device
WO2023284437A1 (zh) 媒体文件处理方法、装置、设备、可读存储介质及产品
CN110290398B (zh) 视频下发方法、装置、存储介质及电子设备
WO2024125329A1 (zh) 直播特效渲染方法、装置、设备、可读存储介质及产品
US20240275997A1 (en) Method, apparatus, storage media and program product for video decoding
JP7553582B2 (ja) 画像特殊効果の処理方法及び装置
US20220159197A1 (en) Image special effect processing method and apparatus, and electronic device and computer readable storage medium
US12041379B2 (en) Image special effect processing method, apparatus, and electronic device, and computer-readable storage medium
WO2023273615A1 (zh) 媒体文件处理方法、装置、设备、可读存储介质及产品
CN109547839A (zh) 直播内容的动态预览处理方法及装置、介质、电子设备
CN115761090A (zh) 特效渲染方法、装置、设备、计算机可读存储介质及产品
WO2022242482A1 (zh) 播放控制方法、设备、存储介质及程序产品
CN113839829A (zh) 云游戏延时测试方法、装置、系统及电子设备
WO2024139843A1 (zh) 跨语言的视频处理方法、装置、设备、介质及产品
CN114445600A (zh) 一种特效道具的展示方法、装置、设备及存储介质
WO2023179404A1 (zh) 直播启播方法、设备及程序产品
CN114501041B (zh) 特效显示方法、装置、设备及存储介质
WO2024120396A1 (zh) 视频编码方法、装置、电子设备及存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23890740

Country of ref document: EP

Kind code of ref document: A1