WO2021078182A1 - 一种播放方法以及播放系统 - Google Patents

一种播放方法以及播放系统 Download PDF

Info

Publication number
WO2021078182A1
WO2021078182A1 PCT/CN2020/122672 CN2020122672W WO2021078182A1 WO 2021078182 A1 WO2021078182 A1 WO 2021078182A1 CN 2020122672 W CN2020122672 W CN 2020122672W WO 2021078182 A1 WO2021078182 A1 WO 2021078182A1
Authority
WO
WIPO (PCT)
Prior art keywords
terminal device
cloud server
rendering
motion capture
voice information
Prior art date
Application number
PCT/CN2020/122672
Other languages
English (en)
French (fr)
Inventor
王斌
赵其勇
史浩
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Publication of WO2021078182A1 publication Critical patent/WO2021078182A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/4302Content synchronisation processes, e.g. decoder synchronisation
    • H04N21/4307Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/65Transmission of management data between client and server
    • H04N21/658Transmission by the client directed to the server
    • H04N21/6587Control parameters, e.g. trick play commands, viewpoint selection
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8146Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics
    • H04N21/8153Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics comprising still images, e.g. texture, background image
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8547Content authoring involving timestamps for synchronizing content

Definitions

  • This application relates to the field of communication technology, and specifically to a playback method.
  • Virtual idols are media stars generated by computer technology, such as “Hatsune Miku”, “Luo Tianyi”, “Le Zheng Ling” and so on.
  • Virtual idol live broadcast refers to the use of motion capture devices and sensors placed on the head and limbs of the real person to show the actions of the real person to the virtual character.
  • virtual idols can also communicate and interact with fans in the real world through physical or language methods.
  • the current motion capture system and the rendering host are deployed in the same place, and the two are connected by cables or both are in the same local area network. This is to a certain extent This limits the portability of the motion capture system deployment and increases the deployment cost of the motion capture system.
  • the current virtual anchor can only be watched in the form of live video, which means that all viewers can watch the virtual anchor from the same angle of view controlled by the director, and cannot watch from the angle of view they need, lacking a sense of presence, and the viewer's experience is poor.
  • the embodiment of the present application provides a playback method, so that the motion capture system and the rendering host do not need to be deployed in the same place, and the user can watch the digital image in a holographic manner, and the viewing angle can be adjusted by itself to improve the user's viewing experience.
  • the first aspect of the present application provides a playback method, which may include: a first terminal device receives motion capture data through a wide area network, and the motion capture data is sent by a second terminal device to the first terminal device through a cloud server.
  • the first terminal device receives voice information through the wide area network, the voice information is sent by the second terminal device to the first terminal device through the cloud server, and the first terminal device shares the voice information with the second terminal device and the cloud server in real time.
  • the first terminal device performs augmented reality AR rendering or virtual reality VR rendering on the digital image according to the received motion capture data to obtain the first animation.
  • the first terminal device synchronously plays the first animation and voice information.
  • the second terminal device obtains motion capture data and transmits the motion capture data through the WAN.
  • the first terminal device renders the digital image according to the motion capture data.
  • the second terminal device and the first terminal device do not need to be deployed in the same location.
  • this solution performs VR or AR rendering on the avatar driven by motion capture data in real time, so that the first terminal device can view the digital image in a holographic manner.
  • the user’s viewing angle can be adjusted by himself to enhance the user’s viewing experience.
  • the voice information and the motion capture data may include the same time stamp.
  • Playing the first animation and voice information synchronously by the first terminal device may include: the first terminal device synchronously playing the first animation and voice information according to the same time stamp.
  • a second aspect of the present application provides a playback method, which may include: a cloud server receives first information sent by a first terminal device through a wide area network, where the first information is used to calculate 3D rendering angle parameters of the digital image.
  • the cloud server receives the motion capture data and voice information sent by the second terminal device through the wide area network, and the first terminal device shares the voice information with the second terminal device and the cloud server in real time.
  • the cloud server performs augmented reality AR rendering or virtual reality VR rendering on the digital image according to the motion capture data and the first information to obtain the first animation.
  • the cloud server encodes the first animation and voice information and performs IP encapsulation to obtain the video.
  • the cloud server sends the video to the first terminal device, and the video is decoded and played by the first terminal device.
  • the cloud server obtains the IP video stream after rendering the digital image, and the cloud server sends the encoded IP video stream to the first terminal device.
  • the cloud server obtains the control information of the first terminal device, it can be based on The control information re-renders the digital image and the updated IP video stream of the scene, and sends the updated IP video stream to the first terminal device.
  • the solution provided by the second aspect solves the problem that the model accuracy of the digital image is too high, and the first terminal device cannot meet its rendering requirements.
  • a third aspect of the present application provides a playback method, which may include: a first terminal device sends first information to a cloud server, the first information is used to calculate 3D rendering angle parameters of the digital image, and the 3D rendering angle parameters are used for cloud server combination actions
  • the captured data performs augmented reality AR rendering or virtual reality VR rendering on the digital image to obtain the first animation.
  • the first terminal device receives the video sent by the cloud server, and the video is obtained by the cloud server according to the first animation encoding.
  • the fourth aspect of the present application provides a playback system, which may include: a first terminal device, a second terminal device, and a cloud server.
  • the second terminal device is used to send motion capture data and voice information to the cloud server through the wide area network.
  • the cloud server is used to send the motion capture data and voice information to the first terminal device.
  • the first terminal device is configured to perform augmented reality AR rendering or virtual reality VR rendering on the digital image according to the received motion capture data to obtain the first animation.
  • the first terminal device is also used to synchronously play the first animation and voice information.
  • the fifth aspect of the present application provides a playback system, which may include: a first terminal device, a second terminal device, and a cloud server.
  • the first terminal device is the first terminal device described in the third aspect
  • the second terminal device is the second terminal device described in the second aspect
  • the cloud server is the cloud server described in the second aspect.
  • a sixth aspect of the present application provides a terminal device, which may include: a transceiving unit configured to receive motion capture data through a wide area network, and the motion capture data is sent by the second terminal device to the first terminal device through the cloud server.
  • the transceiver unit is also used to receive voice information through the wide area network.
  • the voice information is sent from the second terminal device to the first terminal device through the cloud server, and the first terminal device shares the voice information with the second terminal device and the cloud server in real time.
  • the processing unit is configured to perform augmented reality AR rendering or virtual reality VR rendering on the digital image according to the motion capture data received by the transceiver unit to obtain the first animation.
  • the playing unit is used to synchronously play the first animation and voice information.
  • a seventh aspect of the present application provides a cloud server, which may include: a transceiving unit configured to receive first information sent by a first terminal device through a wide area network, where the first information is used to calculate 3D rendering angle parameters of the digital image.
  • the transceiver unit is also used to receive the motion capture data and voice information sent by the second terminal device through the wide area network, and the first terminal device shares the voice information with the second terminal device and the cloud server in real time.
  • the processing unit is configured to perform augmented reality AR rendering or virtual reality VR rendering on the digital image according to the motion capture data and the first information received by the transceiver unit to obtain the first animation.
  • the processing unit is also used to encode the first animation and voice information and perform IP encapsulation to obtain the video.
  • the transceiving unit is also used to send a video to the first terminal device, and the video is used to be played by the first terminal device after decoding.
  • An eighth aspect of the present application provides a terminal device, which may include: a transceiver unit, configured to send first information to a cloud server, the first information is used to calculate 3D rendering angle parameters of the digital image, and the 3D rendering angle parameters are used in combination with the cloud server
  • the motion capture data performs augmented reality AR rendering or virtual reality VR rendering on the digital image to obtain the first animation.
  • the transceiver unit is also used to receive a video sent by the cloud server, where the video is obtained by the cloud server according to the first animation encoding.
  • a ninth aspect of the present application provides a computer-readable storage medium, which when an instruction runs on a computer device, causes the computer device to execute the playback method of the first aspect.
  • the tenth aspect of the present application provides a computer-readable storage medium, which when the instructions are executed on the computer device, causes the computer device to execute the playback method as in the second aspect.
  • the eleventh aspect of the present application provides a computer-readable storage medium, which when an instruction runs on a computer device, causes the computer device to execute the playback method as in the third aspect.
  • the twelfth aspect of the present application provides a computer program product, which when running on a computer, enables the computer to execute the playback method as in the first aspect.
  • the thirteenth aspect of the present application provides a computer program product, which when running on a computer, enables the computer to execute the playback method as in the second aspect.
  • the fourteenth aspect of the present application provides a computer program product, which when running on a computer, enables the computer to execute the playback method as in the third aspect.
  • a fifteenth aspect of the present application provides a terminal device, including a memory and a processor, the memory stores code and data, the memory is coupled with the processor, and the processor runs the code in the memory so that all The device executes the playback method of the first aspect or the first possible implementation manner of the first aspect, or executes the playback method of the third aspect.
  • a sixteenth aspect of the present application provides a cloud server, including a memory and a processor, the memory stores code and data, the memory is coupled with the processor, and the processor runs the code in the memory so that all The device executes the playback method of the second aspect.
  • the second terminal device and the first terminal device do not need to be deployed in the same place, and the first terminal device can also view the digital image in a holographic manner, and the user's viewing angle can be adjusted by itself, thereby enhancing the user's viewing experience.
  • FIG. 1 is a schematic diagram of the architecture of a playback system provided by an embodiment of the application
  • FIG. 2 is a schematic diagram of the architecture of another playback system provided by an embodiment of the application.
  • FIG. 3 is a schematic diagram of an application scenario provided by an embodiment of the application.
  • FIG. 4 is a schematic diagram of the hardware structure of a communication device provided by an embodiment of the application.
  • FIG. 5 is a schematic structural diagram of a cloud server provided by an embodiment of this application.
  • FIG. 6 is a block diagram of a partial structure of a mobile phone related to a first terminal device and a second terminal device provided by an embodiment of the application;
  • FIG. 7 is a schematic structural diagram of a first terminal device provided by an embodiment of the application.
  • FIG. 8 is a schematic structural diagram of a second terminal device according to an embodiment of the application.
  • the embodiments of the present application provide a playback method, a playback system, and a storage medium.
  • a first terminal device receives motion capture data through a wide area network, and the motion capture data is sent by a second terminal device to the first terminal device through a cloud server.
  • the first terminal device receives voice information through the wide area network, the voice information is sent by the second terminal device to the first terminal device through the cloud server, and the first terminal device shares the voice information with the second terminal device and the cloud server in real time.
  • the first terminal device performs augmented reality AR rendering or virtual reality VR rendering on the digital image according to the received motion capture data to obtain the first animation.
  • the first terminal device synchronously plays the first animation and voice information.
  • the second terminal device and the first terminal device do not need to be deployed in the same place, and the first terminal device can also view the digital image in a holographic manner, and the user's viewing angle can be adjusted by itself, thereby enhancing the user's viewing experience. Detailed descriptions are given below.
  • the naming or numbering of steps appearing in this application does not mean that the steps in the method flow must be executed in the time/logical sequence indicated by the naming or numbering.
  • the named or numbered process steps can be implemented according to the The technical purpose changes the execution order, as long as the same or similar technical effects can be achieved.
  • the division of modules presented in this application is a logical division. In actual applications, there may be other divisions. For example, multiple modules can be combined or integrated in another system, or some features can be ignored , Or not to execute, in addition, the displayed or discussed mutual coupling or direct coupling or communication connection may be through some ports, and the indirect coupling or communication connection between the modules may be electrical or other similar forms. There are no restrictions in the application.
  • modules or sub-modules described as separate components may or may not be physically separated, may or may not be physical modules, or may be distributed to multiple circuit modules, and some or all of them may be selected according to actual needs. Module to achieve the purpose of this application program.
  • Virtual idols are media stars generated by computer technology, such as “Hatsune Miku”, “Luo Tianyi”, “Le Zheng Ling” and so on.
  • Virtual idol live broadcast refers to the use of motion capture devices and sensors placed on the head and limbs of the real person to show the actions of the real person to the virtual character.
  • virtual idols can also communicate and interact with fans in the real world through physical or language methods.
  • the current motion capture system and the rendering host are deployed in the same place, and the two are connected by cables or both are in the same local area network. This is to a certain extent This limits the portability of the motion capture system deployment and increases the deployment cost of the motion capture system.
  • the current virtual anchor can only be watched in the form of live video, which means that all viewers can watch the virtual anchor from the same angle of view controlled by the director, and cannot watch from the angle of view they need, lacking a sense of presence, and the viewer's experience is poor.
  • the present application provides a playback system, which will be described in detail below.
  • FIG. 1 is a schematic diagram of the architecture of a playback system provided by an embodiment of the application.
  • a playback system provided by an embodiment of the present application may include:
  • the second terminal device acquires motion capture data.
  • the motion capture device is set on the human body to capture the motion of the human body.
  • motion capture devices can be set on the sacrum, left thigh, right thigh, left calf, right calf, left foot, right foot, left shoulder, right shoulder, vest, left upper arm, right upper arm, and left forearm of the human body as required. , Right forearm, left palm, right palm, and back of head.
  • the parts of the human body mentioned above are only for illustrative purposes, and do not represent restrictions on the location of the motion capture device.
  • the motion capture device can be set in other positions of the human body as needed, for example, you can
  • the motion capture device is set on the human face to obtain the expression and mouth shape of the human body.
  • the motion capture device collects the human body motion and sends the motion signal to the second terminal device, and the second terminal device processes the motion signal to obtain the motion capture data. It should be noted that the technical solutions of how to obtain motion capture data in the prior art can be adopted in the embodiments of the present application.
  • the cloud server receives the motion capture data sent by the second terminal device through the wide area network.
  • the cloud server receives the motion capture data sent by the second terminal device through the high-speed Internet.
  • the cloud server receives the motion capture data sent by the second terminal device through the 5th generation (5G) mobile communication system.
  • 5G 5th generation
  • the prior art has to spread the motion capture data in the local area network. Therefore, the current motion capture system and the rendering host are both Deployed in the same place, this solution can transmit motion capture data in the wide area network by introducing a 5G communication system.
  • 6G 6th generation
  • 7G 7th generation
  • the cloud server sends the motion capture data to the first terminal device through the wide area network.
  • the first terminal device may be a terminal device registered in a cloud server.
  • the first terminal device may send a registration request to the cloud server.
  • the registration request carries the identification information of the first terminal device.
  • the cloud server receives the registration request sent by N first terminal devices, and N is a positive integer.
  • the server may send the motion capture data to the N first terminal devices.
  • the first terminal device may also be a terminal device installed with the APP corresponding to the cloud server.
  • the first terminal device pre-downloads and installs the APP corresponding to the cloud server, and the user can manipulate the corresponding APP and The cloud server interacts.
  • the second terminal device, the cloud server, and the first terminal device share voice information in real time.
  • the second terminal device, the cloud server, and the first terminal device can perform voice real-time communication.
  • voice real-time communication between multiple terminal devices there are mature technologies in the prior art, such as the voice communication function of WeChat.
  • the technologies for real-time voice communication between various devices in this application can all be adopted.
  • the first terminal device performs augmented reality (AR) rendering or virtual reality (VR) rendering on the digital image according to the received motion capture data.
  • AR augmented reality
  • VR virtual reality
  • the digital image may be a digital model.
  • the digital image may be the 3D digital model of the aforementioned virtual idol, or the digital model of other cartoon characters, or the digital model of other images that need to interact with people. If the digital avatar is pre-stored on the first terminal device, the locally stored digital avatar can be used directly. If the digital avatar is not pre-stored on the first terminal device, the first terminal device can obtain the digital avatar from the cloud server.
  • VR technology refers to a technology that comprehensively utilizes computer graphics devices and various reality and control interface devices to provide immersion in an interactive three-dimensional environment generated on a computer.
  • AR refers to the application of virtual information to the real world through computer technology. The real environment and virtual objects are superimposed on the same screen or space in real time.
  • the first terminal device performs VR rendering on the digital image according to the received motion capture data.
  • the VR engine calculates 3D rendering angle of view parameters based on the user's action information. For example, the user adjusts the angle of view through the handle, and the VR engine obtains the user's control information on the handle, and calculates the 3D rendering based on the obtained user's control information on the handle. Viewing angle parameters.
  • the 3D rendering engine renders the digital model corresponding to the digital image according to the 3D rendering perspective parameters obtained by the VR engine and the obtained motion capture data, that is, the digital image is rendered.
  • unity 3D can be used to render 3D models of digital images.
  • Unity 3D is a development software, one of the popular 3D game development engines. Of course, unity 3D is just for illustration. Any engine that can achieve 3D rendering is implemented in this application. All examples can be used.
  • the VR engine performs VR rendering on the 3D model rendered by the 3D rendering engine.
  • the first terminal device performs AR rendering on the digital image according to the received motion capture data.
  • the AR engine calculates 3D rendering perspective parameters.
  • the AR engine obtains spatial coordinate information and user gesture control, and calculates 3D rendering perspective parameters according to the acquired spatial coordinate information and user gesture control.
  • the 3D rendering engine renders the digital model corresponding to the digital image according to the 3D rendering perspective parameters obtained by the AR engine and the obtained motion capture data, that is, the digital image is rendered.
  • a scene where the digital avatar is located can also be set, and the scene corresponds to a digital model of a 3D scene, and the digital model of the 3D scene is rendered while the digital avatar is rendered.
  • the digital model of the scene may be stored locally by the first terminal device, or may be obtained by the first terminal device from the cloud server. It should be noted that during the live broadcast process, the first terminal device may also receive an instruction to change the scene sent by the second terminal device through the cloud server, and then when the first terminal device receives the instruction to change the scene, it changes to the specified scene , Correspondingly, render the 3D digital model of the replaced scene.
  • the first terminal device synchronously renders the motion capture data and voice information.
  • the motion capture data and voice information received by the first terminal device respectively include corresponding and same time stamps.
  • the first terminal device may determine the voice information matching the motion capture data according to the corresponding timestamp, or determine the same time stamp according to the corresponding time stamp.
  • the motion capture data matched by the voice information realizes the synchronous rendering of the motion capture data and the voice information, ensuring the synchronization of the digital image motion and voice, that is, the first terminal device synchronously plays the first animation and the voice obtained after VR rendering or AR rendering information.
  • the second terminal device obtains the motion capture data and transmits the motion capture data through the WAN.
  • the first terminal device renders the digital image according to the motion capture data, and the second terminal device and the first terminal device do not need Deployed in the same place.
  • this solution performs VR or AR rendering on the avatar driven by motion capture data in real time, so that the first terminal device is holographic.
  • users can adjust their viewing angles to enhance the user's viewing experience.
  • FIG. 2 is a schematic structural diagram of another playback system provided by an embodiment of the application.
  • a playback system provided by an embodiment of the present application may include:
  • the second terminal device acquires motion capture data.
  • the cloud server receives the motion capture data sent by the second terminal device through the wide area network.
  • Step 201 and step 202 can be understood with reference to steps 101 and 102 in the embodiment corresponding to FIG. 1, and will not be repeated here.
  • the second terminal device, the cloud server, and the first terminal device share voice information in real time.
  • the first terminal device may be a terminal device registered in the cloud server.
  • the second terminal device, the cloud server, and the first terminal device can perform voice real-time communication.
  • voice real-time communication between multiple terminal devices there are mature technologies in the prior art, such as the voice communication function of WeChat.
  • the technologies for real-time voice communication between various devices in this application can all be adopted.
  • the cloud server receives the first information sent by the first terminal device.
  • the first information is control information of the user.
  • the user's gesture information or touch screen control information can be sent to the cloud server, and in the VR scene, it can be sent to the cloud server Send the user's action information on the VR helmet and the user's manipulation information on the VR handle.
  • the cloud server performs AR rendering or VR rendering on the digital image according to the received motion capture data and the first information.
  • the digital image may be a digital model.
  • the digital image may be the 3D digital model of the aforementioned virtual idol, or the digital model of other cartoon characters, or the digital model of other images that need to interact with people. If the digital avatar is pre-stored on the first terminal device, the locally stored digital avatar can be used directly. If the digital avatar is not pre-stored on the first terminal device, the first terminal device can obtain the digital avatar from the cloud server.
  • the cloud server performs VR rendering on the digital image according to the received motion capture data.
  • the VR engine calculates the 3D rendering perspective parameters according to the first information, and the 3D rendering engine renders the digital model corresponding to the digital avatar based on the 3D rendering perspective parameters acquired by the VR engine and the acquired motion capture data, that is, rendering the digital avatar.
  • the VR engine performs VR rendering on the 3D model rendered by the 3D rendering engine.
  • the cloud server performs AR rendering on the digital image according to the received motion capture data.
  • the AR engine calculates the 3D rendering angle of view parameters, for example, the AR engine calculates the 3D rendering angle of view parameters according to the first information.
  • the 3D rendering engine renders the digital model corresponding to the digital image according to the 3D rendering perspective parameters obtained by the AR engine and the obtained motion capture data, that is, the digital image is rendered.
  • a scene where the digital avatar is located can also be set, and the scene corresponds to a digital model of a 3D scene, and the digital model of the 3D scene is rendered while the digital avatar is rendered.
  • the cloud server renders the motion capture data and voice information synchronously.
  • the motion capture data and voice information received by the cloud server respectively include the same corresponding time stamp.
  • the cloud server can determine the voice information matching the motion capture data according to the corresponding time stamp, or determine the voice information matching the voice information according to the corresponding time stamp.
  • Motion capture data realize the synchronous rendering of motion capture data and voice information, and ensure the synchronization of digital image motion and voice.
  • the cloud server sends the encoded IP video stream to the first terminal device.
  • the cloud server performs audio and video encoding on the digital image and voice information driven by the motion capture data in real time, the data obtained after VR rendering or AR rendering of the scene, and then IP packet encapsulation of the encoded data, and the encapsulated IP video flows to the first A terminal device sends.
  • the first terminal device receives the IP video stream sent by the cloud server, and decodes and displays the IP video stream.
  • steps 204 to 206 can be executed repeatedly.
  • the first terminal device sends the first information to the cloud server every preset time, and the cloud server uses the received first information and motion capture data every preset time. Perform AR rendering or VR rendering on the digital image, and send the updated IP video stream to the first terminal device.
  • the cloud server renders the digital image to obtain the IP video stream, and the cloud server sends the encoded IP video stream to the first terminal device.
  • the cloud server obtains the control information of the first terminal device
  • the digital image and scene can be re-rendered to obtain the updated IP video stream, and the updated IP video stream can be sent to the first terminal device.
  • the solution corresponding to Figure 2 solves the problem that the model accuracy of the digital image is too high and the first terminal device cannot meet its rendering requirements.
  • the second terminal device needs to obtain motion capture data and voice information.
  • the first terminal device can be regarded as a whole or as a whole. It is composed of several different devices.
  • the second terminal device includes a voice collection device and a motion capture data collection device.
  • the cloud server receives the motion capture data and voice information sent by the second terminal device.
  • the cloud server can be regarded as a whole, or it can be regarded as composed of several different devices, for example,
  • the cloud server may include a voice interaction server and a full-family data server, which will be described separately below.
  • the user can control the APP on the mobile phone to interact with the cloud server.
  • the user can click to enter the APP and select the live broadcast or video that they want to watch.
  • the user can choose the type of watch that he wants to watch.
  • the user can choose to watch the live broadcast or choose the video or playback.
  • the user can further select the host that he wants to watch, which may include a virtual host, as shown in b in Fig.
  • the mobile phone receives motion capture data about the first digital avatar, the digital model of the first digital avatar, and voice information from the cloud. And scene information. According to the above information, the mobile phone combines the user's operation, such as in the VR scene, the user's operation of the handle and the helmet, VR rendering or AR rendering of the avatar and scene model driven by the motion capture data in real time, so that the user can Immersively watch the live broadcast.
  • the mobile phone can obtain the digital model and scene digital model of a certain or several digital images in advance. For example, as shown in b in Figure 3, after the user enters the APP page, 4 is recommended for the user. If the mobile phone does not pre-store the digital model and scene model information of the digital image corresponding to the four live broadcast rooms, the mobile phone can Get these content from the cloud in advance.
  • the foregoing mainly introduces the solution provided by the embodiment of the present application from the perspective of interaction between the first terminal device, the second terminal device, and the cloud server.
  • the first terminal device, the second terminal device, and the cloud server include hardware structures and/or software modules corresponding to each function.
  • the present application can be implemented in the form of hardware or a combination of hardware and computer software. Whether a certain function is executed by hardware or computer software-driven hardware depends on the specific application and design constraint conditions of the technical solution. Professionals and technicians can use different methods for each specific application to implement the described functions, but such implementation should not be considered beyond the scope of this application.
  • the first terminal device, the second terminal device, and the cloud server in Figures 1 to 3 can be implemented by one physical device, or can be implemented by multiple physical devices, or different in one physical device.
  • Logical function module which is not specifically limited in the embodiment of the present application.
  • the cloud server can be implemented by the communication device in FIG. 4.
  • FIG. 4 shows a schematic diagram of the hardware structure of a communication device provided by an embodiment of the application.
  • the communication device includes at least one processor 401.
  • the communication device may further include: a memory 403, a communication line 402, and at least one communication interface 404.
  • the processor 401 can be a general-purpose central processing unit (central processing unit, CPU), a microprocessor, an application-specific integrated circuit (ASIC), or one or more programs used to control the execution of the program of this application. integrated circuit.
  • CPU central processing unit
  • ASIC application-specific integrated circuit
  • the communication line 402 may include a path to transmit information between the aforementioned components.
  • the communication interface 404 using any device such as a transceiver, is used to communicate with other devices or communication networks, such as Ethernet, radio access network (RAN), wireless local area networks (WLAN), etc. , Or it can be a communication interface between the communication module and other modules.
  • RAN radio access network
  • WLAN wireless local area networks
  • the memory 403 may be a read-only memory (ROM) or other types of static storage devices that can store static information and instructions, random access memory (RAM), or other types that can store information and instructions
  • the dynamic storage device can also be electrically erasable programmable read-only memory (electrically programmable read-only memory, EEPROM), compact disc read-only memory (CD-ROM) or other optical disk storage, Optical disc storage (including compressed optical discs, laser discs, optical discs, digital versatile discs, Blu-ray discs, etc.), magnetic disk storage media or other magnetic storage devices, or can be used to carry or store desired program codes in the form of instructions or data structures and can Any other medium accessed by the computer, but not limited to this.
  • the memory can exist independently and is connected to the processor through the communication line 402. The memory can also be integrated with the processor.
  • the memory 403 is used to store computer-executed instructions for executing the solution of the present application, and the processor 401 controls the execution.
  • the processor 401 is configured to execute computer-executable instructions stored in the memory 403, so as to implement the communication method provided in the following embodiments of the present application.
  • the memory 403 may be coupled with the processor 401 or not.
  • the computer-executable instructions in the embodiments of the present application may also be referred to as application program codes, which are not specifically limited in the embodiments of the present application.
  • the processor 401 may include one or more CPUs.
  • the communication device may include multiple processors.
  • processors can be a single-CPU (single-CPU) processor or a multi-core (multi-CPU) processor.
  • the processor here may refer to one or more devices, circuits, and/or processing cores for processing data (for example, computer program instructions).
  • the communication device may further include an output device and an input device.
  • the output device communicates with the processor 401 and can display information in a variety of ways.
  • the output device may be a liquid crystal display (LCD), a light emitting diode (LED) display device, a cathode ray tube (CRT) display device, or a projector, etc.
  • the input device communicates with the processor 401, and can receive user input in a variety of ways.
  • the input device can be a mouse, a keyboard, a touch screen device, or a sensor device.
  • the embodiment of the present application may divide the cloud server into functional modules according to the foregoing method examples.
  • each functional module may be divided corresponding to each function, or two or more functions may be integrated into one processing module.
  • the above-mentioned integrated modules can be implemented in the form of hardware or software function modules. It should be noted that the division of modules in the embodiments of the present application is illustrative, and is only a logical function division, and there may be other division methods in actual implementation.
  • FIG. 5 shows a schematic structural diagram of a cloud server.
  • the cloud server provided by the embodiment of the present application may include a transceiver unit 501,
  • the transceiver unit 501 is configured to perform step 102 in the embodiment corresponding to FIG. 1, step 103 in the embodiment corresponding to FIG. 1, and step 104 in the embodiment corresponding to FIG. 1. , And/or the transceiving unit 501 is further configured to perform other transceiving steps on the cloud server side in the embodiment corresponding to FIG. 1.
  • the transceiver unit 501 is configured to perform step 202 in the embodiment corresponding to FIG. 2 above, step 203 in the embodiment corresponding to FIG. 2 above, and step 204 in the embodiment corresponding to FIG. 2 above.
  • Step 206 in the embodiment corresponding to FIG. 2 above, step 207 in the embodiment corresponding to FIG. 2 above, and/or the transceiving unit 501 are further configured to perform other transceiving steps on the cloud server side in the embodiment corresponding to FIG. 2.
  • the processing unit 502 is configured to execute step 205 in the embodiment corresponding to FIG. 2, and/or the processing unit 502 is further configured to execute other processing steps on the cloud server side in the embodiment corresponding to FIG. 2.
  • the cloud server is presented in the form of dividing various functional modules in an integrated manner.
  • the embodiment of the present application may also divide the functional modules of the CU network element and the DU network element corresponding to each function, which is not specifically limited in the embodiment of the present application.
  • the first terminal device and the second terminal device involved in this application can represent any applicable terminal device, and can include (or can represent) such as a wireless transmit/receive unit (WTRU), mobile station, and mobile node , Mobile devices, fixed or mobile contract units, pagers, mobile phones, handheld devices, in-vehicle devices, wearable devices, personal digital assistants (PDAs), smartphones, notebook computers, computers, touch screen devices, wireless sensors, or Consumer electronic equipment and other equipment.
  • WTRU wireless transmit/receive unit
  • mobile station mobile station
  • mobile node Mobile devices, fixed or mobile contract units, pagers, mobile phones, handheld devices, in-vehicle devices, wearable devices, personal digital assistants (PDAs), smartphones, notebook computers, computers, touch screen devices, wireless sensors, or Consumer electronic equipment and other equipment.
  • PDAs personal digital assistants
  • smartphones notebook computers, computers, touch screen devices, wireless sensors, or Consumer electronic equipment and other equipment.
  • the first terminal device and the second terminal device are mobile phones as an example for description:
  • FIG. 6 shows a block diagram of a part of the structure of a mobile phone related to the first terminal device and the second terminal device provided by an embodiment of the present invention.
  • the mobile phone includes: a radio frequency (RF) circuit 601, a memory 602, a touch screen 603, a sensor 604, an audio circuit 606, a wireless fidelity (WiFi) module 605, a processor 607, and a power supply 608, etc. part.
  • RF radio frequency
  • WiFi wireless fidelity
  • FIG. 6 does not constitute a limitation on the mobile phone, and may include more or less components than those shown in the figure, or a combination of some components, or different component arrangements.
  • the RF circuit 601 can be used to send and receive information or receive and send signals during a call. In particular, after receiving the downlink information of the base station, it is processed by the processor 607; in addition, , Send the designed uplink data to the base station.
  • the RF circuit 601 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier (LNA), a duplexer, and the like.
  • the RF circuit 601 can also communicate with the network and other devices through wireless communication.
  • the above-mentioned wireless communication can use any communication standard or protocol, including but not limited to Global System of Mobile Communication (GSM), General Packet Radio Service (GPRS), Code Division Multiple Access (Code Division) Multiple Access (CDMA), Wideband Code Division Multiple Access (WCDMA), Long Term Evolution (LTE), 5th generation (5G) mobile communication systems or new radio (new radio, NR) communication system and future mobile communication system, e-mail, short message service (Short Messaging Service, SMS), etc.
  • GSM Global System of Mobile Communication
  • GPRS General Packet Radio Service
  • CDMA Code Division Multiple Access
  • WCDMA Wideband Code Division Multiple Access
  • LTE Long Term Evolution
  • 5G 5th generation
  • e-mail short message service
  • SMS Short Messaging Service
  • the memory 602 may be used to store software programs and modules.
  • the processor 607 executes various functional applications and data processing of the mobile phone by running the software programs and modules stored in the memory 602.
  • the memory 602 may mainly include a program storage area and a data storage area.
  • the program storage area may store an operating system, an application program required by at least one function (such as a sound playback function, an image playback function, etc.), etc.; Data created by the use of mobile phones (such as audio data, phone book, etc.), etc.
  • the memory 602 may include a high-speed random access memory, and may also include a non-volatile memory, such as at least one magnetic disk storage device, a flash memory device, or other volatile solid-state storage devices.
  • the touch screen 603 is also called a touch panel.
  • the touch screen 603 may include a touch-sensitive surface 6031 and a display 6032.
  • the touch-sensitive surface 6031 (such as a touch panel) can collect touch events on or near the mobile phone user (for example, the user uses a finger, a stylus, or any other suitable object on the touch-sensitive surface 6031 or on the touch-sensitive surface). 6031), and send the collected touch information to other devices such as the processor 607.
  • the user's touch event near the touch-sensitive surface 6031 can be called floating touch; floating touch can mean that the user does not need to directly touch the touchpad in order to select, move, or drag a target (such as an icon, etc.), and The user only needs to be located near the first terminal device in order to perform the desired function.
  • the terms "touch”, “contact”, etc. do not imply that they are used to directly contact the touch screen, but to contact near or close to it.
  • the touch-sensitive surface 6031 capable of floating touch can be realized by capacitive, infrared light sensing, ultrasonic waves, and the like.
  • the touch-sensitive surface 6031 may include two parts: a touch detection device and a touch controller.
  • the touch detection device detects the user's touch position, and detects the signal brought by the touch operation, and transmits the signal to the touch controller; the touch controller receives the touch information from the touch detection device and converts it into contact coordinates, and then Sent to the processor 607, and the touch controller can also receive and execute instructions sent by the processor 607.
  • multiple types such as resistive, capacitive, infrared, and surface acoustic wave can be used to implement the touch-sensitive surface 6031.
  • the display (also called display screen) 6032 can be used to display information input by the user or information provided to the user and various menus of the mobile phone.
  • the display 6032 can be configured in the form of a liquid crystal display, an organic light emitting diode, or the like.
  • the touch-sensitive surface 6031 can be overlaid on the display 6032.
  • the touch-sensitive surface 6031 detects a touch event on or near it, it is sent to the processor 607 to determine the type of the touch event, and then the processor 606 can follow the touch event Type provides corresponding visual output on display 6032.
  • the touch-sensitive surface 6031 and the display screen 6032 are used as two independent components to implement the input and output functions of the mobile phone, in some embodiments, the touch-sensitive surface 6031 and the display screen 6032 can be integrated. Realize the input and output functions of the mobile phone. It is understandable that the 603 touch screen is formed by stacking multiple layers of materials.
  • the touch-sensitive surface 6031 may cover the display 6032, and the size of the touch-sensitive surface 6031 is larger than the size of the display screen 6032, so that the display screen 6032 completely covers the touch-sensitive surface 6031.
  • the above-mentioned touch-sensitive surface 6031 may be configured on the front of the mobile phone in the form of a full panel, that is, the user's touch on the front of the mobile phone can be sensed by the mobile phone, so that a full touch experience on the front of the mobile phone can be realized.
  • the touch-sensitive surface 6031 is configured on the front of the mobile phone in the form of a full panel
  • the display 6032 can also be configured on the front of the mobile phone in the form of a full panel, so that a frameless structure can be realized on the front of the mobile phone.
  • the mobile phone may also include at least one sensor 604, such as a light sensor, a motion sensor, and other sensors.
  • the light sensor may include an ambient light sensor and a proximity sensor, where the ambient light sensor can adjust the brightness of the display 6032 according to the brightness of the ambient light, and the proximity sensor can turn off the display 603 and/or the backlight when the mobile phone is moved to the ear .
  • the accelerometer sensor can detect the magnitude of acceleration in various directions (usually three-axis), and can detect the magnitude and direction of gravity when it is stationary.
  • the audio circuit 606, the speaker 6062, and the microphone 6061 can provide an audio interface between the user and the mobile phone.
  • the audio circuit 606 can transmit the electrical signal converted from the received audio data to the speaker 6062, which is converted into a sound signal for output by the speaker 6062; on the other hand, the microphone 6061 converts the collected sound signal into an electrical signal, which is then output by the audio circuit 606. After being received, it is converted into audio data, and then processed by the audio data output processor 607, and sent to, for example, another mobile phone via the RF circuit 601, or the audio data is output to the memory 602 for further processing.
  • WiFi is a short-distance wireless transmission technology.
  • the mobile phone can help users send and receive emails, browse web pages, and access streaming media through the WiFi module 605. It provides users with wireless broadband Internet access.
  • FIG. 6 shows the WiFi module 605, it is understandable that it is not a necessary component of the mobile phone, and can be omitted as needed without changing the essence of the invention.
  • the processor 607 is the control center of the mobile phone. It uses various interfaces and lines to connect various parts of the entire mobile phone. It executes by running or executing software programs and/or modules stored in the memory 602, and calling data stored in the memory 602. Various functions and processing data of the mobile phone can be used to monitor the mobile phone as a whole.
  • the processor 607 may include one or more processing units; preferably, the processor 607 may integrate an application processor and a modem processor, where the application processor mainly processes the operating system, user interface, and application programs, etc. , The modem processor mainly deals with wireless communication. It can be understood that the foregoing modem processor may not be integrated into the processor 607.
  • the mobile phone also includes a power supply 608 (such as a battery) for supplying power to various components.
  • a power supply 608 (such as a battery) for supplying power to various components.
  • the power supply can be logically connected to the processor 607 through a power management system, so that functions such as charging, discharging, and power management can be managed through the power management system.
  • the mobile phone may also include a camera, a Bluetooth module, etc., which will not be repeated here.
  • the embodiment of the present application can divide the first terminal device and the second terminal device into functional modules according to the above method examples.
  • each functional module can be divided corresponding to each function, or two or more functions can be integrated into one.
  • Processing module can be implemented in the form of hardware or software function modules. It should be noted that the division of modules in the embodiments of the present application is illustrative, and is only a logical function division, and there may be other division methods in actual implementation.
  • FIG. 7 shows a schematic structural diagram of a first terminal device.
  • the transceiving unit 701 is configured to perform steps 104 and 105 in the embodiment corresponding to FIG. 1, and/or the transceiving unit 701 is also configured to perform the first terminal device side in the embodiment corresponding to FIG. The other sending and receiving steps.
  • the processing unit 702 is configured to execute step 105 in the embodiment corresponding to FIG. 1, and/or the processing unit 702 is further configured to execute other processing steps on the first terminal device side in the embodiment corresponding to FIG. 1.
  • the transceiving unit 701 is used to perform steps 203, 204, 206, and 207 in the embodiment corresponding to FIG. 2, and/or the transceiving unit 701 is also used to perform the first step in the embodiment corresponding to FIG. Other transceiving steps on the side of a terminal device.
  • the playing unit 703 is configured to execute step 207 in the embodiment corresponding to FIG. 2 above.
  • FIG. 8 shows a schematic structural diagram of a second terminal device.
  • the transceiver unit 801 is configured to perform steps 101 and 102 in the embodiment corresponding to FIG. 1, step 103 in the embodiment corresponding to FIG. 1, and step 103 in the embodiment corresponding to FIG. Step 104 and/or the transceiving unit 801 are further configured to perform other transceiving steps on the second terminal device side in the embodiment corresponding to FIG. 1.
  • the processing unit 802 is configured to execute step 101 in the embodiment corresponding to FIG. 1, and/or the processing unit 802 is further configured to execute other transceiving steps on the second terminal device side in the embodiment corresponding to FIG. 1.
  • the transceiver unit 801 is configured to perform steps 201 and 202 in the embodiment corresponding to FIG. 2 above, step 203 in the embodiment corresponding to FIG. 2 above, and/or the transceiver unit 801 is also configured to Perform other transceiving steps on the side of the second terminal device in the embodiment corresponding to FIG. 2.
  • the processing unit 802 is configured to execute step 201 in the embodiment corresponding to FIG. 2, and/or the processing unit 802 is further configured to execute other processing steps on the second terminal device side in the embodiment corresponding to FIG. 2.
  • the cloud server, the first terminal device, and the second terminal terminal device provided by the embodiments of the present application can be used to perform the above-mentioned positioning method, the technical effects that can be obtained can refer to the above-mentioned method embodiment, which will not be repeated here.
  • the computer program product includes one or more computer instructions.
  • the computer may be a general-purpose computer, a special-purpose computer, a computer network, or other programmable devices.
  • the computer instructions may be stored in a computer-readable storage medium, or transmitted from one computer-readable storage medium to another computer-readable storage medium.
  • the computer instructions may be transmitted from a website, computer, server, or data center. Transmission to another website site, computer, server or data center via wired (such as coaxial cable, optical fiber, digital subscriber line (DSL)) or wireless (such as infrared, wireless, microwave, etc.).
  • wired such as coaxial cable, optical fiber, digital subscriber line (DSL)
  • wireless such as infrared, wireless, microwave, etc.
  • the computer-readable storage medium may be any available medium that can be stored by a computer or a data storage device such as a server or a data center integrated with one or more available media.
  • the usable medium may be a magnetic medium (for example, a floppy disk, a hard disk, and a magnetic tape), an optical medium (for example, a DVD), or a semiconductor medium (for example, a solid state disk (SSD)).
  • the program can be stored in a computer-readable storage medium, and the storage medium can include: ROM, RAM, magnetic disk or CD, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Security & Cryptography (AREA)
  • Databases & Information Systems (AREA)
  • Computer Graphics (AREA)
  • Processing Or Creating Images (AREA)
  • Information Transfer Between Computers (AREA)

Abstract

本申请公开了一种播放方法,包括:第一终端设备通过广域网接收动作捕捉数据,动作捕捉数据由第二终端设备通过云服务器向第一终端设备发送。第一终端设备通过广域网接收语音信息,语音信息由第二终端设备通过云服务器向第一终端设备发送,第一终端设备与第二终端设备以及云服务器实时共享语音信息。第一终端设备根据接收到的动作捕捉数据对数字形象进行增强现实AR渲染或者虚拟现实VR渲染得到第一动画。第一终端设备同步播放第一动画和语音信息。通过本申请提供的方案,第二终端设备和第一终端设备不需要部署在同一个地方,还可以实现第一终端设备以全息方式观看数字形象,用户观看视角自行调整,提升用户的观看体验。

Description

一种播放方法以及播放系统
本申请要求于2019年10月22日提交中国专利局,申请号为201911007650.8、发明名称为“一种播放方法以及播放系统”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及通信技术领域,具体涉及一种播放方法。
背景技术
虚拟偶像是用计算机技术生成的媒体明星,比如“初音未来”、“洛天依”、“乐正绫”等等。2018年,虚拟偶像“绊爱”的走红,衍生出了以直播为表现形式的虚拟偶像。虚拟偶像直播是指借助安置在真实人物的头部与肢体上的动作捕捉设备以及传感器将真实人物动作展现到虚拟角色上。借助于实时动作捕捉的机制,虚拟偶像还可以通过肢体或者语言等方式与现实世界中的粉丝进行交流互动。
由于在直播的场景下,对实时性要求比较高,因此目前动作捕捉系统和渲染主机都是在同一个地方部署,二者通过线缆连接或者二者处于同一个局域网络,这在一定程度上限制了动作捕捉系统部署的便携性,增加了动作捕捉系统的部署成本。此外,目前虚拟主播只能以视频直播的方式观看,这意味着所有观看者都以导播控制的相同视角观看虚拟主播,不能以自己需要的视角观看,缺少临场感,观看者的体验感差。
发明内容
本申请实施例提供一种播放方法,使得动作捕捉系统和渲染主机不需要部署在同一个地方,用户可以以全息方式观看数字形象,观看视角自行调整,提升用户的观看体验。
为达到上述目的,本申请实施例提供如下技术方案:
本申请第一方面提供一种播放方法,可以包括:第一终端设备通过广域网接收动作捕捉数据,动作捕捉数据由第二终端设备通过云服务器向第一终端设备发送。第一终端设备通过广域网接收语音信息,语音信息由第二终端设备通过云服务器向第一终端设备发送,第一终端设备与第二终端设备以及云服务器实时共享语音信息。第一终端设备根据接收到的动作捕捉数据对数字形象进行增强现实AR渲染或者虚拟现实VR渲染得到第一动画。第一终端设备同步播放第一动画和语音信息。由第一方面可知,第二终端设备获取动作捕捉数据,通过广域网传输动作捕捉数据,第一终端设备根据动作捕捉数据对数字形象进行渲染,第二终端设备和第一终端设备不需要部署在同一个地方,此外,相比于现有技术对视频内容进行VR渲染或者AR渲染,本方案对动作捕捉数据实时驱动的虚拟形象进行VR渲染或者AR渲染,实现第一终端设备以全息方式观看数字形象,用户观看视角自行调整,提升用户的观看体验。
可选地,结合上述第一方面,在第一种可能的实现方式中,语音信息和动作捕捉数据可以包括相同的时间戳。第一终端设备同步播放第一动画和语音信息,可以包括:第一终端设备根据相同的时间戳同步播放第一动画和语音信息。
本申请第二方面提供一种播放方法,可以包括:云服务器通过广域网接收第一终端设 备发送的第一信息,第一信息用于计算数字形象的3D渲染视角参数。云服务器通过广域网接收第二终端设备发送的动作捕捉数据和语音信息,第一终端设备与第二终端设备以及云服务器实时共享语音信息。云服务器根据动作捕捉数据和第一信息对数字形象进行增强现实AR渲染或者虚拟现实VR渲染得到第一动画。云服务器对第一动画和语音信息进行编码并进行IP封装得到视频。云服务器向第一终端设备发送视频,视频用于第一终端设备解码后播放。由第二方面可知,云服务器对数字形象进行渲染后得到IP视频流,云服务器向第一终端设备发送编码后的IP视频流,当云服务器获取到第一终端设备的控制信息后,可以根据该控制信息,重新渲染数字形象以及场景得到更新后的IP视频流,并将更新后的IP视频流向第一终端设备发送。第二方面提供的方案解决了数字形象的模型精度过高,第一终端设备无法满足其渲染要求的问题。
本申请第三方面提供一种播放方法,可以包括:第一终端设备向云服务器发送第一信息,第一信息用于计算数字形象的3D渲染视角参数,3D渲染视角参数用于云服务器结合动作捕捉数据对数字形象进行增强现实AR渲染或者虚拟现实VR渲染得到第一动画。第一终端设备接收云服务器发送的视频,视频为云服务器根据第一动画编码获得。
本申请第四方面提供一种播放系统,可以包括:第一终端设备、第二终端设备以及云服务器。第二终端设备,用于通过广域网向云服务器发送动作捕捉数据和语音信息。云服务器,用于将动作捕捉数据和语音信息向第一终端设备发送。第一终端设备,用于根据接收到的动作捕捉数据对数字形象进行增强现实AR渲染或者虚拟现实VR渲染得到第一动画。第一终端设备,还用于同步播放第一动画和语音信息。
本申请第五方面提供一种播放系统,可以包括:第一终端设备、第二终端设备以及云服务器。第一终端设备为第三方面所描述的第一终端设备,第二终端设备为第二方面所描述的第二终端设备,云服务器为第二方面所描述的云服务器。
本申请第六方面提供一种终端设备,可以包括:收发单元,用于通过广域网接收动作捕捉数据,动作捕捉数据由第二终端设备通过云服务器向第一终端设备发送。收发单元,还用于通过广域网接收语音信息,语音信息由第二终端设备通过云服务器向第一终端设备发送,第一终端设备与第二终端设备以及云服务器实时共享语音信息。处理单元,用于根据收发单元接收到的动作捕捉数据对数字形象进行增强现实AR渲染或者虚拟现实VR渲染得到第一动画。播放单元,用于同步播放第一动画和语音信息。
本申请第七方面提供一种云服务器,可以包括:收发单元,用于通过广域网接收第一终端设备发送的第一信息,第一信息用于计算数字形象的3D渲染视角参数。收发单元,还用于通过广域网接收第二终端设备发送的动作捕捉数据和语音信息,第一终端设备与第二终端设备以及云服务器实时共享语音信息。处理单元,用于根据收发单元接收到的动作捕捉数据和第一信息对数字形象进行增强现实AR渲染或者虚拟现实VR渲染得到第一动画。处理单元,还用于对第一动画和语音信息进行编码并进行IP封装得到视频。收发单元,还用于向第一终端设备发送视频,视频用于第一终端设备解码后播放。
本申请第八方面提供一种终端设备,可以包括:收发单元,用于向云服务器发送第一信息,第一信息用于计算数字形象的3D渲染视角参数,3D渲染视角参数用于云服务器结 合动作捕捉数据对数字形象进行增强现实AR渲染或者虚拟现实VR渲染得到第一动画。收发单元,还用于接收云服务器发送的视频,视频为云服务器根据第一动画编码获得。
本申请第九方面提供一种计算机可读存储介质,当指令在计算机装置上运行时,使得计算机装置执行第一方面的播放方法。
本申请第十方面提供一种计算机可读存储介质,当指令在计算机装置上运行时,使得计算机装置执行如第二方面的播放方法。
本申请第十一方面提供一种计算机可读存储介质,当指令在计算机装置上运行时,使得计算机装置执行如第三方面的播放方法。
本申请第十二方面提供一种计算机程序产品,当其在计算机上运行时,使得计算机可以执行如第一方面的播放方法。
本申请第十三方面提供一种计算机程序产品,当其在计算机上运行时,使得计算机可以执行如第二方面的播放方法。
本申请第十四方面提供一种计算机程序产品,当其在计算机上运行时,使得计算机可以执行如第三方面的播放方法。
本申请第十五方面提供一种终端设备,包括存储器、处理器,所述存储器中存储代码和数据,所述存储器与所述处理器耦合,所述处理器运行所述存储器中的代码使得所述设备执行如第一方面或第一方面第一种可能实现方式的播放方法,或者执行如第三方面的播放方法。
本申请第十六方面提供一种云服务器,包括存储器、处理器,所述存储器中存储代码和数据,所述存储器与所述处理器耦合,所述处理器运行所述存储器中的代码使得所述设备执行如第二方面的播放方法。
通过本申请提供的方案,第二终端设备和第一终端设备不需要部署在同一个地方,还可以实现第一终端设备以全息方式观看数字形象,用户观看视角自行调整,提升用户的观看体验。
附图说明
图1为本申请实施例提供的一种播放系统的架构示意图;
图2为本申请实施例提供的另一种播放系统的架构示意图;
图3为本申请实施例提供的一种应用场景示意图;
图4为本申请实施例提供的通信设备的硬件结构的一个示意图;
图5为本申请实施例提供的一种云服务器的结构示意图;
图6为本申请实施例提供的第一终端设备、第二终端设备相关的手机的部分结构的框图;
图7为本申请实施例提供的第一终端设备的结构示意图;
图8为本申请实施例提供的第二终端设备的结构示意图。
具体实施方式
下面结合附图,对本申请的实施例进行描述,显然,所描述的实施例仅仅是本申请一 部分的实施例,而不是全部的实施例。本领域普通技术人员可知,随着技术的发展和新场景的出现,本申请实施例提供的技术方案对于类似的技术问题,同样适用。
本申请实施例提供一种播放方法、播放系统及存储介质,第一终端设备通过广域网接收动作捕捉数据,动作捕捉数据由第二终端设备通过云服务器向第一终端设备发送。第一终端设备通过广域网接收语音信息,语音信息由第二终端设备通过云服务器向第一终端设备发送,第一终端设备与第二终端设备以及云服务器实时共享语音信息。第一终端设备根据接收到的动作捕捉数据对数字形象进行增强现实AR渲染或者虚拟现实VR渲染得到第一动画。第一终端设备同步播放第一动画和语音信息。通过本申请提供的方案,第二终端设备和第一终端设备不需要部署在同一个地方,还可以实现第一终端设备以全息方式观看数字形象,用户观看视角自行调整,提升用户的观看体验。以下分别进行详细说明。
本申请的说明书和权利要求书及上述附图中的术语“第一”、“第二”等是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。应该理解这样使用的数据在适当情况下可以互换,以便这里描述的实施例能够以除了在这里图示或描述的内容以外的顺序实施。此外,术语“包括”和“具有”以及他们的任何变形,意图在于覆盖不排他的包含,例如,包含了一系列步骤或模块的过程、方法、系统、产品或设备不必限于清楚地列出的那些步骤或模块,而是可包括没有清楚地列出的或对于这些过程、方法、产品或设备固有的其它步骤或模块。在本申请中出现的对步骤进行的命名或者编号,并不意味着必须按照命名或者编号所指示的时间/逻辑先后顺序执行方法流程中的步骤,已经命名或者编号的流程步骤可以根据要实现的技术目的变更执行次序,只要能达到相同或者相类似的技术效果即可。本申请中所出现的模块的划分,是一种逻辑上的划分,实际应用中实现时可以有另外的划分方式,例如多个模块可以结合成或集成在另一个系统中,或一些特征可以忽略,或不执行,另外,所显示的或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些端口,模块之间的间接耦合或通信连接可以是电性或其他类似的形式,本申请中均不作限定。并且,作为分离部件说明的模块或子模块可以是也可以不是物理上的分离,可以是也可以不是物理模块,或者可以分布到多个电路模块中,可以根据实际的需要选择其中的部分或全部模块来实现本申请方案的目的。
需要说明的是,本申请实施例中,名词“网络”和“系统”经常交替使用,但本领域的技术人员可以理解其含义。信息(information),信号(signal),消息(message)有时可以混用,应当指出的是,在不强调其区别时,其所要表达的含义是一致的。
还需要说明的是,在本申请实施例中,“上报”和“反馈”以及“响应”经常交替使用,但本领域的技术人员可以理解其含义。因此,在本申请实施例中,在不强调其区别时,其所要表达的含义是一致的。
还需要说明的是,在本申请实施例中,“数据”和“信息”经常交替使用,但本领域的技术人员可以理解其含义。因此,在本申请实施例中,在不强调其区别时,其所要表达的含义是一致的。
虚拟偶像是用计算机技术生成的媒体明星,比如“初音未来”、“洛天依”、“乐正绫”等等。2018年,虚拟偶像“绊爱”的走红,衍生出了以直播为表现形式的虚拟偶像。 虚拟偶像直播是指借助安置在真实人物的头部与肢体上的动作捕捉设备以及传感器将真实人物动作展现到虚拟角色上。借助于实时动作捕捉的机制,虚拟偶像还可以通过肢体或者语言等方式与现实世界中的粉丝进行交流互动。
由于在直播的场景下,对实时性要求比较高,因此目前动作捕捉系统和渲染主机都是在同一个地方部署,二者通过线缆连接或者二者处于同一个局域网络,这在一定程度上限制了动作捕捉系统部署的便携性,增加了动作捕捉系统的部署成本。此外,目前虚拟主播只能以视频直播的方式观看,这意味着所有观看者都以导播控制的相同视角观看虚拟主播,不能以自己需要的视角观看,缺少临场感,观看者的体验感差。为了解决上述问题,本申请提供一种播放系统,以下进行详细的说明。
在本申请提供的方案中,通过引入云服务器,动作捕捉系统和渲染主机不需要部署在同一个地方,其中一种方式,渲染过程在用户的终端设置上进行,另一种方式,渲染过程在云服务器上进行,以下将针对这两种方式分别进行说明。
图1为本申请实施例提供的一种播放系统的架构示意图。
如图1所示,本申请实施例提供的一种播放系统,可以包括:
101、第二终端设备获取动作捕捉数据。
将动作捕捉设备设于人体上,以采集人体的动作。比如可以将动作捕捉设备按照需求分别设置于人体的骶骨、左大腿、右大腿、左小腿、右小腿、左脚、右脚、左肩膀、右肩膀、背心、左上臂、右上臂、左小臂、右小臂、左手掌、右手掌、后脑勺中的任意几个部位上。当然上述提到的几个人体部分仅仅是为了举例说明,并不代表对动作捕捉设备设置位置的限制,在实际应用场景中,可以根据需要将动作捕捉设备设置在人体的其他位置,比如还可以将动作捕捉设备设置在人脸上,以获取人体的表情以及嘴型。动作捕捉设备采集人体动作后将动作信号发送至第二终端设备,第二终端设备对该动作信号进行处理得到动作捕捉数据。需要说明的是,现有技术中关于如何获取动作捕捉数据的技术方案,本申请实施例均可以采用。
102、云服务器通过广域网接收第二终端设备发送的动作捕捉数据。
云服务器通过高速互联网接收第二终端设备发送的动作捕捉数据。比如,云服务器通过第五代(5th generation,5G)移动通信系统接收第二终端设备发送的动作捕捉数据。相比于现有技术为了降低动作捕捉流程和渲染流程之间的时延,实现数字形象直播的顺畅,现有技术要在局域网中传播动作捕捉数据,所以,目前动作捕捉系统和渲染主机都是在同一个地方部署,本方案中通过引入5G通信系统,可以在广域网中传输动作捕捉数据。此外,需要说明的是,随着技术的发展,将来可能会发展到第六代(6th generation,6G)移动通信系统,第七代(7th generation,7G)移动通信系统等等,本申请实施例均可以采用。
103、云服务器通过广域网向第一终端设备发送动作捕捉数据。
在本申请实施例中,第一终端设备可以是在云服务器中注册的终端设备。比如第一终端设备可以向云服务器发送注册请求,该注册请求中携带第一终端设备的标识信息,比如云服务器接收到了N个第一终端设备发送的注册请求,N为正整数,则当云服务器接收到第二终端设备发送的动作捕捉数据后,可以将该动作捕捉数据向该N个第一终端设备发送。 或者,第一终端设备也可以是安装了云服务器对应的APP的终端设备,比如第一终端设备预先下载并安装了云服务器对应的APP,则用户可以通过操控第二终端终端上对应的APP和云服务器进行交互。
104、第二终端设备、云服务器以及第一终端设备实时共享语音信息。
第二终端设备、云服务器以及第一终端设备可以进行语音实时通信,关于多个终端设备之间进行语音实时通信,现有技术中已经有成熟的技术,比如微信的语音通信功能,现有技术中关于各个设备之间进行语音实时通信的技术本申请均可以采用。
105、第一终端设备根据接收到的动作捕捉数据对数字形象进行增强现实(augmented reality,AR)渲染或者虚拟现实(virtual reality,VR)渲染。
本申请中,该数字形象可以是一个数字模型。该数字形象可能是上述提到的虚拟偶像的3D数字模型,或者是其他卡通人物的数字模型,或者是其他需要与人互动的形象的数字模型。若第一终端设备上预先存储了该数字形象,则可以直接使用本地存储数字形象,若第一终端设备上没有预先存储该数字形象,第一终端设备可以从云服务器中获取该数字形象。
VR技术是指综合利用计算机图形装置和各种现实及控制等接口设备,在计算机上生成的、可交互的三维环境中提供沉浸感觉的技术。AR是指通过电脑技术,将虚拟的信息应用到真实世界,真实的环境和虚拟的物体实时地叠加到了同一个画面或空间同时存在。
在一个具体的实施方式中,第一终端设备根据接收到的动作捕捉数据对数字形象进行VR渲染。具体的,VR引擎根据用户的动作信息计算3D渲染视角参数,比如用户通过手柄进行视角的调控,则VR引擎获取用户对手柄的控制信息,并根据获取到的用户对手柄的控制信息计算3D渲染视角参数。3D渲染引擎根据VR引擎获取到的3D渲染视角参数,以及获取到的动作捕捉数据渲染数字形象对应的数字模型,即渲染数字形象。比如可以采用unity 3D渲染数字形象的3D模型,unity 3D是一个开发软件,当今流行的3D游戏开发引擎之一,当然unity 3D只是为了为了举例说明,任意一个可以实现3D渲染的引擎,本申请实施例均可以采用。VR引擎对3D渲染引擎渲染后得到的3D模型进行VR渲染。
在一个具体的实施方式中,第一终端设备根据接收到的动作捕捉数据对数字形象进行AR渲染。具体的,AR引擎计算3D渲染视角参数,比如AR引擎获取空间坐标信息以及用户的手势控制,并根据获取到的空间坐标信息以及用户的手势控制计算3D渲染视角参数。3D渲染引擎根据AR引擎获取到的3D渲染视角参数,以及获取到的动作捕捉数据渲染数字形象对应的数字模型,即渲染数字形象。
在一个具体的实施方式中,还可以设置数字形象所在的场景,该场景对应一个3D场景的数字模型,在对数字形象进行渲染的同时对该3D场景的数字模型进行渲染。场景的数字模型可以是第一终端设备本地保存的,也可以是第一终端设备从云服务器中获取的。需要说明的是,在直播过程中,第一终端设备还可以接收第二终端设备通过云服务器发送的更换场景的指令,则当第一终端设备接收到更换场景的指令后,更换到指定的场景,对应的,对更换后的场景的3D数字模型进行渲染。
此外,为了保证数字形象的动作和语音的同步,在一个具体的实施方式中,还包括106、 第一终端设备对动作捕捉数据和语音信息同步渲染。
第一终端设备接收到的动作捕捉数据和语音信息分别包括对应的相同的时间戳,第一终端设备可以根据对应的时间戳确定与动作捕捉数据匹配的语音信息,或者根据对应的时间戳确定与语音信息匹配的动作捕捉数据,实现动作捕捉数据和语音信息的同步渲染,保证数字形象动作和语音的同步,即第一终端设备同步播放VR渲染或者AR渲染后得到的第一动画和所述语音信息。
由图1对应的实施例可知,第二终端设备获取动作捕捉数据,通过广域网传输动作捕捉数据,第一终端设备根据动作捕捉数据对数字形象进行渲染,第二终端设备和第一终端设备不需要部署在同一个地方,此外,相比于现有技术对视频内容进行VR渲染或者AR渲染,本方案对动作捕捉数据实时驱动的虚拟形象进行VR渲染或者AR渲染,实现第一终端设备以全息方式观看数字形象,用户观看视角自行调整,提升用户的观看体验。
图2为本申请实施例提供的另一种播放系统的架构示意图。
如图2所示,本申请实施例提供的一种播放系统,可以包括:
201、第二终端设备获取动作捕捉数据。
202、云服务器通过广域网接收第二终端设备发送的动作捕捉数据。
步骤201和步骤202可以参考图1对应的实施例中的步骤101和102进行理解,此处不再重复赘述。
203、第二终端设备、云服务器以及第一终端设备实时共享语音信息。
第一终端设备可以是在云服务器中注册的终端设备。第二终端设备、云服务器以及第一终端设备可以进行语音实时通信,关于多个终端设备之间进行语音实时通信,现有技术中已经有成熟的技术,比如微信的语音通信功能,现有技术中关于各个设备之间进行语音实时通信的技术本申请均可以采用。
204、云服务器接收第一终端设备发送的第一信息。
该第一信息为用户的控制信息。比如获取用户对观看角度的调整信息,对画面的缩放信息等等,举例说明,在AR场景中,可以向云服务器发送用户的手势信息或者触屏控制信息,在VR场景中,可以向云服务器发送用户对VR头盔的动作信息以及用户对VR手柄的操控信息。
205、云服务器根据接收到的动作捕捉数据以及第一信息对数字形象进行AR渲染或者VR渲染。
本申请中,该数字形象可以是一个数字模型。该数字形象可能是上述提到的虚拟偶像的3D数字模型,或者是其他卡通人物的数字模型,或者是其他需要与人互动的形象的数字模型。若第一终端设备上预先存储了该数字形象,则可以直接使用本地存储数字形象,若第一终端设备上没有预先存储该数字形象,第一终端设备可以从云服务器中获取该数字形象。
在一个具体的实施方式中,云服务器根据接收到的动作捕捉数据对数字形象进行VR渲染。具体的,VR引擎根据第一信息计算3D渲染视角参数,3D渲染引擎根据VR引擎获取到的3D渲染视角参数,以及获取到的动作捕捉数据渲染数字形象对应的数字模型,即渲染 数字形象。VR引擎对3D渲染引擎渲染后得到的3D模型进行VR渲染。
在一个具体的实施方式中,云服务器根据接收到的动作捕捉数据对数字形象进行AR渲染。具体的,AR引擎计算3D渲染视角参数,比如AR引擎根据第一信息计算3D渲染视角参数。3D渲染引擎根据AR引擎获取到的3D渲染视角参数,以及获取到的动作捕捉数据渲染数字形象对应的数字模型,即渲染数字形象。
在一个具体的实施方式中,还可以设置数字形象所在的场景,该场景对应一个3D场景的数字模型,在对数字形象进行渲染的同时对该3D场景的数字模型进行渲染。
此外,为了保证数字形象的动作和语音的同步,在一个具体的实施方式中,云服务器对动作捕捉数据和语音信息同步渲染。
云服务器接收到的动作捕捉数据和语音信息分别包括对应的相同的时间戳,云服务器可以根据对应的时间戳确定与动作捕捉数据匹配的语音信息,或者根据对应的时间戳确定与语音信息匹配的动作捕捉数据,实现动作捕捉数据和语音信息的同步渲染,保证数字形象动作和语音的同步。
206、云服务器向第一终端设备发送编码后的IP视频流。
云服务器对动作捕捉数据实时驱动的数字形象以及语音信息,场景进行VR渲染或者AR渲染后得到的数据进行音视频编码,再对编码后的数据进行IP包封装,将封装后的IP视频流向第一终端设备发送。
207、第一终端设备接收云服务器发送的IP视频流,并对该IP视频流进行解码并显示。
需要说明的是,步骤204至步骤206可以重复执行,比如第一终端设备每隔预设时间向云服务器发送第一信息,云服务器每隔预设时间根据接收到的第一信息以及动作捕捉数据对数字形象进行AR渲染或者VR渲染,并将更新后的IP视频流向第一终端设备发送。
由图2对应的实施例可知,云服务器对数字形象进行渲染后得到IP视频流,云服务器向第一终端设备发送编码后的IP视频流,当云服务器获取到第一终端设备的控制信息后,可以根据该控制信息,重新渲染数字形象以及场景得到更新后的IP视频流,并将更新后的IP视频流向第一终端设备发送。图2对应的方案,解决了数字形象的模型精度过高,第一终端设备无法满足其渲染要求的问题。
在上述图1和图2对应的实施例中,第二终端设备要获取动作捕捉数据以及语音信息,在一些具体的应用场景中,第一终端设端可以看做是一个整体,也可以看做是由几个不同的设备组成,比如第二终端设备包括语音采集设备和动作捕捉数据采集设备。此外,云服务器接收第二终端设备发送的动作捕捉数据和语音信息,在一些具体的应用场景中,云服务器可以看做是一个整体,也可以看做是由几个不同的设备组成,比如,云服务器可以包括语音交互服务器和全系数据服务器,以下分别进行说明。
如3所示,为本申请实施例提供的一种应用场景示意图。
以手机为例进行说明,用户可以操控手机上的APP与云服务器进行交互。如图3中的a所示,用户可以点击进入APP,选择想要观看的直播或者视频。比如,用户点击APP后进入到如图3中的b所示的画面,用户可以选择想要观看的种类,比如用户可以选择观看直播也可以选择视频或者回放,假设用户选择观看直播,在直播界面中,用户可以进一步选 择想要观看的主播,其中可以包括虚拟主播,如图3中的b示意性的展示了第一直播间,第二直播间,第三直播间以及第四直播间。假设用户选择进入了第一直播间,第一直播间的主播为第一数字形象,则手机接收来自云端的关于该第一数字形象的动作捕捉数据、该第一数字形象的数字模型、语音信息以及场景信息,根据上述信息,手机结合用户的操作,比如在VR场景中,用户对手柄以及头盔的操作,对动作捕捉数据实时驱动的虚拟形象以及场景模型进行VR渲染或者AR渲染,使得用户可以沉浸式的观看直播。需要说明的是,如果手机本地保存了第一数字形象的数字模型和场景对应的数字模型的相关数据,也可以不从云端获取,这在图1和图2对应的实施例中已经进行了说明,此处不再重复赘述。此外,需要说明的是,手机可以提前获取某一个或者某几个数字形象的数字模型以及场景数字模型,比如如图3中的b所示,在用户进入到APP页面后,为用户推荐了4个直播间,第一直播间,第二直播间,第三直播间以及第四直播间,则如果手机上没有预先存储该四个直播间对应的数字形象的数字模型和场景模型信息,手机可以预先从云端获取这些内容,换句话说,不用等到用户进入直播间后再从云端获取数字形象以及场景对应的数字模型的数据,而是提前从云端获取这些内容。此外,需要说明的,VR渲染或者AR渲染的过程可以在云服务器上进行,这在图1和图2对应的实施例中已经进行了描述,此处不再重复赘述。
上述主要从第一终端设备、第二终端设备以及云服务器之间交互的角度对本申请实施例提供的方案进行了介绍。可以理解的是,上述第一终端设备、第二终端设备以及云服务器为了实现上述功能,其包含了执行各个功能相应的硬件结构和/或软件模块。本领域技术人员应该很容易意识到,结合本文中所公开的实施例描述的各示例的模块及算法步骤,本申请能够以硬件或硬件和计算机软件的结合形式来实现。某个功能究竟以硬件还是计算机软件驱动硬件的方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。
从硬件结构上来描述,图1至图3中第一终端设备、第二终端设备以及云服务器可以由一个实体设备实现,也可以由多个实体设备共同实现,还可以是一个实体设备内的不同逻辑功能模块,本申请实施例对此不作具体限定。
例如,云服务器可以通过图4中的通信设备来实现。图4所示为本申请实施例提供的通信设备的硬件结构示意图。该通信设备包括至少一个处理器401。可选地,该通信设备还可以包括:存储器403,通信线路402以及至少一个通信接口404。
处理器401可以是一个通用中央处理器(central processing unit,CPU),微处理器,特定应用集成电路(application-specific integrated circuit,ASIC),或一个或多个用于控制本申请方案程序执行的集成电路。
通信线路402可包括一通路,在上述组件之间传送信息。
通信接口404,使用任何收发器一类的装置,用于与其他设备或通信网络通信,如以太网,无线接入网(radio access network,RAN),无线局域网(wireless local area networks,WLAN)等,或者也可以是通信模块和其他模块之间的通信接口。
存储器403可以是只读存储器(read-only memory,ROM)或可存储静态信息和指令的其他类型的静态存储设备,随机存取存储器(random access memory,RAM)或者可存储信息和指令的其他类型的动态存储设备,也可以是电可擦可编程只读存储器(electrically er服务器able programmable read-only memory,EEPROM)、只读光盘(compact disc read-only memory,CD-ROM)或其他光盘存储、光碟存储(包括压缩光碟、激光碟、光碟、数字通用光碟、蓝光光碟等)、磁盘存储介质或者其他磁存储设备、或者能够用于携带或存储具有指令或数据结构形式的期望的程序代码并能够由计算机存取的任何其他介质,但不限于此。存储器可以是独立存在,通过通信线路402与处理器相连接。存储器也可以和处理器集成在一起。
其中,存储器403用于存储执行本申请方案的计算机执行指令,并由处理器401来控制执行。处理器401用于执行存储器403中存储的计算机执行指令,从而实现本申请下述实施例提供的通信的方法。存储器403可以跟处理器401耦合在一起,也可以不耦合在一起。
可选的,本申请实施例中的计算机执行指令也可以称之为应用程序代码,本申请实施例对此不作具体限定。
在具体实现中,作为一种实施例,处理器401可以包括一个或多个CPU。
在具体实现中,作为一种实施例,通信设备可以包括多个处理器。这些处理器中的每一个可以是一个单核(single-CPU)处理器,也可以是一个多核(multi-CPU)处理器。这里的处理器可以指一个或多个设备、电路、和/或用于处理数据(例如计算机程序指令)的处理核。
在具体实现中,作为一种实施例,通信设备还可以包括输出设备和输入设备。输出设备和处理器401通信,可以以多种方式来显示信息。例如,输出设备可以是液晶显示器(liquid crystal display,LCD),发光二级管(light emitting diode,LED)显示设备,阴极射线管(cathode ray tube,CRT)显示设备,或投影仪(projector)等。输入设备和处理器401通信,可以以多种方式接收用户的输入。例如,输入设备可以是鼠标、键盘、触摸屏设备或传感设备等。
本申请实施例可以根据上述方法示例对云服务器进行功能模块的划分,例如,可以对应各个功能划分各个功能模块,也可以将两个或两个以上的功能集成在一个处理模块中。上述集成的模块既可以采用硬件的形式实现,也可以采用软件功能模块的形式实现。需要说明的是,本申请实施例中对模块的划分是示意性的,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式。
比如,以采用集成的方式划分各个功能模块的情况下,图5示出了一种云服务器的结构示意图。
如图5所示,本申请实施例提供的云服务器可以包括收发单元501,
在一个具体的实施方式中,收发单元501,用于执行上述图1对应的实施例中的步骤102,上述图1对应的实施例中的步骤103,上述图1对应的实施例中的步骤104,和/或收发单元501 还用于执行图1对应的实施例中云服务器侧的其他收发步骤。
在一个具体的实施方式中,收发单元501,用于执行上述图2对应的实施例中的步骤202,上述图2对应的实施例中的步骤203,上述图2对应的实施例中的步骤204,上述图2对应的实施例中的步骤206,上述图2对应的实施例中的步骤207,和/或收发单元501还用于执行图2对应的实施例中云服务器侧的其他收发步骤。处理单元502,用于执行上述图2对应的实施例中的步骤205,和/或处理单元502还用于执行图2对应的实施例中云服务器侧的其他处理步骤。
上述实施例中,云服务器以采用集成的方式划分各个功能模块的形式来呈现。当然,本申请实施例也可以对应各个功能划分CU网元和DU网元的各个功能模块,本申请实施例对此不作具体限定。
本申请所涉及的第一终端设备、第二终端设备可以表示任意适用的终端设备,可以包括(或可以表示)诸如无线发送/接收单元(wireless transmit/receive unit,WTRU)、移动站、移动节点、移动设备、固定或移动签约单元、寻呼机、移动电话、手持设备、车载设备、可穿戴设备、掌上电脑(personal digital assistant,PDA)、智能手机、笔记本型电脑、计算机、触摸屏设备、无线传感器或消费电子设备等设备。下面,以第一终端设备、第二终端设备为手机为例进行说明:
图6示出的是与本发明实施例提供的第一终端设备、第二终端设备相关的手机的部分结构的框图。参考图6,手机包括:射频(Radio Frequency,RF)电路601、存储器602、触摸屏603、传感器604、音频电路606、无线保真(wireless fidelity,WiFi)模块605、处理器607、以及电源608等部件。本领域技术人员可以理解,图6中示出的手机结构并不构成对手机的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件布置。
下面结合图6对手机的各个构成部件进行具体的介绍:RF电路601可用于收发信息或通话过程中信号的接收和发送,特别地,将基站的下行信息接收后,给处理器607处理;另外,将设计上行的数据发送给基站。通常,RF电路601包括但不限于天线、至少一个放大器、收发信机、耦合器、低噪声放大器(Low Noise Amplifier,LNA)、双工器等。此外,RF电路601还可以通过无线通信与网络和其他设备通信。上述无线通信可以使用任一通信标准或协议,包括但不限于全球移动通讯系统(Global System of Mobile communication,GSM)、通用分组无线服务(General Packet Radio Service,GPRS)、码分多址(Code Division Multiple Access,CDMA)、宽带码分多址(Wideband Code Division Multiple Access,WCDMA)、长期演进(Long Term Evolution,LTE)、第五代(5th generation,5G)移动通信系统或新无线(new radio,NR)通信系统以及未来的移动通信系统,电子邮件、短消息服务(Short Messaging Service,SMS)等。
存储器602可用于存储软件程序以及模块,处理器607通过运行存储在存储器602的软件程序以及模块,从而执行手机的各种功能应用以及数据处理。存储器602可主要包括存储程序区和存储数据区,其中,存储程序区可存储操作系统、至少一个功能所需的应用程序(比如声音播放功能、图像播放功能等)等;存储数据区可存储根据手机的使用所创 建的数据(比如音频数据、电话本等)等。此外,存储器602可以包括高速随机存取存储器,还可以包括非易失性存储器,例如至少一个磁盘存储器件、闪存器件、或其他易失性固态存储器件。
触摸屏603,也称为触控面板,触摸屏603可以包括触敏表面6031和显示器6032。其中,触敏表面6031(例如触控面板)可采集手机的用户在其上或附近的触摸事件(比如用户使用手指、触控笔等任何适合的物体在触敏表面6031上或在触敏表面6031附近的操作),并将采集到的触摸信息发送给其他器件例如处理器607。其中,用户在触敏表面6031附近的触摸事件可以称之为悬浮触控;悬浮触控可以是指,用户无需为了选择、移动或拖动目标(例如图标等)而直接接触触控板,而只需用户位于第一终端设备附近以便执行所想要的功能。在悬浮触控的应用场景下,术语“触摸”、“接触”等不会暗示用于直接接触触摸屏,而是在其附近或接近的接触。能够进行悬浮触控的触敏表面6031可以采用电容式、红外光感以及超声波等实现。触敏表面6031可包括触摸检测装置和触摸控制器两个部分。其中,触摸检测装置检测用户的触摸方位,并检测触摸操作带来的信号,将信号传送给触摸控制器;触摸控制器从触摸检测装置上接收触摸信息,并将它转换成触点坐标,再发送给处理器607,触摸控制器还可以接收处理器607发送的指令并加以执行。此外,可以采用电阻式、电容式、红外线以及表面声波等多种类型来实现触敏表面6031。显示器(也称为显示屏)6032可用于显示由用户输入的信息或提供给用户的信息以及手机的各种菜单。可以采用液晶显示器、有机发光二极管等形式来配置显示器6032。触敏表面6031可以覆盖在显示器6032之上,当触敏表面6031检测到在其上或附近的触摸事件后,传送给处理器607以确定触摸事件的类型,随后处理器606可以根据触摸事件的类型在显示器6032上提供相应的视觉输出。虽然在图6中,触敏表面6031与显示屏6032是作为两个独立的部件来实现手机的输入和输出功能,但是在某些实施例中,可以将触敏表面6031与显示屏6032集成而实现手机的输入和输出功能。可以理解的是,603触摸屏是由多层材料堆叠而成,本申请实施例中只展示出了触敏表面(层)和显示屏(层),其他层在本申请实施例中不予记载。另外,在本申请其他一些实施例中,触敏表面6031可以覆盖在显示器6032之上,并且触敏表面6031的尺寸大于显示屏6032的尺寸,使得显示屏6032全部覆盖在触敏表面6031下面,或者,上述触敏表面6031可以以全面板的形式配置在手机的正面,也即用户在手机正面的触摸均能被手机感知,这样就可以实现手机正面的全触控体验。在其他一些实施例中,触敏表面6031以全面板的形式配置在手机的正面,显示器6032也可以以全面板的形式配置在手机的正面,这样在手机的正面就能够实现无边框的结构。
手机还可包括至少一种传感器604,比如光传感器、运动传感器以及其他传感器。具体地,光传感器可包括环境光传感器及接近传感器,其中,环境光传感器可根据环境光线的明暗来调节显示器6032的亮度,接近传感器可在手机移动到耳边时,关闭显示器603和/或背光。作为运动传感器的一种,加速计传感器可检测各个方向上(一般为三轴)加速度的大小,静止时可检测出重力的大小及方向,可用于识别手机姿态的应用(比如横竖屏切换、相关游戏、磁力计姿态校准)、振动识别相关功能(比如计步器、敲击)等;至于手机还可配置的陀螺仪、气压计、湿度计、温度计、红外线传感器等其他传感器,在此不再 赘述。
音频电路606、扬声器6062,传声器6061可提供用户与手机之间的音频接口。音频电路606可将接收到的音频数据转换后的电信号,传输到扬声器6062,由扬声器6062转换为声音信号输出;另一方面,传声器6061将收集的声音信号转换为电信号,由音频电路606接收后转换为音频数据,再将音频数据输出处理器607处理后,经RF电路601以发送给比如另一手机,或者将音频数据输出至存储器602以便进一步处理。
WiFi属于短距离无线传输技术,手机通过WiFi模块605可以帮助用户收发电子邮件、浏览网页和访问流式媒体等,它为用户提供了无线的宽带互联网访问。虽然图6示出了WiFi模块605,但是可以理解的是,其并不属于手机的必须构成,完全可以根据需要在不改变发明的本质的范围内而省略。
处理器607是手机的控制中心,利用各种接口和线路连接整个手机的各个部分,通过运行或执行存储在存储器602内的软件程序和/或模块,以及调用存储在存储器602内的数据,执行手机的各种功能和处理数据,从而对手机进行整体监控。可选的,处理器607可包括一个或多个处理单元;优选的,处理器607可集成应用处理器和调制解调处理器,其中,应用处理器主要处理操作系统、用户界面和应用程序等,调制解调处理器主要处理无线通信。可以理解的是,上述调制解调处理器也可以不集成到处理器607中。
手机还包括给各个部件供电的电源608(比如电池),优选的,电源可以通过电源管理系统与处理器607逻辑相连,从而通过电源管理系统实现管理充电、放电、以及功耗管理等功能。
尽管未示出,手机还可以包括摄像头、蓝牙模块等,在此不再赘述。
本申请实施例可以根据上述方法示例对第一终端设备和第二终端设备进行功能模块的划分,例如,可以对应各个功能划分各个功能模块,也可以将两个或两个以上的功能集成在一个处理模块中。上述集成的模块既可以采用硬件的形式实现,也可以采用软件功能模块的形式实现。需要说明的是,本申请实施例中对模块的划分是示意性的,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式。
比如,以采用集成的方式划分各个功能模块的情况下,图7示出了一种第一终端设备的结构示意图。
在一个具体的实施方式中,收发单元701用于执行上述图1对应的实施例中的步骤104、105,和/或收发单元701还用于执行图1对应的实施例中第一终端设备侧的其他收发步骤。处理单元702用于执行上述图1对应的实施例中的步骤105,和/或处理单元702还用于执行图1对应的实施例中第一终端设备侧的其他处理步骤。
在一个具体的实施方式中,收发单元701用于执行上述图2对应的实施例中的步骤203,204,206,207,和/或收发单元701还用于执行图1对应的实施例中第一终端设备侧的其他收发步骤。播放单元703用于执行上述图2对应的实施例中的步骤207。
以采用集成的方式划分各个功能模块的情况下,图8示出了一种第二终端设备的结构示意图。
在一个具体的实施方式中,收发单元801,用于执行上述图1对应的实施例中的步骤101、 102,上述图1对应的实施例中的步骤103,上述图1对应的实施例中的步骤104,和/或收发单元801还用于执行图1对应的实施例中第二终端设备侧的其他收发步骤。处理单元802用于执行上述图1对应的实施例中的步骤101,和/或处理单元802还用于执行图1对应的实施例中第二终端设备侧的其他收发步骤。
在一个具体的实施方式中,收发单元801,用于执行上述图2对应的实施例中的步骤201、202,上述图2对应的实施例中的步骤203,和/或收发单元801还用于执行图2对应的实施例中第二终端设备侧的其他收发步骤。处理单元802,用于执行上述图2对应的实施例中的步骤201,和/或处理单元802还用于执行图2对应的实施例中第二终端设备侧的其他处理步骤。
由于本申请实施例提供的云服务器、第一终端设备以及第二终端终端设备可用于执行上述定位的方法,因此其所能获得的技术效果可参考上述方法实施例,在此不再赘述。
在上述实施例中,可以全部或部分地通过软件、硬件、固件或者其任意组合来实现。当使用软件实现时,可以全部或部分地以计算机程序产品的形式实现。
所述计算机程序产品包括一个或多个计算机指令。在计算机上加载和执行所述计算机程序指令时,全部或部分地产生按照本申请实施例所述的流程或功能。所述计算机可以是通用计算机、专用计算机、计算机网络、或者其他可编程装置。所述计算机指令可以存储在计算机可读存储介质中,或者从一个计算机可读存储介质向另一计算机可读存储介质传输,例如,所述计算机指令可以从一个网站站点、计算机、服务器或数据中心通过有线(例如同轴电缆、光纤、数字用户线(DSL))或无线(例如红外、无线、微波等)方式向另一个网站站点、计算机、服务器或数据中心进行传输。所述计算机可读存储介质可以是计算机能够存储的任何可用介质或者是包含一个或多个可用介质集成的服务器、数据中心等数据存储设备。所述可用介质可以是磁性介质,(例如,软盘、硬盘、磁带)、光介质(例如,DVD)、或者半导体介质(例如固态硬盘Solid State Disk(SSD))等。
本领域普通技术人员可以理解上述实施例的各种方法中的全部或部分步骤是可以通过程序来指令相关的硬件来完成,该程序可以存储于一计算机可读存储介质中,存储介质可以包括:ROM、RAM、磁盘或光盘等。
以上对本申请实施例所提供的播放方法、播放系统以及存储介质进行了详细介绍,本文中应用了具体个例对本申请的原理及实施方式进行了阐述,以上实施例的说明只是用于帮助理解本申请的方法及其核心思想;同时,对于本领域的一般技术人员,依据本申请的思想,在具体实施方式及应用范围上均会有改变之处,综上所述,本说明书内容不应理解为对本申请的限制。

Claims (17)

  1. 一种播放方法,其特征在于,包括:
    第一终端设备通过广域网接收动作捕捉数据,所述动作捕捉数据由第二终端设备通过云服务器向所述第一终端设备发送;
    所述第一终端设备通过广域网接收语音信息,所述语音信息由第二终端设备通过云服务器向所述第一终端设备发送,所述第一终端设备与所述第二终端设备以及所述云服务器实时共享语音信息;
    所述第一终端设备根据接收到的动作捕捉数据对数字形象进行增强现实AR渲染或者虚拟现实VR渲染得到第一动画;
    所述第一终端设备同步播放所述第一动画和所述语音信息。
  2. 根据权利要求1所述的播放方法,其特征在于,所述语音信息和所述动作捕捉数据包括相同的时间戳;
    所述第一终端设备同步播放所述第一动画和所述语音信息,包括:
    所述第一终端设备根据所述相同的时间戳同步播放所述第一动画和所述语音信息。
  3. 一种播放方法,其特征在于,包括:
    云服务器通过广域网接收第一终端设备发送的第一信息,所述第一信息用于计算数字形象的3D渲染视角参数;
    所述云服务器通过广域网接收第二终端设备发送的动作捕捉数据和语音信息,所述第一终端设备与所述第二终端设备以及所述云服务器实时共享语音信息;
    所述云服务器根据所述动作捕捉数据和所述第一信息对数字形象进行增强现实AR渲染或者虚拟现实VR渲染得到第一动画;
    所述云服务器对所述第一动画和所述语音信息进行编码并进行IP封装得到视频;
    所述云服务器向所述第一终端设备发送所述视频,所述视频用于所述第一终端设备解码后播放。
  4. 一种播放方法,其特征在于,包括:
    第一终端设备向云服务器发送第一信息,所述第一信息用于计算数字形象的3D渲染视角参数,所述3D渲染视角参数用于所述云服务器结合动作捕捉数据对数字形象进行增强现实AR渲染或者虚拟现实VR渲染得到第一动画;
    所述第一终端设备接收所述云服务器发送的视频,所述视频为所述云服务器根据第一动画编码获得。
  5. 一种播放系统,其特征在于,包括:第一终端设备、第二终端设备以及云服务器;
    所述第二终端设备,用于通过广域网向所述云服务器发送动作捕捉数据和语音信息;
    所述云服务器,用于将所述动作捕捉数据和所述语音信息向所述第一终端设备发送;
    所述第一终端设备,用于根据接收到的动作捕捉数据对数字形象进行增强现实AR渲染或者虚拟现实VR渲染得到第一动画;
    所述第一终端设备,还用于同步播放所述第一动画和所述语音信息。
  6. 一种播放系统,其特征在于,包括:第一终端设备、第二终端设备以及云服务器;
    所述第一终端设备为权利要求4所描述的第一终端设备,所述第二终端设备为权利要求3所描述的第二终端设备,所述云服务器为权利要求3所描述的云服务器。
  7. 一种终端设备,其特征在于,包括:
    收发单元,用于通过广域网接收动作捕捉数据,所述动作捕捉数据由第二终端设备通过云服务器向所述第一终端设备发送;
    所述收发单元,还用于通过广域网接收语音信息,所述语音信息由第二终端设备通过云服务器向所述第一终端设备发送,所述第一终端设备与所述第二终端设备以及所述云服务器实时共享语音信息;
    处理单元,用于根据所述收发单元接收到的动作捕捉数据对数字形象进行增强现实AR渲染或者虚拟现实VR渲染得到第一动画;
    播放单元,用于同步播放所述第一动画和所述语音信息。
  8. 一种云服务器,其特征在于,包括:
    收发单元,用于通过广域网接收第一终端设备发送的第一信息,所述第一信息用于计算数字形象的3D渲染视角参数;
    所述收发单元,还用于通过广域网接收第二终端设备发送的动作捕捉数据和语音信息,所述第一终端设备与所述第二终端设备以及所述云服务器实时共享语音信息;
    处理单元,用于根据所述收发单元接收到的所述动作捕捉数据和所述第一信息对数字形象进行增强现实AR渲染或者虚拟现实VR渲染得到第一动画;
    所述处理单元,还用于所述对所述第一动画和所述语音信息进行编码并进行IP封装得到视频;
    所述收发单元,还用于向所述第一终端设备发送所述视频,所述视频用于所述第一终端设备解码后播放。
  9. 一种终端设备,其特征在于,包括:
    收发单元,用于向云服务器发送第一信息,所述第一信息用于计算数字形象的3D渲染视角参数,所述3D渲染视角参数用于所述云服务器结合动作捕捉数据对数字形象进行增强现实AR渲染或者虚拟现实VR渲染得到第一动画;
    所述收发单元,还用于接收所述云服务器发送的视频,所述视频为所述云服务器根据第一动画编码获得。
  10. 一种计算机可读存储介质,其特征在于,当指令在计算机装置上运行时,使得所述计算机装置执行如权利要求1或2所述的方法。
  11. 一种计算机可读存储介质,其特征在于,当指令在计算机装置上运行时,使得所述计算机装置执行如权利要求3所述的方法。
  12. 一种计算机可读存储介质,其特征在于,当指令在计算机装置上运行时,使得所述计算机装置执行如权利要求4所述的方法。
  13. 一种计算机程序产品,当其在计算机上运行时,使得计算机可以执行如权利要求1或2所述的方法。
  14. 一种计算机程序产品,当其在计算机上运行时,使得计算机可以执行如权利要求 3所述的方法。
  15. 一种计算机程序产品,当其在计算机上运行时,使得计算机可以执行如权利要求4所述的方法。
  16. 一种终端设备,其特征在于,所述设备包括存储器、处理器,所述存储器中存储代码和数据,所述存储器与所述处理器耦合,所述处理器运行所述存储器中的代码使得所述设备执行如权利要求1或2所述的方法,或者执行如权利要求4所述的方法。
  17. 一种云服务器,其特征在于,所述设备包括存储器、处理器,所述存储器中存储代码和数据,所述存储器与所述处理器耦合,所述处理器运行所述存储器中的代码使得所述设备执行如权利要求3所述的方法。
PCT/CN2020/122672 2019-10-22 2020-10-22 一种播放方法以及播放系统 WO2021078182A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201911007650.8 2019-10-22
CN201911007650.8A CN112702611A (zh) 2019-10-22 2019-10-22 一种播放方法以及播放系统

Publications (1)

Publication Number Publication Date
WO2021078182A1 true WO2021078182A1 (zh) 2021-04-29

Family

ID=75504713

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/122672 WO2021078182A1 (zh) 2019-10-22 2020-10-22 一种播放方法以及播放系统

Country Status (2)

Country Link
CN (1) CN112702611A (zh)
WO (1) WO2021078182A1 (zh)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113289332B (zh) * 2021-06-17 2023-08-01 广州虎牙科技有限公司 游戏交互方法、装置、电子设备及计算机可读存储介质

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105915849A (zh) * 2016-05-09 2016-08-31 惠州Tcl移动通信有限公司 一种虚拟现实体育赛事播放方法及系统
CN106790621A (zh) * 2016-12-30 2017-05-31 广州幻境科技有限公司 一种动作捕捉数据的云分享控制方法、及其系统
CN107103801A (zh) * 2017-04-26 2017-08-29 北京大生在线科技有限公司 远程三维场景互动教学系统及控制方法
CN107690799A (zh) * 2017-08-25 2018-02-13 深圳市瑞立视多媒体科技有限公司 一种数据同步的方法、装置及服务器
CN107979763A (zh) * 2016-10-21 2018-05-01 阿里巴巴集团控股有限公司 一种虚拟现实设备生成视频、播放方法、装置及系统
US9984499B1 (en) * 2015-11-30 2018-05-29 Snap Inc. Image and point cloud based tracking and in augmented reality systems

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102340690A (zh) * 2011-09-13 2012-02-01 苏州美娱网络科技有限公司 一种互动电视节目系统及实现方法
CN104796393A (zh) * 2014-05-30 2015-07-22 厦门极致互动网络技术有限公司 一种基于服务器实时渲染的网络游戏系统及方法
CN106385681A (zh) * 2016-09-21 2017-02-08 平越 虚拟现实娱乐系统及其方法
CN109126122A (zh) * 2017-06-16 2019-01-04 上海拆名晃信息科技有限公司 一种用于虚拟现实的云游戏系统实现方法
CN109215132A (zh) * 2017-06-30 2019-01-15 华为技术有限公司 一种增强现实业务的实现方法及设备
CN107613310B (zh) * 2017-09-08 2020-08-04 广州华多网络科技有限公司 一种直播方法、装置及电子设备
CN108597028A (zh) * 2018-04-11 2018-09-28 北京邮电大学 一种连续加载Web AR对象的方法、装置及设备
CN109195020B (zh) * 2018-10-11 2021-07-02 三星电子(中国)研发中心 一种ar增强的游戏直播方法和系统
CN109729368A (zh) * 2018-11-16 2019-05-07 深圳市赛亿科技开发有限公司 Vr直播视频处理的方法和系统

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9984499B1 (en) * 2015-11-30 2018-05-29 Snap Inc. Image and point cloud based tracking and in augmented reality systems
CN105915849A (zh) * 2016-05-09 2016-08-31 惠州Tcl移动通信有限公司 一种虚拟现实体育赛事播放方法及系统
CN107979763A (zh) * 2016-10-21 2018-05-01 阿里巴巴集团控股有限公司 一种虚拟现实设备生成视频、播放方法、装置及系统
CN106790621A (zh) * 2016-12-30 2017-05-31 广州幻境科技有限公司 一种动作捕捉数据的云分享控制方法、及其系统
CN107103801A (zh) * 2017-04-26 2017-08-29 北京大生在线科技有限公司 远程三维场景互动教学系统及控制方法
CN107690799A (zh) * 2017-08-25 2018-02-13 深圳市瑞立视多媒体科技有限公司 一种数据同步的方法、装置及服务器

Also Published As

Publication number Publication date
CN112702611A (zh) 2021-04-23

Similar Documents

Publication Publication Date Title
US11120616B2 (en) Method for implementing virtual scene conversion and related apparatus
WO2020114271A1 (zh) 图像渲染方法、设备及存储介质
WO2020186988A1 (zh) 资讯的展示方法、装置、终端及存储介质
WO2018113639A1 (zh) 用户终端之间的互动方法、终端、服务器、系统及存储介质
US9632683B2 (en) Methods, apparatuses and computer program products for manipulating characteristics of audio objects by using directional gestures
WO2020098462A1 (zh) Ar虚拟人物绘制方法、装置、移动终端及存储介质
WO2016173427A1 (zh) 一种残影效果的实现方法,装置以及计算机可读介质
CN111263181A (zh) 直播互动方法、装置、电子设备、服务器及存储介质
WO2021077923A1 (zh) 控制显示设备的方法及便携设备
WO2013145673A1 (en) Information processing apparatus, information processing method, and program
CN108513671B (zh) 一种2d应用在vr设备中的显示方法及终端
CN111010523B (zh) 一种视频录制方法及电子设备
WO2018216355A1 (ja) 情報処理装置、情報処理方法、及びプログラム
CN113485626A (zh) 一种智能显示设备、移动终端和显示控制方法
WO2021078182A1 (zh) 一种播放方法以及播放系统
CN109117037B (zh) 一种图像处理的方法及终端设备
CN110536236A (zh) 一种通信方法、终端设备和网络设备
WO2021104162A1 (zh) 显示方法及电子设备
KR102261739B1 (ko) 증강 현실 미디어 콘텐츠의 적응적 스트리밍 시스템 및 적응적 스트리밍 방법
CN113014960B (zh) 一种在线制作视频的方法、装置及存储介质
CN111093033B (zh) 一种信息处理方法及设备
WO2019071562A1 (zh) 一种数据处理方法及终端
US11212442B2 (en) Creation and display of a 360 degree looping video file
CN110471895A (zh) 分享方法及终端设备
CN113485596B (zh) 虚拟模型的处理方法、装置、电子设备及存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20880145

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20880145

Country of ref document: EP

Kind code of ref document: A1