CN112702611A - Playing method and playing system - Google Patents

Playing method and playing system Download PDF

Info

Publication number
CN112702611A
CN112702611A CN201911007650.8A CN201911007650A CN112702611A CN 112702611 A CN112702611 A CN 112702611A CN 201911007650 A CN201911007650 A CN 201911007650A CN 112702611 A CN112702611 A CN 112702611A
Authority
CN
China
Prior art keywords
terminal device
cloud server
rendering
motion capture
voice information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911007650.8A
Other languages
Chinese (zh)
Inventor
王斌
赵其勇
史浩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Shanghai Huawei Technologies Co Ltd
Original Assignee
Shanghai Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Huawei Technologies Co Ltd filed Critical Shanghai Huawei Technologies Co Ltd
Priority to CN201911007650.8A priority Critical patent/CN112702611A/en
Priority to PCT/CN2020/122672 priority patent/WO2021078182A1/en
Publication of CN112702611A publication Critical patent/CN112702611A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/4302Content synchronisation processes, e.g. decoder synchronisation
    • H04N21/4307Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/65Transmission of management data between client and server
    • H04N21/658Transmission by the client directed to the server
    • H04N21/6587Control parameters, e.g. trick play commands, viewpoint selection
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8146Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics
    • H04N21/8153Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics comprising still images, e.g. texture, background image
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8547Content authoring involving timestamps for synchronizing content

Abstract

The application discloses a playing method, which comprises the following steps: the first terminal device receives motion capture data through the wide area network, and the motion capture data is sent to the first terminal device by the second terminal device through the cloud server. The first terminal device receives voice information through a wide area network, the voice information is sent to the first terminal device through the cloud server by the second terminal device, and the first terminal device, the second terminal device and the cloud server share the voice information in real time. The first terminal device conducts Augmented Reality (AR) rendering or Virtual Reality (VR) rendering on the digital image according to the received motion capture data to obtain a first animation. The first terminal equipment synchronously plays the first animation and the voice information. Through the scheme provided by the application, the second terminal equipment and the first terminal equipment do not need to be deployed in the same place, digital images can be watched by the first terminal equipment in a holographic mode, the watching visual angle of a user is adjusted by the user, and the watching experience of the user is improved.

Description

Playing method and playing system
Technical Field
The present application relates to the field of communications technologies, and in particular, to a playing method.
Background
Virtual idols are media stars generated with computer technology such as "the beginning of the day", "luotaiyi", "seiko", etc. In 2018, the virtual idol with live broadcast as the expression form is derived by the red trip of the virtual idol. Live virtual idols refer to the presentation of real character motion to a virtual character by means of motion capture devices and sensors placed on the head and limbs of the real character. By means of a real-time motion capture mechanism, the virtual idol can also interact with fans in the real world in a body or language mode.
Because the real-time performance requirement is higher in a live scene, the current motion capture system and the rendering host are deployed in the same place and are connected through a cable or are in the same local area network, so that the portability of the deployment of the motion capture system is limited to a certain extent, and the deployment cost of the motion capture system is increased. In addition, the current virtual anchor can only be watched in a live video mode, which means that all viewers watch the virtual anchor at the same viewing angle controlled by the director, cannot watch the virtual anchor at the viewing angle required by the viewers, and are lack of presence and poor in experience of the viewers.
Disclosure of Invention
The embodiment of the application provides a playing method, so that the motion capture system and the rendering host do not need to be deployed in the same place, a user can watch digital images in a holographic mode, the watching visual angle is automatically adjusted, and the watching experience of the user is improved.
In order to achieve the above purpose, the embodiments of the present application provide the following technical solutions:
a first aspect of the present application provides a playing method, which may include: the first terminal device receives motion capture data through the wide area network, and the motion capture data is sent to the first terminal device by the second terminal device through the cloud server. The first terminal device receives voice information through a wide area network, the voice information is sent to the first terminal device through the cloud server by the second terminal device, and the first terminal device, the second terminal device and the cloud server share the voice information in real time. The first terminal device conducts Augmented Reality (AR) rendering or Virtual Reality (VR) rendering on the digital image according to the received motion capture data to obtain a first animation. The first terminal equipment synchronously plays the first animation and the voice information. Known by the first aspect, second terminal equipment acquires the action capture data, transmit the action capture data through the wide area network, first terminal equipment is according to the action capture data and is makeed up digital image, second terminal equipment and first terminal equipment need not arrange in same place, in addition, compare and carry out VR to video content in prior art and play up or AR and play up, this scheme carries out VR to the virtual image of action capture data real-time drive and plays up or AR and play up, realize that first terminal equipment watches digital image with holographic mode, the user watches the visual angle and adjusts by oneself, promote user's the experience of watching.
Optionally, in combination with the first aspect described above, in a first possible implementation, the voice information and the motion capture data may comprise the same timestamp. The first terminal device synchronously plays the first animation and the voice information, and may include: and the first terminal equipment synchronously plays the first animation and the voice information according to the same timestamp.
A second aspect of the present application provides a playing method, which may include: the cloud server receives first information sent by the first terminal device through the wide area network, and the first information is used for calculating the 3D rendering visual angle parameters of the digital image. The cloud server receives the motion capture data and the voice information sent by the second terminal device through the wide area network, and the first terminal device, the second terminal device and the cloud server share the voice information in real time. And the cloud server performs augmented reality AR rendering or virtual reality VR rendering on the digital image according to the motion capture data and the first information to obtain a first animation. And the cloud server encodes the first animation and the voice information and performs IP packaging to obtain a video. The cloud server sends a video to the first terminal device, and the video is used for playing after the first terminal device decodes the video. According to the second aspect, the cloud server renders the digital image to obtain an IP video stream, and sends the encoded IP video stream to the first terminal device, and after the cloud server obtains the control information of the first terminal device, the cloud server may render the digital image and the scene again to obtain an updated IP video stream according to the control information, and send the updated IP video stream to the first terminal device. The scheme provided by the second aspect solves the problem that the rendering requirement of the first terminal equipment cannot be met due to the fact that the model precision of the digital image is too high.
A third aspect of the present application provides a playing method, which may include: the method comprises the steps that first information is sent to a cloud server by a first terminal device, the first information is used for calculating a 3D rendering visual angle parameter of a digital image, and the 3D rendering visual angle parameter is used for carrying out augmented reality AR rendering or virtual reality VR rendering on the digital image by the cloud server in combination with motion capture data to obtain a first animation. The first terminal device receives a video sent by the cloud server, and the video is obtained by the cloud server according to the first animation code.
A fourth aspect of the present application provides a playback system, which may include: the terminal comprises a first terminal device, a second terminal device and a cloud server. And the second terminal equipment is used for sending the motion capture data and the voice information to the cloud server through the wide area network. And the cloud server is used for sending the motion capture data and the voice information to the first terminal equipment. And the first terminal equipment is used for performing augmented reality AR rendering or virtual reality VR rendering on the digital image according to the received motion capture data to obtain a first animation. The first terminal equipment is also used for synchronously playing the first animation and the voice information.
A fifth aspect of the present application provides a playback system, which may include: the terminal comprises a first terminal device, a second terminal device and a cloud server. The first terminal device is the first terminal device described in the third aspect, the second terminal device is the second terminal device described in the second aspect, and the cloud server is the cloud server described in the second aspect.
A sixth aspect of the present application provides a terminal device, which may include: and the transceiving unit is used for receiving the motion capture data through the wide area network, and the motion capture data is sent to the first terminal equipment by the second terminal equipment through the cloud server. The receiving and sending unit is further used for receiving voice information through the wide area network, the voice information is sent to the first terminal device through the cloud server by the second terminal device, and the first terminal device, the second terminal device and the cloud server share the voice information in real time. And the processing unit is used for performing augmented reality AR rendering or virtual reality VR rendering on the digital image according to the motion capture data received by the receiving and sending unit to obtain a first animation. And the playing unit is used for synchronously playing the first animation and the voice information.
A seventh aspect of the present application provides a cloud server, which may include: and the transceiving unit is used for receiving first information sent by the first terminal equipment through a wide area network, and the first information is used for calculating the 3D rendering visual angle parameter of the digital image. The receiving and sending unit is further used for receiving the motion capture data and the voice information sent by the second terminal device through the wide area network, and the first terminal device, the second terminal device and the cloud server share the voice information in real time. And the processing unit is used for performing augmented reality AR rendering or virtual reality VR rendering on the digital image according to the motion capture data and the first information received by the transceiving unit to obtain a first animation. And the processing unit is also used for coding the first animation and the voice information and carrying out IP (Internet protocol) encapsulation to obtain a video. And the transceiving unit is also used for sending a video to the first terminal equipment, and the video is played after being decoded by the first terminal equipment.
An eighth aspect of the present application provides a terminal device, which may include: the receiving and sending unit is used for sending first information to the cloud server, the first information is used for calculating a 3D rendering visual angle parameter of the digital image, and the 3D rendering visual angle parameter is used for carrying out augmented reality AR rendering or virtual reality VR rendering on the digital image by the cloud server in combination with the motion capture data to obtain a first animation. And the receiving and sending unit is further used for receiving the video sent by the cloud server, and the video is obtained by the cloud server according to the first animation code.
A ninth aspect of the present application provides a computer-readable storage medium, which when executed on a computer device, causes the computer device to execute the playback method of the first aspect.
A tenth aspect of the present application provides a computer-readable storage medium, which when executed on a computer device, causes the computer device to perform the playback method of the second aspect.
An eleventh aspect of the present application provides a computer-readable storage medium, which when executed on a computer apparatus, causes the computer apparatus to perform the playback method of the third aspect.
A twelfth aspect of the present application provides a computer program product, which when run on a computer, enables the computer to perform the playback method as in the first aspect.
A thirteenth aspect of the present application provides a computer program product, which when run on a computer, enables the computer to perform the playback method as in the second aspect.
A fourteenth aspect of the present application provides a computer program product, which when run on a computer, enables the computer to perform the playback method as in the third aspect.
A fifteenth aspect of the present application provides a terminal device, including a memory, and a processor, where the memory stores codes and data, and the memory is coupled to the processor, and the processor executes the codes in the memory to make the device execute a playing method according to the first aspect or the first possible implementation manner of the first aspect, or execute a playing method according to the third aspect.
A sixteenth aspect of the present application provides a cloud server, comprising a memory, a processor, the memory storing code and data therein, the memory being coupled to the processor, the processor executing the code in the memory to make the device execute the playing method according to the second aspect.
Through the scheme provided by the application, the second terminal equipment and the first terminal equipment do not need to be deployed in the same place, digital images can be watched by the first terminal equipment in a holographic mode, the watching visual angle of a user is adjusted by the user, and the watching experience of the user is improved.
Drawings
Fig. 1 is a schematic structural diagram of a playing system according to an embodiment of the present application;
fig. 2 is a schematic structural diagram of another playback system according to an embodiment of the present application;
fig. 3 is a schematic view of an application scenario provided in an embodiment of the present application;
fig. 4 is a schematic diagram of a hardware structure of a communication device according to an embodiment of the present application;
fig. 5 is a schematic structural diagram of a cloud server according to an embodiment of the present application;
fig. 6 is a block diagram of a partial structure of a mobile phone related to a first terminal device and a second terminal device provided in the embodiment of the present application;
fig. 7 is a schematic structural diagram of a first terminal device according to an embodiment of the present application;
fig. 8 is a schematic structural diagram of a second terminal device according to an embodiment of the present application.
Detailed Description
Embodiments of the present application will now be described with reference to the accompanying drawings, and it is to be understood that the described embodiments are merely illustrative of some, but not all, embodiments of the present application. As can be known to those skilled in the art, with the development of technology and the emergence of new scenarios, the technical solution provided in the embodiments of the present application is also applicable to similar technical problems.
The embodiment of the application provides a playing method, a playing system and a storage medium. The first terminal device receives voice information through a wide area network, the voice information is sent to the first terminal device through the cloud server by the second terminal device, and the first terminal device, the second terminal device and the cloud server share the voice information in real time. The first terminal device conducts Augmented Reality (AR) rendering or Virtual Reality (VR) rendering on the digital image according to the received motion capture data to obtain a first animation. The first terminal equipment synchronously plays the first animation and the voice information. Through the scheme provided by the application, the second terminal equipment and the first terminal equipment do not need to be deployed in the same place, digital images can be watched by the first terminal equipment in a holographic mode, the watching visual angle of a user is adjusted by the user, and the watching experience of the user is improved. The following are detailed below.
The terms "first," "second," and the like in the description and in the claims of the present application and in the above-described drawings are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It will be appreciated that the data so used may be interchanged under appropriate circumstances such that the embodiments described herein may be practiced otherwise than as specifically illustrated or described herein. Moreover, the terms "comprises," "comprising," and any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or modules is not necessarily limited to those steps or modules explicitly listed, but may include other steps or modules not expressly listed or inherent to such process, method, article, or apparatus. The naming or numbering of the steps appearing in the present application does not mean that the steps in the method flow have to be executed in the chronological/logical order indicated by the naming or numbering, and the named or numbered process steps may be executed in a modified order depending on the technical purpose to be achieved, as long as the same or similar technical effects are achieved. The division of the modules presented in this application is a logical division, and in practical applications, there may be another division, for example, multiple modules may be combined or integrated into another system, or some features may be omitted, or not executed, and in addition, the shown or discussed coupling or direct coupling or communication connection between each other may be through some ports, and the indirect coupling or communication connection between the modules may be in an electrical or other similar form, which is not limited in this application. The modules or sub-modules described as separate components may or may not be physically separated, may or may not be physical modules, or may be distributed in a plurality of circuit modules, and some or all of the modules may be selected according to actual needs to achieve the purpose of the present disclosure.
It should be noted that, in the embodiments of the present application, the terms "network" and "system" are often used interchangeably, but those skilled in the art can understand the meaning thereof. Information (information), signal (signal), message (message) may sometimes be mixed, it being noted that the intended meaning is consistent when no distinction is emphasized.
It should be noted that, in the embodiments of the present application, "reporting" and "feedback" are often used interchangeably, but those skilled in the art can understand the meaning of the "reporting" and the "feedback" as well as the "response". Therefore, in the embodiments of the present application, the intended meanings thereof are consistent when the differences are not emphasized.
It should be noted that, in the embodiments of the present application, "data" and "information" are often used interchangeably, but those skilled in the art can understand the meaning thereof. Therefore, in the embodiments of the present application, the intended meanings thereof are consistent when the differences are not emphasized.
Virtual idols are media stars generated with computer technology such as "the beginning of the day", "luotaiyi", "seiko", etc. In 2018, the virtual idol with live broadcast as the expression form is derived by the red trip of the virtual idol. Live virtual idols refer to the presentation of real character motion to a virtual character by means of motion capture devices and sensors placed on the head and limbs of the real character. By means of a real-time motion capture mechanism, the virtual idol can also interact with fans in the real world in a body or language mode.
Because the real-time performance requirement is higher in a live scene, the current motion capture system and the rendering host are deployed in the same place and are connected through a cable or are in the same local area network, so that the portability of the deployment of the motion capture system is limited to a certain extent, and the deployment cost of the motion capture system is increased. In addition, the current virtual anchor can only be watched in a live video mode, which means that all viewers watch the virtual anchor at the same viewing angle controlled by the director, cannot watch the virtual anchor at the viewing angle required by the viewers, and are lack of presence and poor in experience of the viewers. In order to solve the above problem, the present application provides a playback system, which will be described in detail below.
In the scheme provided by the application, by introducing the cloud server, the motion capture system and the rendering host do not need to be deployed in the same place, wherein in one mode, the rendering process is performed on the terminal setting of the user, in the other mode, the rendering process is performed on the cloud server, and the two modes are respectively explained below.
Fig. 1 is a schematic structural diagram of a playing system according to an embodiment of the present application.
As shown in fig. 1, a playing system provided in an embodiment of the present application may include:
101. the second terminal device obtains motion capture data.
Motion capture devices are placed on the human body to capture the motion of the human body. For example, the motion capture device can be respectively arranged on any parts of the sacrum, the left thigh, the right thigh, the left calf, the right calf, the left foot, the right foot, the left shoulder, the right shoulder, the vest, the left upper arm, the right upper arm, the left forearm, the right forearm, the left palm, the right palm and the back head scoop of the human body according to requirements. Of course, the above-mentioned human body parts are only for illustration and do not represent a limitation on the position of the motion capture device, and in practical application scenarios, the motion capture device may be disposed at other positions of the human body as required, for example, the motion capture device may also be disposed on the face of a human body to obtain the expression and mouth shape of the human body. The motion capture equipment collects human body motions and then sends motion signals to the second terminal equipment, and the second terminal equipment processes the motion signals to obtain motion capture data. It should be noted that, in the prior art, the embodiments of the present application may be applied to any technical solutions on how to acquire motion capture data.
102. And the cloud server receives the motion capture data sent by the second terminal equipment through the wide area network.
And the cloud server receives the motion capture data sent by the second terminal equipment through the high-speed internet. For example, the cloud server receives the motion capture data sent by the second terminal device through a fifth generation (5G) mobile communication system. Compared with the prior art, in order to reduce time delay between the motion capture process and the rendering process and realize smoothness of digital image live broadcast, motion capture data are transmitted in a local area network in the prior art, so that the current motion capture system and the rendering host are deployed in the same place, and the motion capture data can be transmitted in a wide area network by introducing a 5G communication system in the scheme. In addition, with the development of technology, a sixth generation (6th generation, 6G) mobile communication system, a seventh generation (7th generation, 7G) mobile communication system, and the like may be developed in the future, and the embodiments of the present application may be applied.
103. The cloud server sends the motion capture data to the first terminal device through the wide area network.
In this embodiment, the first terminal device may be a terminal device registered in a cloud server. For example, the first terminal device may send a registration request to the cloud server, where the registration request carries identification information of the first terminal device, and for example, when the cloud server receives registration requests sent by N first terminal devices, where N is a positive integer, the cloud server may send motion capture data to the N first terminal devices after receiving the motion capture data sent by the second terminal device. Or, the first terminal device may also be a terminal device installed with an APP corresponding to the cloud server, for example, the APP corresponding to the cloud server is downloaded and installed in advance by the first terminal device, and then the user may interact with the cloud server by controlling the APP corresponding to the second terminal.
104. The second terminal device, the cloud server and the first terminal device share the voice information in real time.
The second terminal device, the cloud server and the first terminal device can perform voice real-time communication, and regarding the voice real-time communication among the plurality of terminal devices, mature technologies such as a voice communication function of WeChat exist in the prior art, and technologies related to the voice real-time communication among the devices in the prior art can be adopted in the application.
105. The first terminal device performs Augmented Reality (AR) rendering or Virtual Reality (VR) rendering of the digital image according to the received motion capture data.
In this application, the digital representation may be a digital model. The digital representation may be a 3D digital model of the virtual idol mentioned above, or a digital model of another cartoon character, or a digital model of another representation that requires human interaction. If the digital image is pre-stored on the first terminal device, the locally stored digital image can be directly used, and if the digital image is not pre-stored on the first terminal device, the first terminal device can acquire the digital image from the cloud server.
VR technology refers to a technology for providing an immersive sensation in an interactive three-dimensional environment generated on a computer by comprehensively using a computer graphics device and various interface devices such as reality and control. The AR means that virtual information is applied to the real world by a computer technology, and a real environment and a virtual object coexist while being superimposed on the same picture or space in real time.
In a particular embodiment, the first terminal device VR renders the digital image from the received motion capture data. Specifically, the VR engine calculates the 3D rendering view angle parameter according to the action information of the user, for example, if the user adjusts and controls the view angle through the handle, the VR engine acquires control information of the handle by the user, and calculates the 3D rendering view angle parameter according to the acquired control information of the handle by the user. And the 3D rendering engine renders a digital model corresponding to the digital image, namely the rendered digital image, according to the 3D rendering visual angle parameters acquired by the VR engine and the acquired motion capture data. For example, a unity 3D model of a digital image can be rendered by using unity 3D, where unity 3D is a development software, and is one of popular 3D game development engines today, and of course, unity 3D is only for illustration, any engine that can implement 3D rendering may be used in the embodiments of the present application. And the VR engine carries out VR rendering on the 3D model obtained after the 3D rendering engine renders.
In a particular embodiment, the first terminal device AR renders the digital image from the received motion capture data. Specifically, the AR engine calculates the 3D rendering view angle parameter, for example, the AR engine acquires space coordinate information and gesture control of a user, and calculates the 3D rendering view angle parameter according to the acquired space coordinate information and the gesture control of the user. And the 3D rendering engine renders a digital model corresponding to the digital image, namely the rendered digital image, according to the 3D rendering visual angle parameters acquired by the AR engine and the acquired motion capture data.
In a specific implementation manner, a scene where the digital image is located may also be set, where the scene corresponds to a digital model of a 3D scene, and the digital model of the 3D scene is rendered while the digital image is rendered. The digital model of the scene may be locally stored by the first terminal device, or may be obtained by the first terminal device from a cloud server. It should be noted that, in the live broadcast process, the first terminal device may further receive a scene change instruction sent by the second terminal device through the cloud server, and after receiving the scene change instruction, the first terminal device changes to the specified scene, and correspondingly renders the 3D digital model of the changed scene.
In addition, in order to ensure the synchronization of the motion and the voice of the digital image, in a specific embodiment, the method further comprises 106, the first terminal device synchronously rendering the motion capture data and the voice information.
The motion capture data and the voice information received by the first terminal device respectively comprise corresponding same timestamps, the first terminal device can determine the voice information matched with the motion capture data according to the corresponding timestamps or determine the motion capture data matched with the voice information according to the corresponding timestamps, synchronous rendering of the motion capture data and the voice information is realized, synchronization of digital image motion and voice is ensured, namely, the first terminal device synchronously plays a first animation obtained after VR rendering or AR rendering and the voice information.
It can be known from the embodiment that fig. 1 corresponds, the second terminal equipment acquires the motion capture data, the motion capture data is transmitted through the wide area network, the first terminal equipment renders the digital image according to the motion capture data, the second terminal equipment and the first terminal equipment do not need to be deployed in the same place, in addition, compare and carry out VR rendering or AR rendering to the video content in prior art, this scheme carries out VR rendering or AR rendering to the virtual image that the motion capture data real-time drive, realize that the digital image is watched with holographic mode to the first terminal equipment, the user watches the visual angle and adjusts by oneself, promote user's viewing experience.
Fig. 2 is a schematic structural diagram of another playing system according to an embodiment of the present application.
As shown in fig. 2, a playing system provided in an embodiment of the present application may include:
201. the second terminal device obtains motion capture data.
202. And the cloud server receives the motion capture data sent by the second terminal equipment through the wide area network.
Step 201 and step 202 can be understood by referring to steps 101 and 102 in the embodiment corresponding to fig. 1, and are not repeated herein.
203. The second terminal device, the cloud server and the first terminal device share the voice information in real time.
The first terminal device may be a terminal device registered in the cloud server. The second terminal device, the cloud server and the first terminal device can perform voice real-time communication, and regarding the voice real-time communication among the plurality of terminal devices, mature technologies such as a voice communication function of WeChat exist in the prior art, and technologies related to the voice real-time communication among the devices in the prior art can be adopted in the application.
204. The cloud server receives first information sent by the first terminal device.
The first information is control information of a user. For example, the adjustment information of the viewing angle of the user, the zoom information of the picture, and the like are acquired, for example, in an AR scene, gesture information or touch screen control information of the user may be sent to the cloud server, and in a VR scene, action information of the user on the VR helmet and control information of the user on the VR handle may be sent to the cloud server.
205. And the cloud server performs AR rendering or VR rendering on the digital image according to the received motion capture data and the first information.
In this application, the digital representation may be a digital model. The digital representation may be a 3D digital model of the virtual idol mentioned above, or a digital model of another cartoon character, or a digital model of another representation that requires human interaction. If the digital image is pre-stored on the first terminal device, the locally stored digital image can be directly used, and if the digital image is not pre-stored on the first terminal device, the first terminal device can acquire the digital image from the cloud server.
In a particular embodiment, the cloud server VR renders the digital image from the received motion capture data. Specifically, the VR engine calculates a 3D rendering visual angle parameter according to the first information, the 3D rendering engine renders a digital model corresponding to the digital image according to the 3D rendering visual angle parameter acquired by the VR engine and the acquired motion capture data, namely renders the digital image. And the VR engine carries out VR rendering on the 3D model obtained after the 3D rendering engine renders.
In a particular embodiment, the cloud server performs AR rendering of the digital image from the received motion capture data. Specifically, the AR engine calculates the 3D rendering view angle parameter, for example, the AR engine calculates the 3D rendering view angle parameter according to the first information. And the 3D rendering engine renders a digital model corresponding to the digital image, namely the rendered digital image, according to the 3D rendering visual angle parameters acquired by the AR engine and the acquired motion capture data.
In a specific implementation manner, a scene where the digital image is located may also be set, where the scene corresponds to a digital model of a 3D scene, and the digital model of the 3D scene is rendered while the digital image is rendered.
Furthermore, to ensure synchronization of motion and voice of the digital character, in one particular embodiment, the cloud server renders the motion capture data and the voice information synchronously.
The motion capture data and the voice information received by the cloud server respectively comprise corresponding same timestamps, the cloud server can determine the voice information matched with the motion capture data according to the corresponding timestamps or determine the motion capture data matched with the voice information according to the corresponding timestamps, synchronous rendering of the motion capture data and the voice information is achieved, and synchronization of digital image motion and voice is guaranteed.
206. And the cloud server sends the coded IP video stream to the first terminal equipment.
The cloud server carries out VR rendering or AR rendering on the digital image and voice information driven by the motion capture data in real time, audio and video coding on the data obtained after the scene is subjected to the VR rendering or the AR rendering, IP encapsulation is carried out on the coded data, and the encapsulated IP video stream is sent to the first terminal equipment.
207. And the first terminal equipment receives the IP video stream sent by the cloud server, decodes the IP video stream and displays the IP video stream.
It should be noted that, steps 204 to 206 may be repeatedly performed, for example, the first terminal device sends the first information to the cloud server at preset intervals, the cloud server performs AR rendering or VR rendering on the digital video according to the received first information and the motion capture data at preset intervals, and sends the updated IP video stream to the first terminal device.
As can be seen from the embodiment corresponding to fig. 2, the cloud server renders the digital image to obtain an IP video stream, and sends the encoded IP video stream to the first terminal device, and after the cloud server obtains the control information of the first terminal device, the cloud server may render the digital image and the scene again to obtain an updated IP video stream according to the control information, and send the updated IP video stream to the first terminal device. The scheme corresponding to fig. 2 solves the problem that the rendering requirement of the first terminal device cannot be met due to the fact that the model precision of the digital image is too high.
In the embodiments corresponding to fig. 1 and fig. 2, the second terminal device needs to acquire the motion capture data and the voice information, and in some specific application scenarios, the first terminal device may be regarded as a whole, or may be regarded as being composed of several different devices, for example, the second terminal device includes a voice acquisition device and a motion capture data acquisition device. In addition, the cloud server receives the motion capture data and the voice information sent by the second terminal device, and in some specific application scenarios, the cloud server may be regarded as a whole, or may be regarded as composed of several different devices, for example, the cloud server may include a voice interaction server and a system data server, which are described below separately.
As shown in fig. 3, a schematic view of an application scenario provided in the embodiment of the present application is shown.
Taking a mobile phone as an example for explanation, a user can control an APP on the mobile phone to interact with a cloud server. As shown in fig. 3 a, the user may click into the APP to select the desired live or video to watch. For example, after the user clicks the APP, the user enters a picture shown in b in fig. 3, and the user may select a category that the user wants to watch, for example, the user may select to watch live broadcast or to select video or playback, and assuming that the user selects to watch live broadcast, in a live broadcast interface, the user may further select a main broadcast that the user wants to watch, where a virtual main broadcast may be included, and b in fig. 3 schematically shows a first live broadcast room, a second live broadcast room, a third live broadcast room, and a fourth live broadcast room. Assuming that a user selects to enter a first live broadcast room, the anchor of the first live broadcast room is a first digital image, the mobile phone receives motion capture data, a digital model of the first digital image, voice information and scene information from a cloud, and according to the information, the mobile phone combines with the operation of the user, such as the operation of the user on a handle and a helmet in a VR scene, and performs VR rendering or AR rendering on the virtual image and the scene model driven by the motion capture data in real time, so that the user can view live broadcast in an immersive mode. It should be noted that, if the mobile phone locally stores the data related to the digital model of the first digital image and the digital model corresponding to the scene, the data may not be acquired from the cloud, which has already been described in the embodiment corresponding to fig. 1 and 2, and details are not repeated here. In addition, it should be noted that the mobile phone may obtain a digital model and a scene digital model of one or some digital images in advance, for example, as shown in b in fig. 3, after the user enters the APP page, 4 live broadcast rooms, a first live broadcast room, a second live broadcast room, a third live broadcast room and a fourth live broadcast room are recommended for the user, if the digital model and the scene model information of the digital images corresponding to the four live broadcast rooms are not stored in advance on the mobile phone, the mobile phone may obtain the contents in advance from the cloud, in other words, the mobile phone does not need to obtain the data of the digital images and the digital models corresponding to the scenes from the cloud after the user enters the live broadcast rooms, but obtains the contents from the cloud in advance. In addition, it should be noted that the VR rendering or AR rendering process may be performed on the cloud server, which is already described in the embodiments corresponding to fig. 1 and fig. 2, and is not repeated herein.
The scheme provided by the embodiment of the application is introduced mainly from the perspective of interaction among the first terminal device, the second terminal device and the cloud server. It is to be understood that, in order to implement the above functions, the first terminal device, the second terminal device, and the cloud server include hardware structures and/or software modules corresponding to the respective functions. Those of skill in the art will readily appreciate that the various illustrative modules and algorithm steps described in connection with the embodiments disclosed herein may be implemented as hardware or combinations of hardware and computer software. Whether a function is performed as hardware or computer software drives hardware depends upon the particular application and design constraints imposed on the solution. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application.
Described from a hardware structure, in fig. 1 to fig. 3, the first terminal device, the second terminal device, and the cloud server may be implemented by one entity device, may also be implemented by multiple entity devices together, and may also be different logic function modules in one entity device, which is not specifically limited in this embodiment of the present application.
For example, the cloud server may be implemented by the communication device in fig. 4. Fig. 4 is a schematic diagram illustrating a hardware structure of a communication device according to an embodiment of the present application. The communication device comprises at least one processor 401. Optionally, the communication device may further include: a memory 403, a communication link 402, and at least one communication interface 404.
The processor 401 may be a general-purpose Central Processing Unit (CPU), a microprocessor, an application-specific integrated circuit (ASIC), or one or more ics for controlling the execution of programs in accordance with the present disclosure.
The communication link 402 may include a path for communicating information between the aforementioned components.
The communication interface 404 may be any device, such as a transceiver, for communicating with other devices or communication networks, such as ethernet, Radio Access Network (RAN), Wireless Local Area Network (WLAN), etc., or may be a communication interface between the communication module and other modules.
The memory 403 may be a read-only memory (ROM) or other types of static storage devices that can store static information and instructions, a Random Access Memory (RAM) or other types of dynamic storage devices that can store information and instructions, an electrically erasable programmable read-only memory (EEPROM), a compact disc read-only memory (CD-ROM) or other optical disc storage, optical disc storage (including compact disc, laser disc, optical disc, digital versatile disc, blu-ray disc, etc.), magnetic disk storage media or other magnetic storage devices, or any other medium that can be used to carry or store desired program code in the form of instructions or data structures and that can be accessed by a computer, but is not limited to such. The memory may be separate and coupled to the processor via a communication line 402. The memory may also be integral to the processor.
The memory 403 is used for storing computer-executable instructions for executing the present invention, and is controlled by the processor 401. Processor 401 is configured to execute computer-executable instructions stored in memory 403 to implement the methods of communication provided by the embodiments described below in the present application. The memory 403 may or may not be coupled to the processor 401.
Optionally, the computer-executable instructions in the embodiments of the present application may also be referred to as application program codes, which are not specifically limited in the embodiments of the present application.
In particular implementations, processor 401 may include one or more CPUs, as one embodiment.
In particular implementations, a communication device may include multiple processors, as one embodiment. Each of these processors may be a single-core (single-CPU) processor or a multi-core (multi-CPU) processor. A processor herein may refer to one or more devices, circuits, and/or processing cores for processing data (e.g., computer program instructions).
In a particular implementation, the communication device may further include an output device and an input device, as one embodiment. An output device is in communication with the processor 401 and may display information in a variety of ways. For example, the output device may be a Liquid Crystal Display (LCD), a Light Emitting Diode (LED) display device, a Cathode Ray Tube (CRT) display device, a projector (projector), or the like. The input device is in communication with the processor 401 and may receive user input in a variety of ways. For example, the input device may be a mouse, a keyboard, a touch screen device, or a sensing device, among others.
In the embodiment of the present application, the cloud server may be divided into the functional modules according to the above method example, for example, each functional module may be divided corresponding to each function, or two or more functions may be integrated into one processing module. The integrated module can be realized in a hardware mode, and can also be realized in a software functional module mode. It should be noted that, in the embodiment of the present application, the division of the module is schematic, and is only one logic function division, and there may be another division manner in actual implementation.
For example, in a case where each functional module is divided in an integrated manner, fig. 5 shows a schematic structural diagram of a cloud server.
As shown in fig. 5, the cloud server provided by the embodiment of the present application may include a transceiver unit 501,
in a specific embodiment, the transceiver 501 is configured to execute step 102 in the embodiment corresponding to fig. 1, step 103 in the embodiment corresponding to fig. 1, step 104 in the embodiment corresponding to fig. 1, and/or the transceiver 501 is further configured to execute other transceiver steps on the cloud server side in the embodiment corresponding to fig. 1.
In a specific embodiment, the transceiver 501 is configured to execute step 202 in the embodiment corresponding to fig. 2, step 203 in the embodiment corresponding to fig. 2, step 204 in the embodiment corresponding to fig. 2, step 206 in the embodiment corresponding to fig. 2, step 207 in the embodiment corresponding to fig. 2, and/or the transceiver 501 is further configured to execute other transceiver steps on the cloud server side in the embodiment corresponding to fig. 2. Processing unit 502 is configured to execute step 205 in the embodiment corresponding to fig. 2, and/or processing unit 502 is further configured to execute other processing steps on the cloud server side in the embodiment corresponding to fig. 2.
In the above embodiment, the cloud server is presented in a form of dividing each functional module in an integrated manner. Of course, in the embodiment of the present application, each function module of the CU network element and the DU network element may also be divided corresponding to each function, which is not specifically limited in the embodiment of the present application.
The first terminal device and the second terminal device referred to in the present application may represent any suitable terminal device, and may include (or may represent) devices such as a wireless transmit/receive unit (WTRU), a mobile station, a mobile node, a mobile device, a fixed or mobile subscription unit, a pager, a mobile phone, a handheld device, a vehicle-mounted device, a wearable device, a Personal Digital Assistant (PDA), a smart phone, a notebook computer, a touch screen device, a wireless sensor, or a consumer electronics device. In the following, the first terminal device and the second terminal device are taken as mobile phones for example:
fig. 6 is a block diagram illustrating a partial structure of a mobile phone related to a first terminal device and a second terminal device provided in an embodiment of the present invention. Referring to fig. 6, the handset includes: radio Frequency (RF) circuitry 601, memory 602, touch screen 603, sensor 604, audio circuitry 606, wireless fidelity (WiFi) module 605, processor 607, and power supply 608. Those skilled in the art will appreciate that the handset configuration shown in fig. 6 is not intended to be limiting and may include more or fewer components than those shown, or some components may be combined, or a different arrangement of components.
The following describes each component of the mobile phone in detail with reference to fig. 6: the RF circuit 601 may be used for receiving and transmitting information or receiving and transmitting signals during a call, and in particular, receives downlink information of a base station and then sends the received downlink information to the processor 607; in addition, the data for designing uplink is transmitted to the base station. In general, the RF circuit 601 includes, but is not limited to, an antenna, at least one Amplifier, a transceiver, a coupler, a Low Noise Amplifier (LNA), a duplexer, and the like. In addition, the RF circuit 601 may also communicate with networks and other devices via wireless communications. The wireless communication may use any communication standard or protocol, including but not limited to Global System for Mobile communication (GSM), General Packet Radio Service (GPRS), Code Division Multiple Access (Code Division Multiple Access, CDMA), Wideband Code Division Multiple Access (WCDMA), Long Term Evolution (LTE), 5th generation (5G) Mobile communication System or New Radio (NR) communication System, and future Mobile communication System, e-mail, Short Message Service (SMS), etc.
The memory 602 may be used to store software programs and modules, and the processor 607 executes various functional applications and data processing of the mobile phone by operating the software programs and modules stored in the memory 602. The memory 602 may mainly include a program storage area and a data storage area, wherein the program storage area may store an operating system, an application program required by at least one function (such as a sound playing function, an image playing function, etc.), and the like; the storage data area may store data (such as audio data, a phonebook, etc.) created according to the use of the cellular phone, and the like. Further, the memory 602 may include high speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid state storage device.
Touch screen 603, also referred to as a touch panel, touch screen 603 may include a touch sensitive surface 6031 and a display 6032. Among other things, touch-sensitive surface 6031 (e.g., a touch panel) may capture touch events on or near the touch-sensitive surface 6031 by a user of the cell phone (e.g., user manipulation on or near touch-sensitive surface 6031 using a finger, a stylus, or any other suitable object) and send the captured touch information to another device, such as processor 607. Among them, a touch event of a user near the touch-sensitive surface 6031 may be referred to as a hover touch; hover touch may refer to a user not having to directly contact the touchpad in order to select, move, or drag a target (e.g., an icon, etc.), but rather only having to be located near the first terminal device in order to perform a desired function. In the context of a hover touch application, the terms "touch," "contact," and the like do not imply a contact that is used to directly contact the touch screen, but rather a contact that is near or in proximity thereto. The touch-sensitive surface 6031 capable of floating touch control can be implemented by using capacitance, infrared light sensing, ultrasonic waves, or the like. Touch-sensitive surface 6031 may comprise two portions, a touch detection device and a touch controller. The touch detection device detects the touch direction of a user, detects a signal brought by touch operation and transmits the signal to the touch controller; the touch controller receives touch information from the touch sensing device, converts the touch information into touch point coordinates, and sends the touch point coordinates to the processor 607, and the touch controller can also receive and execute instructions sent by the processor 607. Additionally, touch-sensitive surface 6031 can be implemented using various types of resistive, capacitive, infrared, and surface acoustic waves. A display (also referred to as a display screen) 6032 may be used to display information entered by or provided to the user as well as various menus of the handset. The display 6032 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like. Touch-sensitive surface 6031 may be overlaid on display 6032 and, upon detection of a touch event on or near touch-sensitive surface 6031, communicated to processor 607 for determination of the type of touch event, and processor 606 may then provide a corresponding visual output on display 6032 in accordance with the type of touch event. Although in fig. 6, touch-sensitive surface 6031 and display 6032 are two separate components to implement the input and output functions of the handset, in some embodiments, touch-sensitive surface 6031 and display 6032 may be integrated to implement the input and output functions of the handset. It is understood that 603 the touch screen is formed by stacking multiple layers of materials, only the touch sensitive surface (layer) and the display screen (layer) are shown in the embodiment of the present application, and other layers are not described in the embodiment of the present application. In addition, in some other embodiments of the present application, the touch-sensitive surface 6031 may be covered on the display 6032, and the size of the touch-sensitive surface 6031 is larger than the size of the display 6032, so that the display 6032 is completely covered under the touch-sensitive surface 6031, or the touch-sensitive surface 6031 may be configured on the front of the mobile phone in a full panel manner, that is, the touch of the user on the front of the mobile phone can be sensed by the mobile phone, so that the full touch experience on the front of the mobile phone can be achieved. In other embodiments, the touch-sensitive surface 6031 can be disposed on the front of the handset in a full-panel manner, and the display 6032 can also be disposed on the front of the handset in a full-panel manner, so that a bezel-free structure can be implemented on the front of the handset.
The handset may also include at least one sensor 604, such as a light sensor, motion sensor, and other sensors. Specifically, the light sensor may include an ambient light sensor that adjusts the brightness of the display 6032 based on the intensity of ambient light, and a proximity sensor that turns off the display 603 and/or backlight when the phone is moved to the ear. As one of the motion sensors, the accelerometer sensor can detect the magnitude of acceleration in each direction (generally, three axes), can detect the magnitude and direction of gravity when stationary, and can be used for applications of recognizing the posture of a mobile phone (such as horizontal and vertical screen switching, related games, magnetometer posture calibration), vibration recognition related functions (such as pedometer and tapping), and the like; as for other sensors such as a gyroscope, a barometer, a hygrometer, a thermometer, and an infrared sensor, which can be configured on the mobile phone, further description is omitted here.
Audio circuitry 606, speaker 6062, and microphone 6061 may provide an audio interface between the user and the handset. The audio circuit 606 may transmit the electrical signal converted from the received audio data to the speaker 6062, and convert the electrical signal into a sound signal by the speaker 6062 and output the sound signal; on the other hand, the microphone 6061 converts the collected sound signal into an electric signal, which is received by the audio circuit 606 and converted into audio data, and the audio data is processed by the audio data output processor 607 and then transmitted to, for example, another cellular phone via the RF circuit 601, or the audio data is output to the memory 602 for further processing.
WiFi belongs to short-range wireless transmission technology, and the handset can help the user to send and receive e-mail, browse web pages, access streaming media, etc. through the WiFi module 605, which provides wireless broadband internet access for the user. Although fig. 6 shows the WiFi module 605, it is understood that it does not belong to the essential constitution of the handset, and may be omitted entirely as needed within the scope not changing the essence of the invention.
The processor 607 is a control center of the mobile phone, connects various parts of the entire mobile phone by using various interfaces and lines, and performs various functions of the mobile phone and processes data by operating or executing software programs and/or modules stored in the memory 602 and calling data stored in the memory 602, thereby performing overall monitoring of the mobile phone. Optionally, processor 607 may include one or more processing units; preferably, the processor 607 may integrate an application processor, which mainly handles operating systems, user interfaces, application programs, etc., and a modem processor, which mainly handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into processor 607.
The handset also includes a power supply 608 (e.g., a battery) for powering the various components, which may preferably be logically coupled to the processor 607 via a power management system to manage charging, discharging, and power consumption via the power management system.
Although not shown, the mobile phone may further include a camera, a bluetooth module, etc., which are not described herein.
In the embodiment of the present application, the first terminal device and the second terminal device may be divided into the functional modules according to the above method example, for example, each functional module may be divided corresponding to each function, or two or more functions may be integrated into one processing module. The integrated module can be realized in a hardware mode, and can also be realized in a software functional module mode. It should be noted that, in the embodiment of the present application, the division of the module is schematic, and is only one logic function division, and there may be another division manner in actual implementation.
For example, in the case of dividing each functional module in an integrated manner, fig. 7 shows a schematic structural diagram of a first terminal device.
In a specific embodiment, the transceiver 701 is configured to perform steps 104 and 105 in the embodiment corresponding to fig. 1, and/or the transceiver 701 is further configured to perform other transceiving steps on the first terminal device side in the embodiment corresponding to fig. 1. The processing unit 702 is configured to execute step 105 in the above embodiment corresponding to fig. 1, and/or the processing unit 702 is further configured to execute other processing steps on the first terminal device side in the embodiment corresponding to fig. 1.
In a specific embodiment, the transceiver 701 is configured to perform steps 203, 204, 206, and 207 in the embodiment corresponding to fig. 2, and/or the transceiver 701 is further configured to perform other transceiving steps on the first terminal device side in the embodiment corresponding to fig. 1. The playing unit 703 is configured to execute step 207 in the embodiment corresponding to fig. 2.
In the case where the functional modules are divided in an integrated manner, fig. 8 shows a schematic structural diagram of a second terminal device.
In a specific embodiment, the transceiver unit 801 is configured to execute steps 101 and 102 in the embodiment corresponding to fig. 1, step 103 in the embodiment corresponding to fig. 1, step 104 in the embodiment corresponding to fig. 1, and/or the transceiver unit 801 is further configured to execute other transceiver steps on the second terminal device side in the embodiment corresponding to fig. 1. The processing unit 802 is configured to execute the step 101 in the embodiment corresponding to fig. 1, and/or the processing unit 802 is further configured to execute other transceiving steps on the second terminal device side in the embodiment corresponding to fig. 1.
In a specific embodiment, the transceiver unit 801 is configured to execute steps 201 and 202 in the embodiment corresponding to fig. 2, step 203 in the embodiment corresponding to fig. 2, and/or the transceiver unit 801 is further configured to execute other transceiver steps on the second terminal device side in the embodiment corresponding to fig. 2. A processing unit 802, configured to execute step 201 in the embodiment corresponding to fig. 2, and/or the processing unit 802 is further configured to execute other processing steps on the second terminal device side in the embodiment corresponding to fig. 2.
Since the cloud server, the first terminal device, and the second terminal device provided in the embodiment of the present application may be used to execute the above positioning method, the technical effect obtained by the method may refer to the above method embodiment, and details are not described herein again.
In the above embodiments, the implementation may be wholly or partially realized by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product.
The computer program product includes one or more computer instructions. When loaded and executed on a computer, cause the processes or functions described in accordance with the embodiments of the application to occur, in whole or in part. The computer may be a general purpose computer, a special purpose computer, a network of computers, or other programmable device. The computer instructions may be stored in a computer readable storage medium or transmitted from one computer readable storage medium to another, for example, from one website site, computer, server, or data center to another website site, computer, server, or data center via wired (e.g., coaxial cable, fiber optic, Digital Subscriber Line (DSL)) or wireless (e.g., infrared, wireless, microwave, etc.). The computer-readable storage medium can be any available medium that a computer can store or a data storage device, such as a server, a data center, etc., that is integrated with one or more available media. The usable medium may be a magnetic medium (e.g., floppy Disk, hard Disk, magnetic tape), an optical medium (e.g., DVD), or a semiconductor medium (e.g., Solid State Disk (SSD)), among others.
Those skilled in the art will appreciate that all or part of the steps in the methods of the above embodiments may be implemented by associated hardware instructed by a program, which may be stored in a computer-readable storage medium, and the storage medium may include: ROM, RAM, magnetic or optical disks, and the like.
The foregoing detailed description is directed to a playing method, a playing system, and a storage medium provided in the embodiments of the present application, and a specific example is applied in the present application to explain the principles and embodiments of the present application, and the description of the foregoing embodiments is only used to help understand the method and the core ideas of the present application; meanwhile, for a person skilled in the art, according to the idea of the present application, there may be variations in the specific embodiments and the application scope, and in summary, the content of the present specification should not be construed as a limitation to the present application.

Claims (17)

1. A playback method, comprising:
the method comprises the steps that a first terminal device receives motion capture data through a wide area network, and the motion capture data are sent to the first terminal device by a second terminal device through a cloud server;
the first terminal equipment receives voice information through a wide area network, the voice information is sent to the first terminal equipment by second terminal equipment through a cloud server, and the first terminal equipment, the second terminal equipment and the cloud server share the voice information in real time;
the first terminal equipment performs Augmented Reality (AR) rendering or Virtual Reality (VR) rendering on the digital image according to the received motion capture data to obtain a first animation;
and the first terminal equipment synchronously plays the first animation and the voice information.
2. The playback method according to claim 1, wherein the voice information and the motion capture data include the same time stamp;
the first terminal device synchronously plays the first animation and the voice information, and the method comprises the following steps:
and the first terminal equipment synchronously plays the first animation and the voice information according to the same timestamp.
3. A playback method, comprising:
the method comprises the steps that a cloud server receives first information sent by first terminal equipment through a wide area network, wherein the first information is used for calculating a 3D rendering visual angle parameter of a digital image;
the cloud server receives motion capture data and voice information sent by second terminal equipment through a wide area network, and the first terminal equipment, the second terminal equipment and the cloud server share the voice information in real time;
the cloud server performs Augmented Reality (AR) rendering or Virtual Reality (VR) rendering on the digital image according to the motion capture data and the first information to obtain a first animation;
the cloud server encodes the first animation and the voice information and performs IP packaging to obtain a video;
and the cloud server sends the video to the first terminal device, and the video is used for playing after being decoded by the first terminal device.
4. A playback method, comprising:
the method comprises the steps that first information is sent to a cloud server by a first terminal device, the first information is used for calculating a 3D rendering visual angle parameter of a digital image, and the 3D rendering visual angle parameter is used for carrying out augmented reality AR rendering or virtual reality VR rendering on the digital image by the cloud server in combination with motion capture data to obtain a first animation;
and the first terminal equipment receives a video sent by the cloud server, wherein the video is obtained by the cloud server according to a first animation code.
5. A playback system, comprising: the system comprises a first terminal device, a second terminal device and a cloud server;
the second terminal device is used for sending motion capture data and voice information to the cloud server through a wide area network;
the cloud server is used for sending the motion capture data and the voice information to the first terminal equipment;
the first terminal device is used for performing augmented reality AR rendering or virtual reality VR rendering on the digital image according to the received motion capture data to obtain a first animation;
the first terminal device is further configured to play the first animation and the voice information synchronously.
6. A playback system, comprising: the system comprises a first terminal device, a second terminal device and a cloud server;
the first terminal device is the first terminal device described in claim 4, the second terminal device is the second terminal device described in claim 3, and the cloud server is the cloud server described in claim 3.
7. A terminal device, comprising:
a transceiving unit, configured to receive motion capture data via a wide area network, where the motion capture data is sent to the first terminal device by a second terminal device via a cloud server;
the receiving and sending unit is further configured to receive voice information through a wide area network, the voice information is sent to the first terminal device by a second terminal device through a cloud server, and the first terminal device, the second terminal device and the cloud server share the voice information in real time;
the processing unit is used for performing augmented reality AR rendering or virtual reality VR rendering on the digital image according to the motion capture data received by the transceiving unit to obtain a first animation;
and the playing unit is used for synchronously playing the first animation and the voice information.
8. A cloud server, comprising:
the receiving and sending unit is used for receiving first information sent by first terminal equipment through a wide area network, and the first information is used for calculating a 3D rendering visual angle parameter of a digital image;
the receiving and sending unit is further used for receiving motion capture data and voice information sent by second terminal equipment through a wide area network, and the first terminal equipment, the second terminal equipment and the cloud server share the voice information in real time;
the processing unit is used for performing Augmented Reality (AR) rendering or Virtual Reality (VR) rendering on the digital image according to the motion capture data and the first information received by the transceiving unit to obtain a first animation;
the processing unit is further configured to encode the first animation and the voice information and perform IP encapsulation to obtain a video;
the transceiver unit is further configured to send the video to the first terminal device, where the video is used for playing after being decoded by the first terminal device.
9. A terminal device, comprising:
the receiving and sending unit is used for sending first information to the cloud server, the first information is used for calculating a 3D rendering visual angle parameter of the digital image, and the 3D rendering visual angle parameter is used for performing Augmented Reality (AR) rendering or Virtual Reality (VR) rendering on the digital image by the cloud server in combination with motion capture data to obtain a first animation;
the receiving and sending unit is further configured to receive a video sent by the cloud server, where the video is obtained by the cloud server according to a first animation code.
10. A computer-readable storage medium, which when executed on a computer device, causes the computer device to perform the method of claim 1 or 2.
11. A computer-readable storage medium, which when executed on a computer device, causes the computer device to perform the method of claim 3.
12. A computer-readable storage medium, which when executed on a computer device, causes the computer device to perform the method of claim 4.
13. A computer program product enabling a computer to carry out the method of claim 1 or 2 when run on the computer.
14. A computer program product enabling a computer to carry out the method of claim 3 when run on the computer.
15. A computer program product enabling a computer to carry out the method of claim 4 when run on the computer.
16. A terminal device, characterized in that the device comprises a memory, a processor, the memory having code and data stored therein, the memory being coupled to the processor, the processor executing the code in the memory causing the device to perform the method according to claim 1 or 2, or to perform the method according to claim 4.
17. A cloud server, wherein the device comprises a memory, a processor, the memory having code and data stored therein, the memory coupled to the processor, the processor executing the code in the memory to cause the device to perform the method of claim 3.
CN201911007650.8A 2019-10-22 2019-10-22 Playing method and playing system Pending CN112702611A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201911007650.8A CN112702611A (en) 2019-10-22 2019-10-22 Playing method and playing system
PCT/CN2020/122672 WO2021078182A1 (en) 2019-10-22 2020-10-22 Playback method and playback system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911007650.8A CN112702611A (en) 2019-10-22 2019-10-22 Playing method and playing system

Publications (1)

Publication Number Publication Date
CN112702611A true CN112702611A (en) 2021-04-23

Family

ID=75504713

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911007650.8A Pending CN112702611A (en) 2019-10-22 2019-10-22 Playing method and playing system

Country Status (2)

Country Link
CN (1) CN112702611A (en)
WO (1) WO2021078182A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113289332A (en) * 2021-06-17 2021-08-24 广州虎牙科技有限公司 Game interaction method and device, electronic equipment and computer-readable storage medium

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102340690A (en) * 2011-09-13 2012-02-01 苏州美娱网络科技有限公司 Interactive television program system and realization method
CN104796393A (en) * 2014-05-30 2015-07-22 厦门极致互动网络技术有限公司 Online game system and method based on server real-time rendering
CN106385681A (en) * 2016-09-21 2017-02-08 平越 Virtual reality entertainment system and method thereof
CN107613310A (en) * 2017-09-08 2018-01-19 广州华多网络科技有限公司 A kind of live broadcasting method, device and electronic equipment
CN108597028A (en) * 2018-04-11 2018-09-28 北京邮电大学 A kind of method, apparatus and equipment of continuous load Web AR objects
CN109126122A (en) * 2017-06-16 2019-01-04 上海拆名晃信息科技有限公司 A kind of cloud game network system realization for virtual reality
CN109195020A (en) * 2018-10-11 2019-01-11 三星电子(中国)研发中心 A kind of the game live broadcasting method and system of AR enhancing
CN109215132A (en) * 2017-06-30 2019-01-15 华为技术有限公司 A kind of implementation method and equipment of augmented reality business
CN109729368A (en) * 2018-11-16 2019-05-07 深圳市赛亿科技开发有限公司 The method and system of VR live video processing

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9984499B1 (en) * 2015-11-30 2018-05-29 Snap Inc. Image and point cloud based tracking and in augmented reality systems
CN105915849A (en) * 2016-05-09 2016-08-31 惠州Tcl移动通信有限公司 Virtual reality sports event play method and system
CN107979763B (en) * 2016-10-21 2021-07-06 阿里巴巴集团控股有限公司 Virtual reality equipment video generation and playing method, device and system
CN106790621A (en) * 2016-12-30 2017-05-31 广州幻境科技有限公司 A kind of cloud of motion capture data shares control method and its system
CN107103801B (en) * 2017-04-26 2020-09-18 北京大生在线科技有限公司 Remote three-dimensional scene interactive teaching system and control method
WO2019037073A1 (en) * 2017-08-25 2019-02-28 深圳市瑞立视多媒体科技有限公司 Method, device and sever for data synchronization

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102340690A (en) * 2011-09-13 2012-02-01 苏州美娱网络科技有限公司 Interactive television program system and realization method
CN104796393A (en) * 2014-05-30 2015-07-22 厦门极致互动网络技术有限公司 Online game system and method based on server real-time rendering
CN106385681A (en) * 2016-09-21 2017-02-08 平越 Virtual reality entertainment system and method thereof
CN109126122A (en) * 2017-06-16 2019-01-04 上海拆名晃信息科技有限公司 A kind of cloud game network system realization for virtual reality
CN109215132A (en) * 2017-06-30 2019-01-15 华为技术有限公司 A kind of implementation method and equipment of augmented reality business
CN107613310A (en) * 2017-09-08 2018-01-19 广州华多网络科技有限公司 A kind of live broadcasting method, device and electronic equipment
CN108597028A (en) * 2018-04-11 2018-09-28 北京邮电大学 A kind of method, apparatus and equipment of continuous load Web AR objects
CN109195020A (en) * 2018-10-11 2019-01-11 三星电子(中国)研发中心 A kind of the game live broadcasting method and system of AR enhancing
CN109729368A (en) * 2018-11-16 2019-05-07 深圳市赛亿科技开发有限公司 The method and system of VR live video processing

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113289332A (en) * 2021-06-17 2021-08-24 广州虎牙科技有限公司 Game interaction method and device, electronic equipment and computer-readable storage medium

Also Published As

Publication number Publication date
WO2021078182A1 (en) 2021-04-29

Similar Documents

Publication Publication Date Title
US11120616B2 (en) Method for implementing virtual scene conversion and related apparatus
CN107741809B (en) Interaction method, terminal, server and system between virtual images
WO2019184889A1 (en) Method and apparatus for adjusting augmented reality model, storage medium, and electronic device
US11790612B2 (en) Information display method and device, terminal, and storage medium
WO2019034142A1 (en) Three-dimensional image display method and device, terminal, and storage medium
WO2016173427A1 (en) Method, device and computer readable medium for creating motion blur effect
WO2021077923A1 (en) Method of controlling display device, and portable device
CN107087137B (en) Method and device for presenting video and terminal equipment
CN109646944B (en) Control information processing method, control information processing device, electronic equipment and storage medium
CN108513671B (en) Display method and terminal for 2D application in VR equipment
WO2015131767A1 (en) Video processing method and apparatus
CN111311757B (en) Scene synthesis method and device, storage medium and mobile terminal
WO2021136266A1 (en) Virtual image synchronization method and wearable device
WO2021073293A1 (en) Animation file generating method and device, and storage medium
KR102633468B1 (en) Method and device for displaying hotspot maps, and computer devices and readable storage media
CN110750734A (en) Weather display method and device, computer equipment and computer-readable storage medium
CN109032468A (en) A kind of method and terminal of adjustment equipment parameter
CN108564613A (en) A kind of depth data acquisition methods and mobile terminal
CN111045945B (en) Method, device, terminal, storage medium and program product for simulating live broadcast
CN111083554A (en) Method and device for displaying live gift
WO2021104162A1 (en) Display method and electronic device
WO2021078182A1 (en) Playback method and playback system
CN109117037A (en) A kind of method and terminal device of image procossing
CN113485596B (en) Virtual model processing method and device, electronic equipment and storage medium
CN109343782A (en) A kind of display methods and terminal

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20210423