CN114079800A - Virtual character performance method, device, system and computer readable storage medium - Google Patents

Virtual character performance method, device, system and computer readable storage medium Download PDF

Info

Publication number
CN114079800A
CN114079800A CN202111132440.9A CN202111132440A CN114079800A CN 114079800 A CN114079800 A CN 114079800A CN 202111132440 A CN202111132440 A CN 202111132440A CN 114079800 A CN114079800 A CN 114079800A
Authority
CN
China
Prior art keywords
performance
information
virtual character
control information
client
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111132440.9A
Other languages
Chinese (zh)
Inventor
曹传宇
唐崎森
江海明
陈锦锋
周松泉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Youban Technology Co ltd
Original Assignee
Shenzhen Youban Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Youban Technology Co ltd filed Critical Shenzhen Youban Technology Co ltd
Priority to CN202111132440.9A priority Critical patent/CN114079800A/en
Publication of CN114079800A publication Critical patent/CN114079800A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8126Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8547Content authoring involving timestamps for synchronizing content
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Abstract

The invention belongs to the technical field of network live broadcast, and particularly relates to a virtual character performance method, an anchor terminal, a virtual character performance system and a computer readable storage medium. The virtual character performance method comprises the following steps: acquiring performance triggering information, wherein the performance triggering information comprises voice information and/or performance instructions; acquiring performance control information according to the performance trigger information; and outputting the performance control information to the client so that the client controls the virtual character to perform the performance according to the performance control information. Therefore, the invention can obviously improve the performance effect of the virtual character, simultaneously reduce the live broadcast cost required by the virtual anchor, the network cost for users to watch the virtual character performance and the like, can be popularized and applied in a larger range in the form of the virtual character performance, and bring considerable social and economic benefits.

Description

Virtual character performance method, device, system and computer readable storage medium
Technical Field
The invention belongs to the technical field of network live broadcast, and particularly relates to a virtual character performance method, an anchor terminal, a virtual character performance and a computer readable storage medium.
Background
In daily life, various live broadcasts enrich people's leisure time. In various live broadcasts, a live broadcast method through virtual character images is adopted, the facial expressions or limb actions and the like of a main broadcast are captured through a motion capture technology, and video output is realized through 3D or 2D virtual character images so as to achieve the effect of live broadcast. Live broadcast is carried out through the virtual character image, and the requirements of partial anchor and audiences are met, so that the live broadcast is welcomed and popularized to a certain extent. The motion capture is a tracker arranged at the key position of a moving object. The method comprises the following steps: motion capture english Motion capture, abbreviated as Mocap. It relates to the measurement of dimensions, the positioning of objects in physical space, and the determination of orientation, etc. data that can be directly understood and processed by a computer. A tracker is arranged at a key part of a moving object, the position of the tracker is captured by a Motion capture system, and data of three-dimensional space coordinates are obtained after the data are processed by a computer, so that the animation in the virtual world is obtained.
However, the input accuracy and the expressiveness of the information collected by the capturing technique are ceiling, and the degree of exaggeration of the expressive power of the face and the motion is limited, so that the effect of the performance is not good. And the capturing technology has high cost and threshold for operation, and is difficult to be widely used. In addition, the live broadcasting cost of the video stream is high, and the requirements on equipment and networks of the client are high. How to solve the above technical problems needs to be considered urgently by those skilled in the art.
In view of the above problems, those skilled in the art have sought solutions.
The foregoing description is provided for general background information and is not admitted to be prior art.
Disclosure of Invention
The invention solves the technical problems that in the prior art, the problems caused by technologies such as motion capture, face capture and the like are needed for virtual character performance, wherein the problems include but are not limited to the problem that the accuracy of capture technology acquisition is not high, so that the performance effect is wrong or the performance effect is poor; professional equipment is expensive, and operation and implementation have certain thresholds, so that the popularization is difficult. Meanwhile, for live broadcasting, the live broadcasting cost of the video stream is high, and the fluency has a certain demand on the equipment and the network of the client. In view of the above, a virtual character performing method, an anchor terminal, a virtual character performing system, and a computer-readable storage medium are provided to solve the above technical problems.
The technical problem to be solved by the invention is realized by adopting the following technical scheme:
the invention provides a virtual character performance method, which comprises the following steps: acquiring performance triggering information, wherein the performance triggering information comprises voice information and/or performance instructions; acquiring performance control information according to the performance trigger information; and outputting the performance control information to the client so that the client controls the virtual character to perform the performance according to the performance control information.
Further, the step of obtaining the show trigger information includes: responding to user operation, matching corresponding performance instructions, wherein the performance instructions comprise one of single instructions and/or combined instructions; and/or automatically acquiring preset performance instructions from the natural performance database.
Further, the step of acquiring the show control information according to the show trigger information includes: generating performance control information according to the voice information; and/or matching the corresponding performance control information according to the performance instruction.
Further, the step of acquiring the show control information according to the show trigger information includes: acquiring a timestamp of the performance control information according to the timestamp in the voice information; and/or setting a timestamp of the show control information by a preset protocol.
Further, the show control information includes: at least one item of action information, expression information, sound effect information, special effect information, background information, prop information, pendant information and control information.
Further, after the step of obtaining the show control information according to the show trigger information, the method includes: controlling the virtual character to realize performance according to the performance control information; and displaying the performance of the virtual character in real time.
Further, the method further comprises: when the number of the virtual roles is more than or equal to two, acquiring target role information; the step of acquiring the show control information according to the show trigger information includes: and acquiring the performance control information of the target role information according to the performance trigger information.
Further, the virtual character performing method further includes the following steps: acquiring interaction request information sent by a client; responding to the interaction request information, and generating interaction feedback information; and outputting the interactive feedback information to the client.
The invention also provides an anchor terminal, which comprises a processor and a memory: the processor is configured to execute a computer program stored in the memory to implement the virtual character rendering method steps as described above.
The invention also provides a virtual anchor performance system which is characterized by comprising an anchor end and a client end; the system comprises a main broadcasting terminal, a client and a broadcasting terminal, wherein the main broadcasting terminal is used for acquiring performance triggering information, the performance triggering information comprises voice information and/or performance instructions, acquiring performance control information according to the performance triggering information and outputting the performance control information to the client; and the client is used for receiving the performance control information transmitted by the main broadcasting terminal.
Further, in the virtual character performance system described above, after the step of receiving the performance control information transmitted by the main broadcasting terminal, the client includes: rendering by local resources through performance control information to realize performance of virtual characters
The present invention also provides a computer readable storage medium storing a computer program which, when executed by a processor, implements the steps of the virtual character rendering method as described above.
The invention also provides a virtual character performance method, an anchor terminal, a virtual character performance system and a computer readable storage medium. The virtual character performance method comprises the following steps: acquiring performance triggering information, wherein the performance triggering information comprises voice information and/or performance instructions; acquiring performance control information according to the performance trigger information; and outputting the performance control information to the client so that the client controls the virtual character to perform the performance according to the performance control information. Therefore, the invention triggers the performance through the instruction and voice input mode, reduces the cost of virtual character performance, particularly for live broadcasting, no special acquisition equipment is needed, and the performance effect can be obviously improved, and the action and expression can be more flexible, magical, cool, exaggerated, delicate and smooth. Meanwhile, for live broadcasting, the network pressure output by a live broadcasting end is reduced by outputting a data stream instead of a video stream, the network cost and consumption of data received by a client are reduced, and the virtual character performance form can be popularized and applied in a larger range, so that considerable social and economic benefits are brought. In addition, the virtual character performance method provided by the embodiment of the invention can also acquire performance instructions in various ways, and can meet the manual performance requirement of the anchor end or automatically acquire preset performance instructions from the database through the anchor self-defining performance actions, so that natural actions or linkage actions are added, and the virtual character performance is smoother. Meanwhile, when the voice input triggers the performance, the voice information and the performance control information are bound in a timestamp mode, so that the virtual character can play the upper mouth shape more naturally. Meanwhile, the performance of the virtual character is monitored in real time, so that timely processing can be carried out before the virtual character is transmitted when an accident happens. In addition, the simultaneous and on-screen performance of one or more virtual characters is supported, so that the performance is richer. And the interaction request information sent by the client side can be received to correspond to the response, so that the interaction between the anchor and the audience is realized, various requirements are met, the interestingness of performance is increased, the operation of a user is reduced, the convenience of the user is increased, and the use experience of the user is improved.
The foregoing description is only an overview of the technical solutions of the present invention, and in order to make the technical means of the present invention more clearly understood, the present invention may be implemented in accordance with the content of the description, and in order to make the above and other objects, features, and advantages of the present invention more clearly understood, the following preferred embodiments are specifically described in detail with reference to the accompanying drawings.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings needed to be used in the description of the embodiments or the prior art will be briefly introduced below, and it is obvious that the drawings in the following description are some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to these drawings without creative efforts.
Fig. 1 is a schematic flow chart of a virtual character performing method according to a first embodiment of the present invention;
fig. 2 is a schematic diagram of a first structure of a anchor according to a second embodiment of the present invention;
fig. 3 is a second structural diagram of the anchor according to a second embodiment of the present invention;
fig. 4 is a schematic structural diagram of a virtual character performance system according to a third embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is described in further detail below with reference to the accompanying drawings and embodiments. It is to be understood that the described embodiments are merely a few embodiments of the invention, and not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The embodiments of the present invention will be described in detail with reference to the accompanying drawings.
First embodiment
Fig. 1 is a flowchart illustrating a virtual character performing method according to a first embodiment of the present invention. For a clear description of the virtual character rendering method according to the first embodiment of the present invention, please refer to fig. 1.
The virtual character performing method provided by the embodiment is applied to a main broadcasting end. The anchor terminal can be but not limited to including at least one in PC, mobile terminal, server etc. can gather the information that obtains triggering the performance, and can pass through high performance transport protocol packing synthetic data stream, initiate remote network transmission, and can real-time visual monitoring to the effect preview of audio input and instruction input to can realize: one anchor operates one control console to support a plurality of anchors to input audio and instructions simultaneously, data streams are packed and synthesized through a high-performance transmission protocol and uploaded to a server, and the server distributes the data streams to massive user clients after confluence. Preferably, the anchor terminal may include a console of the anchor and a server, the console acquires the performance trigger information, the console and/or the server acquires the performance control information, and the server transmits the performance control information to a large number of clients. Specifically, the virtual character may include one or more characters, and the specific character may also have various forms such as 2D or 3D.
The specific implementation manner of the virtual character rendering method provided by this embodiment includes the following steps:
step S1: and acquiring performance trigger information, wherein the performance trigger information comprises voice information and/or performance instructions.
In one embodiment, at step S1: the step of obtaining the show trigger information comprises the following steps: responding to user operation, matching corresponding performance instructions, wherein the performance instructions comprise one of single instructions and/or combined instructions; and/or automatically acquiring preset performance instructions from the natural performance database.
In one embodiment, step S1 is mainly used to collect information that can trigger the performance, which may include, but is not limited to, performance instructions or voice information. The voice information collection technology is relatively existing, and preferably, the voice is collected through a microphone, and the audio output stream through other software, for example, the audio played by other software in the host terminal is collected.
In one embodiment, the means for obtaining show instructions may include, but is not limited to, user manipulation or automatic retrieval from a database of natural shows. Specifically, the obtaining path of the user operation is obtained through a peripheral device, an input device, and the like connected to the host side, where the peripheral device or the input device may include, but is not limited to, a mouse, a keyboard, a handle, a touch screen, and the like. The show command can also comprise a single command and/or a combined command. Specifically, after the performance command and the performance control information are matched and correspond to each other, the virtual character can be correspondingly controlled. Wherein, the single instruction is a simple instruction, such as controlling a virtual character or running or jumping; the combination command may be a command to implement a plurality of continuous show control information, for example, a combination command can control the virtual character to run and jump again. Furthermore, after editing or customizing, a combination of several instructions can be realized, and more contents are added, for example, by triggering the combination instruction, the virtual character is controlled to run and jump again, and the connection action is also realized. However, it is understood that for the implementation of the combination command, a corresponding matching relationship between the show command and the subsequent show control information is required. And the user can edit and customize the performance instruction, and can support various operation modes according to the instruction type and characteristics, wherein the operation modes include but are not limited to one-time performance instruction triggered by clicking, instruction cycle triggered by clicking/cancellation clicked again and the like. The instructions can be edited and customized by a user, and after the subsequent instructions correspond to the performances in an associated manner, the abundance of the virtual character performances can be greatly enriched, so that the requirements of the program effects which are as exaggerated and as cool as possible during the virtual character performances are met. In addition to manual acquisition, there is also included an approach to automatically acquire from the database of natural performances. It will be appreciated that for avatars in a performance, rather than being rigid and standing motionless, there may be some natural actions, behaviors, expressions, e.g. natural swaying, slight changes in facial expressions, such as blinking eyes, etc. The method can also comprise natural connection of some back and forth actions, for example, two actions for controlling the virtual character to jump first and then run are input by the anchor, in order to make the performance natural and smooth, a connection action needs to be added between the two actions, and the performance instruction corresponding to the connection action can be automatically obtained from the natural performance database without the secondary intervention of the anchor. Furthermore, the natural performance database can be stored in the local anchor terminal or in the cloud terminal, the cloud terminal shared by a plurality of anchor terminals stores data, the specific storage mode is not limited, and the natural performance database can be preferably customized, namely, a natural connection action can be added in the database by a user, the automatically acquired time can be determined, and the like, so that the requirement of enriching the performance program effect can be met.
Step S2: and acquiring performance control information according to the performance trigger information.
In one embodiment, it is understood that the information triggering the performance is obtained in step S1, and the performance control information for controlling the virtual character to complete the performance is performance control information capable of controlling the virtual character to perform the performance, where the performance control information includes: at least one item of action information, expression information, sound effect information, special effect information, background information, prop information, pendant information and control information. Specifically, the motion information is information for controlling the limb movement of the virtual character, and specifically includes but is not limited to: walking, running, jumping, rotating the body, etc.; the expression information may specifically include but is not limited to: the facial expression of the virtual character is expressed, the information of the mood such as joy, anger, sadness, fear and the like is embodied, the face-to-mouth shape matched with the voice information, the expression action such as blink and the like in natural action and the like can also be provided; the sound effect information may include some atmosphere of performance, such as explosion sound when a booming message is displayed, can laughter for increasing comedy effect, and the like, and may include sound effect information for controlling the birth of the virtual character, and the sound effect information may be generated according to the voice information, such as processing the sound to obtain a sound effect for performance, or may be directly the voice information, and is not particularly limited; the special effect information can be matched with explosion light effect and the like when bombing messages are displayed or light effect and the like of some live broadcast rooms; the live broadcast room is not limited to the virtual character, but also includes a background, which is the content included in the background information, and the background can be composed of a long shot, a medium shot, a short shot, and the like, for example, a table in front of the virtual character is the short shot, a stool at the back is the medium shot, a scene such as a classroom or a remote mountain behind the virtual character is the long shot, and the like, and is generally static and does not interact with the virtual character; the pendant information may be some arrangements placed on the avatar, such as a pen in a hand, a stick, or a wing behind, or may be some virtual articles that can interact with the avatar in some scenes, such as a cup on a table, a doll, etc.; the property information is that the property can be a disposable property generally used for a user to realize interaction through a client and/or a main broadcasting terminal, and the disposable property comprises a rocket brushed in a live broadcasting room and the like, and the properties are displayed at the client; the control information is small plug-ins used for showing the client side in some live broadcast rooms, such as voting and purchasing, so as to realize interaction between the client side and the main broadcast side. It can be understood that the above show control information basically covers all actions required by the virtual character to complete the performance and the requirements of the arrangement of the live broadcast room, so as to fully meet the performance requirements required by the user. And all the performance control information can be controlled by the user through the performance trigger information, for example, the background can be switched through the performance instruction, for example, for a live broadcast room of a network course, the user can control the background information in the performance control information to complete the switching from the PPT to the demonstration video, and the like; and the interaction such as questioning and the like can be completed by controlling the control information. It will be appreciated that the patterns of performance performed in real life are complex and difficult to be exhaustive, and thus the above description is illustrative of the technology and not limiting.
In one embodiment, at step S2: the step of obtaining the show control information according to the show trigger information includes: generating performance control information according to the voice information; and/or matching the corresponding performance control information according to the performance instruction.
In one embodiment, at step S2: the step of obtaining the show control information according to the show trigger information includes: acquiring a timestamp of the performance control information according to the timestamp in the voice information; and/or setting a timestamp of the show control information by a preset protocol.
In one embodiment, specifically, the performance control information is generated by voice information, that is, voice content can be identified by voice analysis and other technologies, and the performance control information including expression information is correspondingly generated by detecting the size of sound wave and other technologies, so that the technical effect of the virtual character on the mouth shape according to voice can be realized. The corresponding performance control information can be specifically called through the type of voice recognition, which is similar to a voice assistant, for example, in live broadcast in a network classroom, a main broadcast can acquire the performance control information of the control information through voice information matching so as to control the playing or page changing of the PPT, and the like. Further, when the performance control information is acquired based on the voice information, the time stamp of the performance control information may also be acquired based on the time stamp in the voice information. The time stamps are synchronized, so that when voice information is displayed on the virtual character, the effect of completely synchronizing with voice output when mouth movement is called can be realized, and the performance of the virtual character is more natural, careful and flexible. The specific implementation mode can include that when the voice information is acquired, the time stamp in the voice information is acquired at the same time, and when the performance control information is generated, the time stamp is synchronized on the corresponding performance control information so as to realize the synchronization of the time stamp; it is also possible to synchronize the time stamps of the speech information and the show control information when the show control information is transmitted. In another embodiment, a set of protocols can be designed to uniformly transmit the voice information and the performance control information in sequence in the same connection, so that the voice information and the performance control information are naturally synchronized. Specifically, the sequence of the actions and sounds of the character is determined according to the sequence of the timestamps of the voice information and the performance control information, so that perfect synchronization of the sound and the painting of the performance is achieved. Besides the time stamp, due to the ingenious design of the communication protocol, the voice information and the performance control information can be sent through the same communication channel in the transmission process, and the synchronization of the voice information and the performance control information is also ensured.
In another embodiment, the show control information may be obtained by performing matching according to the show command. As described above, the performance control information includes a plurality of categories, and the categories are respectively awakened and called by corresponding performance instructions to control virtual character performance, setting of scenes, or playing of sound effects.
In one embodiment, at step S2: after the step of obtaining show control information according to the show trigger information, the method includes: controlling the virtual character to realize performance according to the performance control information; and displaying the performance of the virtual character in real time.
In an embodiment, after the performance control information is generated, the anchor terminal may control and implement the performance of the virtual character according to the performance control information, and display the performance in real time in the display device, so as to implement the monitoring and previewing of the anchor terminal, so as to implement monitoring before outputting to the client terminal, so as to be able to process in time before a live broadcast accident occurs, or to modify and adjust the performance with poor effect at any time.
In an embodiment, the virtual character rendering method further includes: when the number of the virtual roles is more than or equal to two, acquiring target role information; at step S2: the step of obtaining the show control information according to the show trigger information includes: and acquiring the performance control information of the target role information according to the performance trigger information.
In an embodiment, it can be understood that when a plurality of virtual characters appear in one live broadcast room, corresponding control is required, so that target character information can be obtained, and performance control information of the target character information is obtained according to performance trigger information, so as to realize one-to-one correspondence between the performance control information and the target characters, prevent problems of performance confusion, control errors and the like, and support the plurality of virtual characters to perform simultaneously.
Step S3: and outputting the performance control information to the client so that the client controls the virtual character to perform the performance according to the performance control information.
In one embodiment, the anchor terminal is connected with the client terminal through a communication technology, and after the anchor terminal acquires the performance control information, the performance control information can be processed through technologies such as coding compression, fault tolerance and the like and then sent to the user client terminal through a real-time audio transmission network; and the client analyzes the data through the unpacking protocol after receiving the data. It can be understood that, in this embodiment, the data stream including the performance control information is sent by the anchor to the client instead of the video stream, and the transmitted data stream includes not only the data stream including the performance control information but also the voice information associated with the performance control information, so that the client can not only perform the virtual character but also play audio synchronized with the virtual character when parsing. That is, it is preferable that the anchor terminal transmits a data stream including the performance control information to the client terminal, and the client terminal performs self-rendering using a local resource according to the received performance control information to render the performance of the virtual character. In another embodiment, the host may output the data of the video stream to the client after rendering the performance on the display device thereof, for example, the data of the video stream may be output to a third party, and the data and the path of outputting the data to the client may be synchronized and parallel, so as to realize a wider spread of the performance of the virtual character and a performance of a wide spread virtual character. Meanwhile, in the process of outputting the audio to the client, in the implementation mode that the anchor side comprises the anchor console and the server, the server can render the audio in real time, including processing of adding background sound, adding sound effect, changing sound and the like, so that the processing mode can be conveniently and timely modified, and performance pressure brought by the processing of the audio by the anchor console and delay caused by the performance pressure can be relieved. Because only the data stream of the performance control information is output, the pressure of the data stream output by the main broadcasting end can be greatly reduced, the network load is reduced, the network cost required by the client for receiving the high-definition video is reduced, and the like, and the network requirement threshold for watching the performance by the user is greatly reduced.
In an embodiment, the virtual character performing method provided in this embodiment further includes the following steps: acquiring interaction request information sent by a client; responding to the interaction request information, and generating interaction feedback information; and outputting the interactive feedback information to the client.
In one embodiment, to increase the interactivity of the performance between the user and the main broadcast, some interactive request buttons are often provided, such as connecting to a wheat, a mini game, or a name of a question in a live classroom. In this embodiment, it is necessary to obtain the interaction request information sent by the client, where the client is a broader client, and includes an intelligent terminal of a general audience watching the performance, and may also include other anchor terminals. Thus, not only the interaction between the anchor and the audience, but also the interaction between the anchor and the anchor, such as connecting to the wheat, PK, mini-games, etc., can be realized. And then generating interaction feedback information according to the interaction request information response, and then outputting the interaction feedback information to the client. That is, the request for interaction is satisfied, and the corresponding response is executed according to the request for interaction. It will be appreciated that the interaction received by the anchor is varied and therefore the corresponding response is varied and is described herein for purposes of illustration and not limitation. Specifically, the anchor terminal randomly selects a client terminal to perform the microphone connection response, and then selects a microphone connection interaction request message from the plurality of client terminals to perform the response, thereby realizing the response process between the anchor terminal and the client terminal. And correspondingly generating interactive feedback information, such as communication between the two, and finally feeding the interactive feedback information back to all the clients, so that all the clients can confirm the interactive response. The requirements that the user interacts with the virtual role, the function and the content in real time and interactive data is synchronized to the anchor terminal and other user client terminals through a real-time audio transmission network are met. It should be understood that, in the client that sends the last interactive feedback information, it is preferable that all the clients connected to the anchor side, that is, all the clients are able to receive the interactive feedback information generated by this interaction.
The virtual character performing method provided by the first embodiment of the present invention is applied to a main broadcasting end, and includes the following steps: step S1: acquiring performance triggering information, wherein the performance triggering information comprises voice information and/or performance instructions; step S2: acquiring performance control information according to the performance trigger information; step S3: and outputting the performance control information to the client so that the client controls the virtual character to perform the performance according to the performance control information. Therefore, the invention triggers the performance through the instruction and voice input mode, reduces the cost of virtual character performance, particularly for live broadcasting, no special acquisition equipment is needed, and the performance effect can be obviously improved, and the action and expression can be more flexible, magical, cool, exaggerated, delicate and smooth. Meanwhile, for live broadcasting, the network pressure output by a live broadcasting end is reduced by outputting a data stream instead of a video stream, the network cost and consumption of data received by a client are reduced, and the virtual character performance form can be popularized and applied in a larger range, so that considerable social and economic benefits are brought. In addition, the virtual character performance method provided by the embodiment of the invention can also acquire performance instructions in various ways, and can meet the manual performance requirement of the anchor end or automatically acquire preset performance instructions from the database through the anchor self-defining performance actions, so that natural actions or linkage actions are added, and the virtual character performance is smoother. Meanwhile, when the voice input triggers the performance, the voice information and the performance control information are bound in a timestamp mode, so that the virtual character can play the upper mouth shape more naturally. Meanwhile, the performance of the virtual character is monitored in real time, so that timely processing can be carried out before the virtual character is transmitted when an accident happens. In addition, the simultaneous and on-screen performance of one or more virtual characters is supported, so that the performance is richer. And the interaction request information sent by the client side can be received to correspond to the response, so that the interaction between the anchor and the audience is realized, various requirements are met, the interestingness of performance is increased, the operation of a user is reduced, the convenience of the user is increased, and the use experience of the user is improved.
Second embodiment
Fig. 2 is a schematic diagram of a first structure of a anchor according to a second embodiment of the present invention. For a clear description of the anchor 110 provided in the second embodiment of the present invention, please refer to fig. 1 and fig. 2.
The anchor terminal 110 according to the second embodiment of the present invention includes: a processor a101 and a memory a201, wherein the processor a101 is configured to execute the computer program a6 stored in the memory a201 to implement the steps of the virtual character rendering method as described in the first embodiment.
In an embodiment, the anchor terminal 110 provided in this embodiment includes at least one processor a101 and at least one memory a 201. Wherein, at least one processor A101 may be referred to as a processing unit A1, and at least one memory A201 may be referred to as a memory unit A2. Specifically, the storage unit a2 stores the computer program a6, and when the computer program a6 is executed by the processing unit a1, the anchor terminal 110 provided in the present embodiment implements the steps of the virtual character rendering method described in the first embodiment. For example, step S1 shown in fig. 1: acquiring performance triggering information, wherein the performance triggering information comprises voice information and/or performance instructions; step S2: acquiring performance control information according to the performance trigger information; step S3: and outputting the performance control information to the client so that the client controls the virtual character to perform the performance according to the performance control information.
In one embodiment, the anchor 110 provided in the present embodiment may include a plurality of memories a201 (simply referred to as storage unit A2). Storage unit a2 may be either volatile memory or nonvolatile memory, or may include both volatile and nonvolatile memory. Among them, the nonvolatile Memory may be a Read Only Memory (ROM), a Programmable Read Only Memory (PROM), an Erasable Programmable Read-Only Memory (EPROM), an Electrically Erasable Programmable Read-Only Memory (EEPROM), a magnetic random access Memory (FRAM), a Flash Memory (Flash Memory), a magnetic surface Memory, an optical disk, or a Compact Disc Read-Only Memory (CD-ROM); the magnetic surface storage may be disk storage or tape storage. Volatile Memory can be Random Access Memory (RAM), which acts as external cache Memory. By way of illustration and not limitation, many forms of RAM are available, such as Static Random Access Memory (SRAM), Synchronous Static Random Access Memory (SSRAM), Dynamic Random Access Memory (DRAM), Synchronous Dynamic Random Access Memory (SDRAM), Double Data Rate Synchronous Dynamic Random Access Memory (DDRSDRAM), Enhanced Synchronous Dynamic Random Access Memory (ESDRAM), Enhanced Synchronous Dynamic Random Access Memory (Enhanced DRAM), Synchronous Dynamic Random Access Memory (SLDRAM), Direct Memory (DRmb Access), and Random Access Memory (DRAM). The memory unit a2 described in the embodiments of the present invention is intended to comprise, without being limited to, these and any other suitable types of memory.
In one embodiment, the anchor 110 also includes a bus connecting the various components (e.g., processor a101 and memory a201, display a3, etc.). After the anchor terminal 110 obtains the performance control information, the display device A3 can first render the performance of the virtual character according to the local resource of the anchor terminal 110, and display the performance through the display device A3, so as to monitor the live broadcast effect.
In one embodiment, the anchor 110 in this embodiment may further include a communication interface (e.g., I/O interface a4) that may be used to communicate with external devices.
In an embodiment, the anchor terminal 110 provided in this embodiment may further include a communication device a 5.
In one embodiment, the anchor 110 is communicatively connected to the client, for example, through a wired or wireless communication method, and therefore needs the participation of the I/O interface a4 or the communication device a 5.
The anchor terminal 110 provided by the second embodiment of the present invention includes a memory a101 and a processor a201, and the processor a101 is configured to execute the computer program a6 stored in the memory a201 to implement the steps of the virtual character performing method described in the first embodiment, so that the anchor terminal 110 provided by this embodiment triggers performing through instructions and voice input, reduces the cost of virtual character performing, in particular for live broadcasting, does not need special acquisition equipment, and can significantly improve the performing effect, and actions and expressions can be more flexible, curious, cool, exaggerated, detailed, and smooth. Meanwhile, for live broadcasting, the network pressure output by a live broadcasting end is reduced by outputting a data stream instead of a video stream, the network cost and consumption of data received by a client are reduced, and the virtual character performance form can be popularized and applied in a larger range, so that considerable social and economic benefits are brought. In addition, the virtual character performance method provided by the embodiment of the invention can also acquire performance instructions in various ways, and can meet the manual performance requirement of the anchor end or automatically acquire preset performance instructions from the database through the anchor self-defining performance actions, so that natural actions or linkage actions are added, and the virtual character performance is smoother. Meanwhile, when the voice input triggers the performance, the voice information and the performance control information are bound in a timestamp mode, so that the virtual character can play the upper mouth shape more naturally. Meanwhile, the performance of the virtual character is monitored in real time, so that timely processing can be carried out before the virtual character is transmitted when an accident happens. In addition, the simultaneous and on-screen performance of one or more virtual characters is supported, so that the performance is richer. And the interaction request information sent by the client side can be received to correspond to the response, so that the interaction between the anchor and the audience is realized, various requirements are met, the interestingness of performance is increased, the operation of a user is reduced, the convenience of the user is increased, and the use experience of the user is improved.
Fig. 3 is a second structural diagram of the anchor terminal 110 according to a second embodiment of the present invention. For a clear description of the anchor 110 provided in the second embodiment of the present invention, please refer to fig. 1 and fig. 3.
Referring to fig. 3, in another embodiment, the anchor 110 according to a third embodiment of the present invention includes: a central controller 801 (CPU for short), a read only memory 802 (ROM for short), a random access memory 803 (RAM for short), a communication bus 804, an I/O interface 805, an input unit 806, an output unit 807, a storage unit 808, and a communication unit 809.
In other embodiments, the central controller 801 (CPU), the read only memory 802 (ROM), and the random access memory 803 (RAM) may constitute a processor.
In other embodiments, the anchor terminal 110 can implement the virtual character rendering method as provided in the first embodiment, and the technical effects that can be achieved are described in detail above and will not be expanded herein.
The second embodiment of the present invention also provides a computer-readable storage medium storing a computer program a6, which when executed by a processor a101, implements the steps of the virtual character rendering method as described in the first embodiment.
In an embodiment, the computer readable storage medium provided by the embodiment may include any entity or device capable of carrying computer program code, a recording medium, such as ROM, RAM, magnetic disk, optical disk, flash memory, and the like.
The technical effects that can be achieved by the computer program a6 stored in the computer-readable storage medium provided by the second embodiment of the present invention when the computer program a6 is executed by the processor a101 have been described in detail above, and will not be expanded herein.
Third embodiment
Fig. 4 is a schematic structural diagram of a virtual character performance system according to a third embodiment of the present invention. For a clear description of the virtual character performing system 100 according to the third embodiment of the present invention, please refer to fig. 1 to 4.
The virtual character performing system 100 according to the third embodiment of the present invention includes an anchor terminal 110 and a client terminal 120.
In one embodiment, the anchor terminal 110 is configured to obtain the performance trigger information, where the performance trigger information includes voice information and/or performance instructions, obtain the performance control information according to the performance trigger information, and output the performance control information to the client terminal 120.
In an embodiment, a specific structure of the anchor terminal 110 has been described in detail in the anchor terminal 110 provided in the second embodiment of the present invention, and for specific reference, it is not repeated herein.
In one embodiment, the client 120 is configured to receive show control information transmitted by the anchor 110.
In one embodiment, the client 120, after the step of receiving the show control information transmitted by the anchor 110, includes: and performing the virtual character by using the rendering of the local resources through the performance control information.
In one embodiment, the client 120 may include a structure similar to the host 110, such as a PC or a mobile terminal. Audio and instruction data can be respectively obtained through an unpacking protocol, and according to the audio data, sound is played, the instruction data is analyzed to control the action and expression of the virtual character, set a scene, play a dynamic effect sound effect and issue an interaction control for user interaction operation; 1 or more 2d or 3d virtual characters are supported to perform the same-screen performance at the same time.
In one embodiment, anchor 110 and client 120 are communicatively coupled via wired and/or wireless communication techniques. Among them, for Wireless Communication technologies, including but not limited to Global System for Mobile Communication (GSM), Enhanced Data GSM Environment (EDGE), wideband Code division multiple Access (W-CDMA), Code Division Multiple Access (CDMA), Time Division Multiple Access (TDMA), bluetooth, Wireless Fidelity (WiFi) (e.g., IEEE802.11 a, IEEE802.11 b, IEEE802.1 g, and/or IEEE802.1 n), Voice over internet protocol (VoIP), Worldwide Interoperability for Microwave Access (Microwave-assisted, Max), and other short message Communication protocols, as well as any other suitable communication protocols, and may even include those that have not yet been developed.
In another embodiment, the number of clients 120 is massive for anchor 110, and therefore a server may be included to act as an intermediary between anchor 110 and clients 120. That is, the data stream of the performance control information output by the anchor terminal 110 is encoded, compressed, and subjected to fault-tolerant processing, and then is sent to the server through the real-time audio transmission network, and the server forwards the data stream to the user client terminal 120; the client 120 receives the data and then parses the data via a depacketization protocol. It is also possible that the host 110 includes a console and a server, wherein the client 120 is directly connected to the server and the server is connected to the console. As in the preferred embodiment described in the first embodiment of the present invention: the console acquires the performance trigger information, and the console and/or server acquires the performance control information, and the server sends the performance control information to the mass clients 120.
In an embodiment, the anchor terminal 110 and the client terminal 120 may further implement an interaction function by obtaining interaction request information and interaction feedback information, so as to satisfy interaction between a user and the anchor terminal.
The virtual character performing system 100 according to the third embodiment of the present invention includes an anchor terminal 110 and a client terminal 120. The anchor terminal 110 is configured to acquire performance trigger information, where the performance trigger information includes voice information and/or performance instructions, acquire performance control information according to the performance trigger information, and output the performance control information to the client terminal 120; the client 120 is configured to receive the show control information transmitted by the anchor 110. Therefore, the technical effects that can be specifically achieved by the virtual character performance method provided by the first embodiment of the present invention have been described in detail in the foregoing, and are not repeated herein.
The technical features of the embodiments described above may be arbitrarily combined, and for the sake of brevity, all possible combinations of the technical features in the embodiments described above are not described, but should be considered as being within the scope of the present specification as long as there is no contradiction between the combinations of the technical features.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, the recitation of an element by the phrase "comprising an … …" does not exclude the presence of additional like elements in the process, method, article, or apparatus that comprises the element, and further, where similarly-named elements, features, or elements in different embodiments of the disclosure may have the same meaning, or may have different meanings, that particular meaning should be determined by their interpretation in the embodiment or further by context with the embodiment. As used herein, the meaning of "a plurality" or "a plurality" is two or more unless otherwise specified.
It should be understood that, although the steps in the flowcharts in the embodiments of the present application are shown in order as indicated by the arrows, the steps are not necessarily performed in order as indicated by the arrows. The steps are not performed in the exact order shown and may be performed in other orders unless explicitly stated herein. Moreover, at least some of the steps in the figures may include multiple sub-steps or multiple stages that are not necessarily performed at the same time, but may be performed at different times, in different orders, and may be performed alternately or at least partially with respect to other steps or sub-steps of other steps.
It will be understood by those skilled in the art that all or part of the steps of implementing the above method embodiments may be implemented by hardware associated with program instructions, and the program may be stored in a computer readable storage medium, and when executed, performs the steps including the above method embodiments. The foregoing storage medium includes: various media that can store program codes, such as ROM, RAM, magnetic or optical disks.
The above description is only for the specific embodiments of the present invention, but the scope of the present invention is not limited thereto, and any person skilled in the art can easily conceive of the changes or substitutions within the technical scope of the present invention, and all the changes or substitutions should be covered within the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the appended claims.

Claims (10)

1. A virtual character performance method is applied to a main broadcasting end and is characterized by comprising the following steps:
acquiring performance triggering information, wherein the performance triggering information comprises voice information and/or performance instructions;
acquiring performance control information according to the performance trigger information;
and outputting the performance control information to a client so that the client controls the virtual character to perform performance according to the performance control information.
2. The virtual character rendering method of claim 1, wherein the step of obtaining rendering trigger information comprises:
responding to user operation, matching the corresponding performance instruction, wherein the performance instruction comprises one of a single instruction and/or a combined instruction; and/or the presence of a gas in the gas,
and automatically acquiring the preset performance instruction from a natural performance database.
3. The virtual character rendering method of claim 1, wherein the step of obtaining rendering control information based on the rendering trigger information comprises:
generating the performance control information according to the voice information; and/or the presence of a gas in the gas,
and matching the corresponding performance control information according to the performance instruction.
4. The virtual character rendering method of claim 1, wherein the step of obtaining rendering control information based on the rendering trigger information comprises:
acquiring a time stamp of the performance control information according to the time stamp in the voice information; and/or the presence of a gas in the gas,
and setting a timestamp of the performance control information through a preset protocol.
5. The virtual character rendering method of claim 1, after the step of obtaining rendering control information based on the rendering trigger information, comprising:
controlling the virtual character to realize performance according to the performance control information;
and displaying the performance of the virtual character in real time.
6. The virtual character rendering method of claim 1, the method further comprising: when the number of the virtual roles is more than or equal to two, acquiring target role information;
the step of acquiring the performance control information according to the performance trigger information includes:
and acquiring the performance control information of the target role information according to the performance trigger information.
7. The virtual character rendering method of claim 1, further comprising the steps of:
acquiring interaction request information sent by the client;
responding to the interaction request information, and generating interaction feedback information;
and outputting the interactive feedback information to the client.
8. An anchor, comprising a processor and a memory:
the processor is configured to execute a computer program stored in the memory to implement the virtual character rendering method steps of any of claims 1 to 7.
9. A virtual character performance system is characterized by comprising an anchor terminal and a client terminal;
the main broadcasting terminal is used for acquiring performance triggering information, the performance triggering information comprises voice information and/or performance instructions, performance control information is acquired according to the performance triggering information, and the performance control information is output to the client;
and the client is used for receiving the performance control information transmitted by the main broadcasting terminal and controlling the virtual character to perform performance according to the performance control information.
10. A computer-readable storage medium, wherein a computer program is stored, which when executed by a processor, performs the steps of the virtual character rendering method of any one of claims 1-7.
CN202111132440.9A 2021-09-18 2021-09-18 Virtual character performance method, device, system and computer readable storage medium Pending CN114079800A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111132440.9A CN114079800A (en) 2021-09-18 2021-09-18 Virtual character performance method, device, system and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111132440.9A CN114079800A (en) 2021-09-18 2021-09-18 Virtual character performance method, device, system and computer readable storage medium

Publications (1)

Publication Number Publication Date
CN114079800A true CN114079800A (en) 2022-02-22

Family

ID=80283484

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111132440.9A Pending CN114079800A (en) 2021-09-18 2021-09-18 Virtual character performance method, device, system and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN114079800A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116389777A (en) * 2023-03-10 2023-07-04 启朔(深圳)科技有限公司 Cloud digital person live broadcasting method, cloud device, anchor terminal device and system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104867176A (en) * 2015-05-05 2015-08-26 中国科学院自动化研究所 Cryengine-based interactive virtual deduction system
CN110308792A (en) * 2019-07-01 2019-10-08 北京百度网讯科技有限公司 Control method, device, equipment and the readable storage medium storing program for executing of virtual role
CN111312240A (en) * 2020-02-10 2020-06-19 北京达佳互联信息技术有限公司 Data control method and device, electronic equipment and storage medium
CN113325951A (en) * 2021-05-27 2021-08-31 百度在线网络技术(北京)有限公司 Operation control method, device, equipment and storage medium based on virtual role
CN113325952A (en) * 2021-05-27 2021-08-31 百度在线网络技术(北京)有限公司 Method, apparatus, device, medium and product for presenting virtual objects

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104867176A (en) * 2015-05-05 2015-08-26 中国科学院自动化研究所 Cryengine-based interactive virtual deduction system
CN110308792A (en) * 2019-07-01 2019-10-08 北京百度网讯科技有限公司 Control method, device, equipment and the readable storage medium storing program for executing of virtual role
CN111312240A (en) * 2020-02-10 2020-06-19 北京达佳互联信息技术有限公司 Data control method and device, electronic equipment and storage medium
CN113325951A (en) * 2021-05-27 2021-08-31 百度在线网络技术(北京)有限公司 Operation control method, device, equipment and storage medium based on virtual role
CN113325952A (en) * 2021-05-27 2021-08-31 百度在线网络技术(北京)有限公司 Method, apparatus, device, medium and product for presenting virtual objects

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116389777A (en) * 2023-03-10 2023-07-04 启朔(深圳)科技有限公司 Cloud digital person live broadcasting method, cloud device, anchor terminal device and system

Similar Documents

Publication Publication Date Title
US10210002B2 (en) Method and apparatus of processing expression information in instant communication
CN106237614B (en) Capturing asynchronous commentary on prerecorded gameplay
TWI554317B (en) System and method for managing audio and video channels for video game players and spectators
US9832516B2 (en) Systems and methods for multiple device interaction with selectably presentable media streams
CN105450642B (en) It is a kind of based on the data processing method being broadcast live online, relevant apparatus and system
JP5603417B2 (en) Shared media selection method and system integrated with avatar
KR101571283B1 (en) Media content transmission method and apparatus, and reception method and apparatus for providing augmenting media content using graphic object
TW201445414A (en) Method, user terminal and server for information exchange in communications
CN110178158B (en) Information processing apparatus, information processing method, and recording medium
JP2020039029A (en) Video distribution system, video distribution method, and video distribution program
US11451858B2 (en) Method and system of processing information flow and method of displaying comment information
CN108322474B (en) Virtual reality system based on shared desktop, related device and method
US20220068010A1 (en) Augmented reality methods and systems
CN113301358A (en) Content providing and displaying method and device, electronic equipment and storage medium
CN115631270A (en) Live broadcast method and device of virtual role, computer storage medium and terminal
CN114079800A (en) Virtual character performance method, device, system and computer readable storage medium
US20220214749A1 (en) Real-time immersive content providing system, and haptic effect transmission method thereof
JP2020102782A (en) Content distribution system, distribution device, reception device, and program
KR102200239B1 (en) Real-time computer graphics video broadcasting service system
WO2012166072A1 (en) Apparatus, systems and methods for enhanced viewing experience using an avatar
CN110602523A (en) VR panoramic live multimedia processing and synthesizing system and method
US20150371661A1 (en) Conveying Audio Messages to Mobile Display Devices
CN108271057A (en) Video interaction method, subscription client, server and readable storage medium storing program for executing
US20190208288A1 (en) Method and system for delivering an interactive video
CN113409431B (en) Content generation method and device based on movement data redirection and computer equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination