WO2023236045A1 - 实现虚拟人物和真人视频混合连麦的系统和方法 - Google Patents

实现虚拟人物和真人视频混合连麦的系统和方法 Download PDF

Info

Publication number
WO2023236045A1
WO2023236045A1 PCT/CN2022/097401 CN2022097401W WO2023236045A1 WO 2023236045 A1 WO2023236045 A1 WO 2023236045A1 CN 2022097401 W CN2022097401 W CN 2022097401W WO 2023236045 A1 WO2023236045 A1 WO 2023236045A1
Authority
WO
WIPO (PCT)
Prior art keywords
anchor
virtual
terminal
avatar
video
Prior art date
Application number
PCT/CN2022/097401
Other languages
English (en)
French (fr)
Inventor
段新盟
陈帅雷
刘宁
Original Assignee
云智联网络科技(北京)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 云智联网络科技(北京)有限公司 filed Critical 云智联网络科技(北京)有限公司
Priority to PCT/CN2022/097401 priority Critical patent/WO2023236045A1/zh
Publication of WO2023236045A1 publication Critical patent/WO2023236045A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting

Definitions

  • This application relates to the field of live broadcasting, and in particular to a system and method for realizing mixed live streaming of virtual characters and real-life videos.
  • the interaction process between live broadcast anchors requires the anchors to turn on their respective cameras, and the cameras capture the anchor's live video and send the live video to the terminals of each audience in the live broadcast room. While watching the host's live broadcast or interacting with the host, the audience can leave messages, tips, or send atmosphere or effect content to the host, which will be presented during the host's live broadcast.
  • this application provides a solution to realize the mixed continuous broadcast of virtual characters and real people's videos.
  • the interactive process of the anchor's continuous broadcast it is no longer limited to the form of interaction between real people, but between the real video anchor and the virtual character anchor. The interaction between them enriches the presentation form of live broadcast interaction and brings a better viewing experience to the audience.
  • a method for implementing mixed broadcasting of virtual characters and real-life videos is provided, which is applied to the server side and includes:
  • the avatar image instruction of the avatar anchor terminal is sent to the real person video anchor terminal and the audience terminal in the real person video anchor's live broadcast room, and the The video stream of the real person video anchor terminal is sent to the virtual character anchor terminal and the audience terminal in the virtual character anchor live broadcast room, wherein the virtual character image instruction is used to call the virtual character anchor terminal, the real person video anchor terminal and Elements in the set of preset virtual elements in the audience terminal.
  • a method for realizing mixed broadcast of virtual characters and real-person videos is provided, which is applied to the virtual character anchor terminal, including:
  • the live image of the live video anchor is displayed.
  • a method for realizing mixed continuous broadcast of virtual characters and real-person videos is provided, which is applied to real-person video anchor terminals, including:
  • the virtual elements in the preset virtual element set are called to display the virtual character image corresponding to the host.
  • a method for realizing mixed broadcast of virtual characters and real-life videos is provided, which is applied to audience terminals, including:
  • the virtual elements in the preset virtual element set are called to display the virtual character image of the virtual character anchor, and the Lianmai image of the real video anchor is displayed according to the live video stream.
  • a system to realize mixed broadcasting of virtual characters and real-person videos, which includes a server, an avatar anchor terminal, a real-person video anchor terminal, and a viewer terminal, wherein the server executes the following: The method described in the first aspect, the virtual character anchor terminal performs the method described in the second aspect, the real person video anchor terminal performs the method described in the third aspect, and the audience terminal performs the method described in the third aspect. methods described in four aspects.
  • avatar elements related to the host's avatar are preset on the host's and audience's terminals respectively, and the avatar elements have corresponding avatar element instructions.
  • the anchor terminal or the audience terminal sends virtual element instructions to other terminals through the server.
  • the anchor terminal and the audience terminal call the preset avatar elements according to the avatar element instructions to present virtual effects.
  • Figure 1 is a schematic diagram of a system for implementing a hybrid video link between virtual characters and real people according to an embodiment of the present application.
  • Figure 2 is a flowchart of a method executed by the server to implement a hybrid video link between a virtual character and a real person according to an embodiment of the present application.
  • Figure 3(a) is a flow chart of a method executed by an avatar anchor terminal to implement a hybrid video connection between an avatar and a real person according to an embodiment of the present application.
  • Figure 3(b) is a flowchart of a method executed by an avatar anchor terminal to implement a hybrid video link between an avatar and a real person according to another embodiment of the present application.
  • Figure 4(a) is a flow chart of a method executed by a real-life video anchor terminal to implement a hybrid connection between virtual characters and real-life videos according to an embodiment of the present application.
  • Figure 4(b) is a flow chart of a method executed by a real-life video anchor terminal to implement a hybrid connection between virtual characters and real-life videos according to another embodiment of the present application.
  • Figure 5 is a flowchart of a method executed by a viewer terminal to implement a hybrid video link between a virtual character and a real person according to an embodiment of the present application.
  • Figure 1 is a schematic diagram of a system for implementing a hybrid video link between virtual characters and real people according to an embodiment of the present application.
  • the system that realizes the mixed broadcast of virtual characters and real-person videos includes a server, a real-person video anchor terminal, an avatar anchor terminal, and an audience terminal.
  • the real-person video anchor terminal, the avatar anchor terminal, and the audience terminal are respectively connected with Server side connection.
  • mixed video of virtual characters and real people refers to the existence of real and virtual characters in the live broadcast process.
  • a set of virtual elements related to the live broadcast and interaction of virtual characters is stored.
  • This set of virtual elements includes the set of virtual elements related to the live broadcast and interaction of virtual characters.
  • Various virtual elements such as character settings related to virtual characters (cartoon characters, animals, etc.), character's body parts (facial features, hair, limbs, etc.), character's clothing (hats, clothes, gloves, etc.), virtual characters
  • the server side, anchor terminal and audience terminal can all issue corresponding instructions to call virtual elements in the virtual element set to present corresponding virtual effects.
  • the avatar anchor In order to implement avatar live broadcast, the avatar anchor needs to preset a set of virtual elements (for example, downloaded from the server) before starting the broadcast, and needs to set up his own avatar.
  • the virtual character anchor can set his own eyes, ears, and his own clothing.
  • the avatar anchor personalizes his own avatar, when the broadcast starts, he will form an avatar image command corresponding to the setting of the avatar image, and send the avatar image command to the server, and the server will send the avatar image to the server.
  • Image instructions are sent to the audience terminals in the host's live broadcast room.
  • the audience terminal In order to present a virtual character live broadcast, the audience terminal needs to preset a set of virtual elements (for example, downloaded from the server) at the same time or before entering the live broadcast room. After entering the live broadcast room, the audience terminal can receive the virtual character image command sent by the host terminal from the server, and call the preset virtual element set according to the virtual character image command to present the effect of the virtual character live broadcast.
  • a set of virtual elements for example, downloaded from the server
  • the virtual character image instructions include static element instructions and dynamic element instructions.
  • the static element instructions include the virtual image presented in the live broadcast, for example, the static state presented by the anchor (what the body parts look like, what clothes and hats are worn) ), the background of the avatar and the atmosphere or rewards sent by the anchor, etc.; while the avatar anchor will have various expressions and actions during the live broadcast, such as blinking, dancing, smiling, etc., the avatar anchor will The terminal (for example, through a camera) will capture these actions, and the resulting instructions are dynamic element instructions.
  • the dynamic element instructions are used to instruct the host's avatar to perform various actions.
  • the anchor's terminal captures the blinking action, forms an instruction instructing the anchor's avatar to blink, and sends the instruction to the server.
  • the server then sends the instruction to the audience terminal, and the audience terminal responds to the instruction.
  • the first anchor (which can be a real video anchor or a virtual character anchor) sends an interaction request to the server through its terminal.
  • the interaction request can specify the object to interact with, such as a second anchor; the interaction request can also not specify the object to interact with, but instructs the server to specify an interaction object or randomly match an interaction object, such as randomly matching a second anchor.
  • the method by which the first anchor obtains the interactive anchor object is not limited by this application.
  • the anchor who initiates the continuous broadcast request can start broadcasting or not.
  • the anchor who initiates the continuous broadcast request sends the continuous broadcast request to the server; when the broadcast is not started, the anchor who initiates the continuous broadcast request first enters the live broadcast room of the target anchor as a viewer, and then passes the server The terminal sends a microphone connection request to the target anchor terminal.
  • the server After determining the interactive object of the first anchor (the second anchor), the server sends the interaction request to the second anchor terminal. After receiving the request, the second anchor terminal presents the interaction request information to the second anchor. If the second anchor does not agree to interact, the second anchor terminal sends a rejection request message to the server, and specifies the interaction object after the first anchor In the mode, the server side feeds back the message that the interaction request is rejected to the first anchor terminal; in the mode of randomly matching interactive objects, the server side continues to match the next interactive object until it matches an object that accepts the interaction request or receives all pending interaction objects. The rejection message for the matching object. If the second anchor agrees to interact, the second anchor terminal sends an interaction confirmation message to the server.
  • the server After receiving the interaction confirmation from the second anchor, the server sends the avatar image instructions of the avatar anchor terminal to the real video anchor terminal and the audience terminals in the real video anchor live broadcast room, and sends the video stream of the real video anchor terminal to The avatar anchor terminal and the audience terminal in the avatar anchor live broadcast room, where the avatar image instruction is used to call the elements in the virtual character anchor terminal, the live video anchor terminal and the audience terminal that are preset in the set of virtual elements.
  • the real-person video anchor terminal After receiving the avatar image instruction sent by the avatar host terminal, the real-person video anchor terminal calls the virtual elements in the preset virtual element set to display the avatar image of the avatar anchor. After receiving the video stream sent by the real video anchor terminal, the virtual character anchor displays the image of the real video anchor based on the video stream. After receiving the avatar image command sent by the avatar anchor terminal, the audience terminal calls the virtual elements in the preset virtual element set to display the avatar image of the avatar anchor and the interaction process between the anchors, such as displaying the virtual character between anchors. Animations such as head touching and virtual tapping. After receiving the video stream sent by the live video host terminal, the audience terminal displays the live video host's continuous wheat image according to the video stream.
  • the real person video anchor can see the avatar image of the avatar anchor
  • the avatar anchor can see the real person image of the real video anchor
  • the audience can see the avatar image of the avatar anchor and the real person image of the real video anchor.
  • Mai image During the live interaction process, the anchor terminal and the audience terminal can display the anchor screen in various forms.
  • the audience will see two forms, one is the form of the avatar anchor, and the other is the form of the real video anchor.
  • Audiences can participate in the process of watching the host's interaction, for example, by enhancing the atmosphere, giving rewards, etc. For example, viewers can tip their favorite anchors and choose the reward effect.
  • the interactive effects available for the interaction between the audience and the live video anchor and the avatar anchor can be different.
  • virtual character anchors can be given unique virtual gifts and virtual special effects
  • real video anchors can be given ordinary continuous gifts and special effects gifts.
  • the interaction forms for the audience to choose are also different depending on the interaction objects.
  • the interaction objects are the real video anchor and the virtual character anchor, and the content of the interactive effect selection list presented to the audience by the audience terminal may be different.
  • the audience terminal sends a common effect instruction representing the selected interaction effect (such as a reward effect) to the server side, and the server side sends the instruction to the anchor terminal and other audience terminals.
  • the anchor terminal After receiving the instruction, other audience terminals will display interactive effects according to the instruction.
  • the audience terminal sends a virtual special effects instruction representing the selected interaction effect (such as a reward effect) to the server side, and the server side sends the instruction to the anchor terminal and other audience terminals, and the anchor terminal and other After receiving the instruction, the audience terminal calls the virtual elements in the preset virtual element set according to the instruction to display the interactive effect.
  • the interactive effects available for the audience to interact with the real video anchor and the avatar anchor are not restricted by whether the anchor is a real video anchor or an avatar anchor. That is, ordinary effect instructions can be sent to the avatar anchor, and the avatar anchor can The terminal presents normal effects; virtual special effects can be sent to live video anchors, such as special effects based on face recognition, and the live video anchor terminal presents virtual effects.
  • real-person video anchors and virtual character anchors can interact.
  • a live video anchor takes a picture of a virtual character anchor, etc.
  • the real-person video anchor interacts with the virtual character anchor.
  • the real-person video anchor selects the content and effect of the interaction, generates virtual interaction instructions and sends them to the server.
  • the server sends the virtual interaction instructions to the virtual character anchor terminal and the audience. terminal.
  • the avatar anchor terminal and the audience terminal call the virtual elements in the preset set of virtual elements according to the virtual interaction instructions to display the interaction effect between the real video anchor and the avatar anchor.
  • the mixed continuous broadcasting of two anchors is used for description.
  • the number of anchors in the mixed broadcast solution of the present application can be arbitrary, and as long as there are real video anchors and virtual character anchors, the number of real video anchors and virtual character anchors No restrictions.
  • both anchors of the video streaming must have the same anchor character presentation form, that is, there can be a mixed form of real video anchors and virtual character anchors, which makes the format of video streaming more flexible and provides greater convenience to the audience.
  • the effect of the instructions sent by the host terminal, the audience terminal and the server to call the virtual element set is more consistent with the virtual character compared to the live broadcast.
  • the audience rewards the host The virtual effect of a hat. This hat can be worn just on the head of the host's avatar. However, if the host is a real person, the matching effect is not very good. In this way, using the hybrid interactive method of this application, it is richer than real-person live broadcast in terms of vision, interaction, playability and other dimensions.
  • this application also provides a method for realizing mixed continuous broadcasting of virtual characters and real-person videos, as follows: As shown in Figure 2, Figure 3(a), Figure 3(b), Figure 4(a), Figure 4(b) and Figure 5.
  • FIG. 2 is a flowchart of a method executed by the server to implement a hybrid video link between a virtual character and a real person according to an embodiment of the present application. As shown in Figure 2, the method includes the following steps.
  • Step S201 Obtain a microphone connection request from any one of the avatar anchor terminal and the real person video anchor terminal.
  • the first anchor (which can be a real video anchor or a virtual character anchor) sends an interaction request to the server through its terminal.
  • the interaction request can specify the object to interact with, such as a second anchor; the interaction request can also not specify the object to interact with, but instructs the server to specify an interaction object or randomly match an interaction object, such as randomly matching a second anchor.
  • the method by which the first anchor obtains the interactive anchor object is not limited by this application.
  • Step S202 Send the microphone connection request to the other host terminal among the virtual character host terminal and the real person video host terminal.
  • the server After determining the interactive object of the first anchor (the second anchor), the server sends the interaction request to the second anchor terminal. After receiving the request, the second anchor terminal presents the interaction request information to the second anchor. If the second anchor does not agree to interact, the second anchor terminal sends a rejection request message to the server, and specifies the interaction object after the first anchor In the mode, the server side feeds back the message that the interaction request is rejected to the first anchor terminal; in the mode of randomly matching interactive objects, the server side continues to match the next interactive object until it matches an object that accepts the interaction request or receives all pending interaction objects. The rejection message for the matching object. If the second anchor agrees to interact, the second anchor terminal sends an interaction confirmation message to the server.
  • Step S203 in response to the other party's anchor terminal confirming the connection request, sending the avatar image instruction of the avatar anchor terminal to the real person video anchor terminal and the audience terminal in the real person video anchor's live broadcast room, and The video stream of the real person video anchor terminal is sent to the avatar anchor terminal and the audience terminal in the avatar anchor live broadcast room, where the avatar image instruction is used to call the avatar anchor terminal, the real person video Elements in the set of preset virtual elements in the host terminal and audience terminal.
  • the server After receiving the interaction confirmation from the second anchor, the server sends the avatar image instructions of the avatar anchor terminal to the real video anchor terminal and the audience terminals in the real video anchor live broadcast room, and sends the video stream of the real video anchor terminal to The avatar anchor terminal and the audience terminal in the avatar anchor live broadcast room, where the avatar image instruction is used to call the elements in the virtual character anchor terminal, the live video anchor terminal and the audience terminal that are preset in the set of virtual elements.
  • the real-person video anchor terminal After receiving the avatar image instruction sent by the avatar host terminal, the real-person video anchor terminal calls the virtual elements in the preset virtual element set to display the avatar image of the avatar anchor. After receiving the video stream sent by the real video anchor terminal, the virtual character anchor displays the image of the real video anchor based on the video stream. After receiving the avatar image command sent by the avatar anchor terminal, the audience terminal calls the virtual elements in the preset virtual element set to display the avatar image of the avatar anchor and the interaction process between the anchors, such as displaying the virtual character between anchors. Animations such as head touching and virtual tapping. After receiving the video stream sent by the live video host terminal, the audience terminal displays the live video host's continuous wheat image according to the video stream.
  • the real person video anchor can see the avatar image of the avatar anchor
  • the avatar anchor can see the real person image of the real video anchor
  • the audience can see the avatar image of the avatar anchor and the real person image of the real video anchor.
  • Mai image During the live interaction process, the anchor terminal and the audience terminal can display the anchor screen in various forms.
  • the audience will see two forms, one is the form of the avatar anchor, and the other is the form of the real video anchor.
  • the server in addition to receiving avatar image instructions and video streams from the host terminal, the server also receives instructions from the audience terminals.
  • the method executed by the server to realize the mixed continuous broadcast of virtual characters and real-life videos also includes:
  • Step S204 Obtain the virtual special effects instruction from the first audience terminal among the audience terminals.
  • Step S205 Send the virtual special effects instruction to the avatar anchor terminal, the real person video anchor terminal, and other audience terminals except the first audience terminal.
  • Audiences can participate in the process of watching the host's interaction, for example, by enhancing the atmosphere, giving rewards, etc. For example, viewers can tip their favorite anchors and choose the reward effect.
  • the interactive effects available for the interaction between the audience and the live video anchor and the avatar anchor can be different.
  • virtual character anchors can be given unique virtual gifts and virtual special effects
  • real video anchors can be given ordinary continuous gifts and special effects gifts.
  • the interaction forms for the audience to choose are also different depending on the interaction objects.
  • the interaction objects are the real video anchor and the virtual character anchor, and the content of the interactive effect selection list presented to the audience by the audience terminal may be different.
  • the audience terminal sends a common effect instruction representing the selected interaction effect (such as a reward effect) to the server side, and the server side sends the instruction to the anchor terminal and other audience terminals.
  • the anchor terminal After receiving the instruction, other audience terminals will display interactive effects according to the instruction.
  • the audience terminal sends a virtual special effects instruction representing the selected interaction effect (such as a reward effect) to the server side, and the server side sends the instruction to the anchor terminal and other audience terminals, and the anchor terminal and other After receiving the instruction, the audience terminal calls the virtual elements in the preset virtual element set according to the instruction to display the interactive effect.
  • the interactive effects available for the audience to interact with the real video anchor and the avatar anchor are not restricted by whether the anchor is a real video anchor or an avatar anchor. That is, ordinary effect instructions can be sent to the avatar anchor, and the avatar anchor can The terminal presents normal effects; virtual special effects instructions can be sent to the live video anchor, and the live video anchor terminal presents virtual effects.
  • the method executed by the server to realize the mixed continuous broadcast of virtual characters and real-life videos also includes:
  • Step S206 Obtain virtual interaction instructions from the live video anchor terminal.
  • Step S207 Send the virtual interaction instruction to the virtual character anchor terminal and the audience terminal.
  • the real-person video anchor interacts with the avatar anchor.
  • the real-person video anchor selects the content and effect of the interaction, generates virtual interaction instructions and sends them to the server.
  • the server sends the virtual interaction instructions to the avatar anchor terminal and the audience terminal.
  • the avatar anchor terminal and the audience terminal call the virtual elements in the preset set of virtual elements according to the virtual interaction instructions to display the interaction effect between the real video anchor and the avatar anchor.
  • the server At the end of the mixed broadcast, the server will stop the interaction between the two anchors' live broadcast rooms.
  • the method executed by the server to realize the mixed continuous broadcast of virtual characters and real-life videos also includes:
  • Step S208 in response to the end of the continuous broadcast, stop sending the avatar image instructions of the avatar anchor terminal to the real person video anchor terminal and the audience terminals in the live broadcast room of the real video anchor, and stop sending the avatar image instructions of the real person video anchor terminal.
  • the video stream is sent to the virtual character anchor terminal and the audience terminal in the virtual character anchor live broadcast room.
  • the audience terminal restores the image of the host before the mixed connection, that is, the audience in the virtual character anchor's live broadcast room restores the avatar image image, and the audience in the real video anchor's live broadcast room restores the real person video image.
  • Figure 3(a) is a flow chart of a method executed by an avatar anchor terminal to implement a hybrid video connection between an avatar and a real person according to an embodiment of the present application. As shown in Figure 3(a), the method includes the following steps.
  • Step S301 Generate an avatar image instruction and send the avatar image instruction to the server.
  • the avatar anchor In order to implement avatar live broadcast, the avatar anchor needs to preset a set of virtual elements (for example, downloaded from the server) before starting the broadcast, and needs to set up his own avatar.
  • the virtual character anchor can set his own eyes, ears, and his own clothing.
  • the avatar anchor personalizes his own avatar, when the broadcast starts, he will form an avatar image command corresponding to the setting of the avatar image, and send the avatar image command to the server, and the server will send the avatar image to the server.
  • Image instructions are sent to the audience terminals in the host's live broadcast room.
  • the audience terminal In order to present a virtual character live broadcast, the audience terminal needs to preset a set of virtual elements (for example, downloaded from the server) at the same time or before entering the live broadcast room. After entering the live broadcast room, the audience terminal can receive the virtual character image command sent by the host terminal from the server, and call the preset virtual element set according to the virtual character image command to present the effect of the virtual character live broadcast.
  • a set of virtual elements for example, downloaded from the server
  • the virtual character image instructions include static element instructions and dynamic element instructions.
  • the static element instructions include the virtual image presented in the live broadcast, for example, the static state presented by the anchor (what the body parts look like, what clothes and hats are worn) ), the background of the avatar and the atmosphere or rewards sent by the anchor, etc.; while the avatar anchor will have various expressions and actions during the live broadcast, such as blinking, dancing, smiling, etc., the avatar anchor will The terminal (for example, through a camera) will capture these actions, and the resulting instructions are dynamic element instructions.
  • the dynamic element instructions are used to instruct the host's avatar to perform various actions.
  • the anchor's terminal captures the blinking action, forms an instruction instructing the anchor's avatar to blink, and sends the instruction to the server.
  • the server then sends the instruction to the audience terminal, and the audience terminal responds to the instruction.
  • Step S302 Receive the video stream from the live video anchor terminal.
  • Step S303 Display the live image of the live video anchor based on the video stream.
  • the server sends the avatar image command of the avatar anchor terminal to the real video anchor terminal and the audience terminal in the live broadcast room of the real video anchor, and sends the video stream of the real video anchor terminal To the virtual character anchor terminal and the audience terminal in the virtual character anchor live broadcast room.
  • the virtual character anchor displays the image of the real video anchor based on the video stream.
  • the virtual character anchor terminal not only receives the video stream from the real video anchor terminal, but also receives instructions from the audience terminal.
  • the method performed by the virtual character anchor terminal to realize the mixed broadcast of virtual characters and real people's videos also includes:
  • Step S304 Receive virtual special effects instructions from the audience terminal.
  • Step S305 Call a virtual element in a preset set of virtual elements to display a virtual special effect according to the virtual special effect instruction.
  • Audiences can participate in the process of watching the host's interaction, for example, by enhancing the atmosphere, giving rewards, etc. For example, viewers can tip their favorite anchors and choose the reward effect.
  • the interactive effects available for the interaction between the audience and the live video anchor and the avatar anchor can be different.
  • virtual character anchors can be given unique virtual gifts and virtual special effects
  • real video anchors can be given ordinary continuous gifts and special effects gifts.
  • the interaction forms for the audience to choose are also different depending on the interaction objects.
  • the interaction objects are the real video anchor and the virtual character anchor, and the content of the interactive effect selection list presented to the audience by the audience terminal may be different.
  • the audience terminal sends a common effect instruction representing the selected interaction effect (such as a reward effect) to the server side, and the server side sends the instruction to the anchor terminal and other audience terminals.
  • the anchor terminal After receiving the instruction, other audience terminals will display interactive effects according to the instruction.
  • the audience terminal sends a virtual special effects instruction representing the selected interaction effect (such as a reward effect) to the server side, and the server side sends the instruction to the anchor terminal and other audience terminals, and the anchor terminal and other After receiving the instruction, the audience terminal calls the virtual elements in the preset virtual element set according to the instruction to display the interactive effect.
  • the interactive effects available for the audience to interact with the real video anchor and the avatar anchor are not restricted by whether the anchor is a real video anchor or an avatar anchor. That is, ordinary effect instructions can be sent to the avatar anchor, and the avatar anchor can The terminal presents normal effects; virtual special effects instructions can be sent to the live video anchor, and the live video anchor terminal presents virtual effects.
  • the method executed by the avatar anchor terminal to realize the mixed connection between the avatar and the real person video also includes:
  • Step S306 Receive virtual interaction instructions from the live video anchor terminal.
  • Step S307 Call the virtual elements in the preset set of virtual elements to display interactive special effects according to the virtual interaction instruction.
  • the real-person video anchor interacts with the avatar anchor.
  • the real-person video anchor selects the content and effect of the interaction, generates virtual interaction instructions and sends them to the server.
  • the server sends the virtual interaction instructions to the avatar anchor terminal and the audience terminal.
  • the avatar anchor terminal and the audience terminal call the virtual elements in the preset set of virtual elements according to the virtual interaction instructions to display the interaction effect between the real video anchor and the avatar anchor.
  • the avatar anchor may or may not start broadcasting.
  • the avatar anchor sends a connection request to the server; when the broadcast is not started, the avatar anchor first enters the live broadcast room of the real video anchor as a viewer, and then sends a request to the real video anchor's terminal through the server. Lian Mai requested.
  • the method executed by the avatar anchor terminal to implement a virtual character and real person video mixed broadcast may also include steps S308 and Step S309.
  • Step S308 enter the live broadcast room of the live video anchor
  • Step S309 Send a microphone connection request to the live video anchor terminal through the server.
  • Figure 4(a) is a flow chart of a method executed by a real-life video anchor terminal to implement a hybrid connection between virtual characters and real-life videos according to an embodiment of the present application. As shown in Figure 4(a), the method includes the following steps.
  • Step S401 Generate a live video stream and send the live video stream to the server.
  • Step S402 Receive an avatar image instruction from the avatar host terminal.
  • Step S403 Call the virtual elements in the preset virtual element set according to the virtual character image instruction to display the virtual character image corresponding to the host.
  • the server sends the avatar image command of the avatar anchor terminal to the real video anchor terminal and the audience terminal in the live broadcast room of the real video anchor, and sends the video stream of the real video anchor terminal to the avatar anchor terminal and the audience terminal in the avatar anchor live broadcast room, where the avatar image instruction is used to call elements in a set of preset virtual elements in the avatar anchor terminal, the real person video anchor terminal and the audience terminal.
  • the real-person video anchor terminal After receiving the avatar image instruction sent by the avatar host terminal, the real-person video anchor terminal calls the virtual elements in the preset virtual element set to display the avatar image of the avatar anchor. After receiving the video stream sent by the real video anchor terminal, the virtual character anchor displays the image of the real video anchor based on the video stream. After receiving the avatar image command sent by the avatar anchor terminal, the audience terminal calls the virtual elements in the preset virtual element set to display the avatar image of the avatar anchor and the interaction process between the anchors, such as displaying the virtual character between anchors. Animations such as head touching and virtual tapping. After receiving the video stream sent by the live video host terminal, the audience terminal displays the live video host's continuous wheat image according to the video stream.
  • the live video anchor terminal in addition to receiving avatar image instructions from the avatar host terminal, the live video anchor terminal also receives instructions from the audience terminal.
  • the method executed by the real-person video anchor terminal to realize the mixed continuous broadcast of virtual characters and real-person videos also includes:
  • Step S404 Receive virtual special effects instructions from the audience terminal.
  • Step S405 Call a virtual element in a preset set of virtual elements to display a virtual special effect according to the virtual special effect instruction.
  • Audiences can participate in the process of watching the host's interaction, for example, by enhancing the atmosphere, giving rewards, etc. For example, viewers can tip their favorite anchors and choose the reward effect.
  • the interactive effects available for the interaction between the audience and the live video anchor and the avatar anchor can be different.
  • virtual character anchors can be given unique virtual gifts and virtual special effects
  • real video anchors can be given ordinary continuous gifts and special effects gifts.
  • the interaction forms for the audience to choose are also different depending on the interaction objects.
  • the interaction objects are the real video anchor and the virtual character anchor, and the content of the interactive effect selection list presented to the audience by the audience terminal may be different.
  • the audience terminal sends a common effect instruction representing the selected interaction effect (such as a reward effect) to the server side, and the server side sends the instruction to the anchor terminal and other audience terminals.
  • the anchor terminal After receiving the instruction, other audience terminals will display interactive effects according to the instruction.
  • the audience terminal sends a virtual special effects instruction representing the selected interaction effect (such as a reward effect) to the server side, and the server side sends the instruction to the anchor terminal and other audience terminals, and the anchor terminal and other After receiving the instruction, the audience terminal calls the virtual elements in the preset virtual element set according to the instruction to display the interactive effect.
  • the interactive effects available for the audience to interact with the real video anchor and the avatar anchor are not restricted by whether the anchor is a real video anchor or an avatar anchor. That is, ordinary effect instructions can be sent to the avatar anchor, and the avatar anchor can The terminal presents normal effects; virtual special effects instructions can be sent to the live video anchor, and the live video anchor terminal presents virtual effects.
  • real-person video anchors and virtual character anchors can interact.
  • the method executed by the real-person video anchor terminal to realize the mixed continuous broadcast of virtual characters and real-person videos also includes:
  • Step S406 Send a virtual interaction instruction to the server.
  • the real-person video anchor interacts with the avatar anchor.
  • the real-person video anchor selects the content and effect of the interaction, generates virtual interaction instructions and sends them to the server.
  • the server sends the virtual interaction instructions to the avatar anchor terminal and the audience terminal.
  • the avatar anchor terminal and the audience terminal call the virtual elements in the preset set of virtual elements according to the virtual interaction instructions to display the interaction effect between the real video anchor and the avatar anchor.
  • the real video anchor may or may not start broadcasting.
  • the live video anchor sends a connection request to the server; when the broadcast is not started, the real video anchor first enters the avatar anchor's live broadcast room as a viewer, and then sends a request to the avatar anchor's terminal through the server. Lian Mai requested.
  • the method executed by the avatar anchor terminal to implement a virtual character and real person video mixed broadcast may also include steps S407 and Step S408.
  • Step S407 enter the live broadcast room of the virtual character anchor
  • Step S408 Send a microphone connection request to the avatar anchor terminal through the server.
  • Figure 5 is a flowchart of a method executed by a viewer terminal to implement a hybrid video link between a virtual character and a real person according to an embodiment of the present application. As shown in Figure 5, the method includes the following steps.
  • Step S501 Receive an avatar image instruction from an avatar host terminal.
  • Step S502 Receive the live video stream from the live video anchor terminal.
  • Step S503 Call the virtual elements in the preset set of virtual elements according to the virtual character image instruction to display the virtual character image of the virtual character host, and display the Lianmai image of the real video host according to the live video stream.
  • the server sends the avatar image command of the avatar anchor terminal to the real video anchor terminal and the audience terminal in the live broadcast room of the real video anchor, and sends the video stream of the real video anchor terminal to the avatar anchor terminal and the audience terminal in the avatar anchor live broadcast room, where the avatar image instruction is used to call elements in a set of preset virtual elements in the avatar anchor terminal, the real person video anchor terminal and the audience terminal.
  • the real-person video anchor terminal After receiving the avatar image instruction sent by the avatar host terminal, the real-person video anchor terminal calls the virtual elements in the preset virtual element set to display the avatar image of the avatar anchor. After receiving the video stream sent by the real video anchor terminal, the virtual character anchor displays the image of the real video anchor based on the video stream. After receiving the avatar image command sent by the avatar anchor terminal, the audience terminal calls the virtual elements in the preset virtual element set to display the avatar image of the avatar anchor and the interaction process between the anchors, such as displaying the virtual character between anchors. Animations such as head touching and virtual tapping. After receiving the video stream sent by the live video host terminal, the audience terminal displays the live video host's continuous wheat image according to the video stream.
  • audience terminals can also participate in interactions, for example, tipping the anchor.
  • the method executed by the audience terminal to realize the mixed continuous broadcast of virtual characters and real-life videos also includes:
  • Step S504 Generate a first virtual special effect instruction and send the first virtual special effect instruction to the server.
  • Step S505 Receive a second virtual special effect instruction from the server.
  • Step S506 Call a virtual element in a preset set of virtual elements to display a virtual special effect according to the second virtual special effect instruction.
  • Audiences can participate in the process of watching the host's interaction, for example, by enhancing the atmosphere, giving rewards, etc. For example, viewers can tip their favorite anchors and choose the reward effect.
  • the interactive effects available for the interaction between the audience and the live video anchor and the avatar anchor can be different.
  • virtual character anchors can be given unique virtual gifts and virtual special effects
  • real video anchors can be given ordinary continuous gifts and special effects gifts.
  • the interaction forms for the audience to choose are also different depending on the interaction objects.
  • the interaction objects are the real video anchor and the virtual character anchor, and the content of the interactive effect selection list presented to the audience by the audience terminal may be different.
  • the audience terminal sends a common effect instruction representing the selected interaction effect (such as a reward effect) to the server side, and the server side sends the instruction to the anchor terminal and other audience terminals.
  • the anchor terminal After receiving the instruction, other audience terminals will display interactive effects according to the instruction.
  • the audience terminal sends a virtual special effects instruction representing the selected interaction effect (such as a reward effect) to the server side, and the server side sends the instruction to the anchor terminal and other audience terminals, and the anchor terminal and other After receiving the instruction, the audience terminal calls the virtual elements in the preset virtual element set according to the instruction to display the interactive effect.
  • the interactive effects available for the audience to interact with the real video anchor and the avatar anchor are not restricted by whether the anchor is a real video anchor or an avatar anchor. That is, ordinary effect instructions can be sent to the avatar anchor, and the avatar anchor can The terminal presents normal effects; virtual special effects instructions can be sent to the live video anchor, and the live video anchor terminal presents virtual effects.
  • the method executed by the audience terminal to realize the mixed continuous broadcast of virtual characters and real-life videos also includes:
  • Step S507 Receive virtual interaction instructions from the live video anchor terminal.
  • Step S508 Call the virtual elements in the preset set of virtual elements to display interactive special effects according to the virtual interaction instruction.
  • the real-person video anchor interacts with the avatar anchor.
  • the real-person video anchor selects the content and effect of the interaction, generates virtual interaction instructions and sends them to the server.
  • the server sends the virtual interaction instructions to the avatar anchor terminal and the audience terminal.
  • the avatar anchor terminal and the audience terminal call the virtual elements in the preset set of virtual elements according to the virtual interaction instructions to display the interaction effect between the real video anchor and the avatar anchor.
  • avatar elements related to the host's avatar are preset on the host's and audience's terminals respectively, and the avatar elements have corresponding avatar element instructions.
  • the anchor terminal or the audience terminal sends virtual element instructions to other terminals through the server.
  • the anchor terminal and the audience terminal call the preset avatar elements according to the avatar element instructions to present virtual effects.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • General Engineering & Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Information Transfer Between Computers (AREA)

Abstract

本申请提供一种实现虚拟人物和真人视频混合连麦的系统和方法,该方法包括:获取来自虚拟人物主播终端和真人视频主播终端中任一者的连麦请求;将所述连麦请求发送至对方主播终端;响应于所述对方主播终端对所述连麦请求的确认,将所述虚拟人物主播终端的虚拟人物形象指令发送至所述真人视频主播终端和真人视频主播直播间的观众终端,以及将所述真人视频主播终端的视频流发送至所述虚拟人物主播终端和虚拟人物主播直播间的观众终端,其中,所述虚拟人物形象指令用于调用所述虚拟人物主播终端、所述真人视频主播终端和观众终端中预置的虚拟元素集中的元素。根据本申请的方案,丰富了直播互动的呈现形式,为观众带来更佳的观看体验。

Description

实现虚拟人物和真人视频混合连麦的系统和方法 技术领域
本申请涉及直播领域,尤其涉及一种实现虚拟人物和真人视频混合连麦的系统和方法。
背景技术
现有技术中,直播的主播之间的互动过程,需要主播打开各自的摄像头,摄像头摄制主播的直播视频,并将直播视频发送到直播间的各个观众的终端。观众在观看主播直播或互动的过程中,可以给主播留言、打赏或发送气氛或效果内容,在主播的直播过程中进行呈现。
然而,这种真人直播形式的互动过程比较单一,直播场景比较单调,观众的观看体验有待进一步丰富。
发明内容
基于此,本申请提供了一种实现虚拟人物和真人视频混合连麦的方案,在主播的连麦互动过程中,不再局限于真人之间互动形式,而是真人视频主播与虚拟人物主播之间的互动,丰富了直播互动的呈现形式,为观众带来更佳的观看体验。
根据本申请的第一个方面,提供一种实现虚拟人物和真人视频混合连麦的方法,应用于服务器端,包括:
获取来自虚拟人物主播终端和真人视频主播终端中任一者的连麦请求;
将所述连麦请求发送至所述虚拟人物主播终端和真人视频主播终端中的对方主播终端;
响应于所述对方主播终端对所述连麦请求的确认,将所述虚拟人物主播终端的虚拟人物形象指令发送至所述真人视频主播终端和真人视频主播直播间的观众终端,以及将所述真人视频主播终端的视频流发送 至所述虚拟人物主播终端和虚拟人物主播直播间的观众终端,其中,所述虚拟人物形象指令用于调用所述虚拟人物主播终端、所述真人视频主播终端和观众终端中预置的虚拟元素集中的元素。
根据本申请的第二个方面,提供一种实现虚拟人物和真人视频混合连麦的方法,应用于虚拟人物主播终端,包括:
生成虚拟人物形象指令,并将所述虚拟人物形象指令发送至服务器端;
接收来自真人视频主播终端的视频流;以及
根据所述视频流,显示真人视频主播的连麦形象。
根据本申请的第三个方面,提供一种实现虚拟人物和真人视频混合连麦的方法,应用于真人视频主播终端,包括:
生成直播视频流,并将所述直播视频流发送至服务器端;
接收来自虚拟人物主播终端的虚拟人物形象指令;以及
根据所述虚拟人物形象指令调用预置的虚拟元素集中的虚拟元素显示对应主播的虚拟人物形象。
根据本申请的第四个方面,提供一种实现虚拟人物和真人视频混合连麦的方法,应用于观众终端,包括:
接收来自虚拟人物主播终端的虚拟人物形象指令;
接收来自真人视频主播终端的直播视频流;以及
根据所述虚拟人物形象指令调用预置的虚拟元素集中的虚拟元素显示所述虚拟人物主播的虚拟人物形象,并根据所述直播视频流显示真人视频主播的连麦形象。
根据本申请的第五个方面,提供一种实现虚拟人物和真人视频混合连麦的系统,其包括服务器端、虚拟人物主播终端、真人视频主播终端和观众终端,其中,所述服务器端执行如第一个方面所述的方法,所述虚拟人物主播终端执行如第二个方面所述的方法,所述真人视频主播终端执行如第三个方面所述的方法,所述观众终端执行如第四个方面所述的方法。
根据本申请提出的实现虚拟人物和真人视频混合连麦的系统和方法,在主播和观众的终端均分别预置与主播虚拟形象相关的虚拟形象元素,虚拟形象元素有对应的虚拟形象元素指令,主播终端或观众终端通过服务器端向其他终端发送虚拟元素指令,主播终端和观众终端根据虚拟形象元素指令,调用预置的虚拟形象元素,呈现虚拟效果。这样,首先,不限制连麦双方主播必须是相同的主播人物呈现形式,即可以存在真人视频主播与虚拟人物主播的混合形式,这样使得连麦的形式更加灵活,为观众呈现新颖的直播观看体验;其次,对于虚拟人物主播的直播互动,由于只需传输虚拟元素指令,无需传输真实人物直播过程中的真实画面,降低了对网络带宽的需求;再者,主播终端和用户终端发送的虚拟效果(例如气氛效果、打赏效果)更匹配主播的虚拟人物形象,从视觉、互动、可玩性等各个维度比真人直播更加丰富。
附图说明
为了更清楚地说明本申请实施例中的技术方案,下面将对实施例描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,还可以根据这些附图获得其他的附图,而并不超出本申请要求保护的范围。
图1是根据本申请实施例的实现虚拟人物和真人视频混合连麦的系统的示意图。
图2是根据本申请实施例的由服务器端执行的实现虚拟人物和真人视频混合连麦的方法的流程图。
图3(a)是根据本申请一个实施例的由虚拟人物主播终端执行的实现虚拟人物和真人视频混合连麦的方法的流程图。
图3(b)是根据另一个本申请实施例的由虚拟人物主播终端执行的实现虚拟人物和真人视频混合连麦的方法的流程图。
图4(a)是根据本申请一个实施例的由真人视频主播终端执行的实现虚拟人物和真人视频混合连麦的方法的流程图。
图4(b)是根据另一个本申请实施例的由真人视频主播终端执行 的实现虚拟人物和真人视频混合连麦的方法的流程图。
图5是根据本申请实施例的由观众终端执行的实现虚拟人物和真人视频混合连麦的方法的流程图。
具体实施方式
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。
图1是根据本申请实施例的实现虚拟人物和真人视频混合连麦的系统的示意图。如图1所示,实现虚拟人物和真人视频混合连麦的系统包括服务器端、真人视频主播终端、虚拟人物主播终端和观众终端,其中,真人视频主播终端、虚拟人物主播终端和观众终端分别与服务器端连接。
在本申请中,“虚拟人物和真人视频混合连麦”指的是主播在直播过程中存在真人形象和虚拟人物形象。在本申请的实现虚拟人物和真人视频混合连麦的系统中(例如在服务器端)存储了与虚拟人物直播和互动相关的虚拟元素集,这个虚拟元素集包括了与虚拟人物直播和互动相关的各个虚拟元素,例如,与虚拟人物相关的人物形象设定(卡通人物、动物等)、人物的身体部位(五官、头发、四肢等)、人物的穿戴(帽子、衣服、手套等)、虚拟人物所处的背景、主播或观众发送的气氛或打赏等。服务器端、主播终端和观众终端都可以发出相应的指令来调用虚拟元素集中的虚拟元素,从而呈现对应的虚拟效果。
为了实现虚拟人物直播,虚拟人物主播在开播前,需要预置虚拟元素集(例如,从服务器端下载),并需要设置自己的虚拟形象。例如,虚拟人物主播可以设置自己的眼睛、耳朵,还可以设置自己的穿戴等。虚拟人物主播在个性化设置自己的虚拟形象后,在开播时,就形成与该虚拟形象的设置对应的虚拟人物形象指令,并将该虚拟人物形象指令发送至服务器端,服务器端将该虚拟人物形象指令发送至主播直播间内的 观众终端。
观众终端为了呈现虚拟人物直播,在进入直播间的同时或之前,需要预置虚拟元素集(例如,从服务器端下载)。在进入直播间后,观众终端可以从服务器端接收到主播终端发送的虚拟人物形象指令,根据该虚拟人物形象指令调用预置的虚拟元素集,呈现虚拟人物直播的效果。
虚拟人物形象指令包括静态元素指令和动态元素指令,其中,静态元素指令包括直播所呈现的虚拟形象,例如,主播所呈现的静态状态(身体部位是什么样的,穿戴的衣服帽子是什么样的),虚拟人物所处的背景以及主播发送的气氛或打赏是什么样的等;而虚拟人物主播在直播的过程中会存在各种表情和动作,比如眨眼、跳舞、微笑等,虚拟人物主播的终端(例如通过摄像头)会捕捉这些动作,所形成的指令为动态元素指令,动态元素指令用于指示主播的虚拟形象进行各种动作。例如,主播眨眼,主播的终端捕捉到眨眼动作,形成指示主播的虚拟人物形象的眨眼的指令,并将该指令发送到服务器端,服务器端再将该指令发送到观众终端,观众终端根据该指令呈现主播的虚拟人物形象眨眼的效果。
在主播直播的过程中,如果需要与另一个主播互动,则需要向服务器端发送互动请求。例如,第一主播(可以是真人视频主播也可以是虚拟人物主播)通过其终端向服务器端发送互动请求。该互动请求可以指定进行互动的对象,例如指定第二主播;该互动请求还可以不指定进行互动的对象,而指示服务器端指定一个互动对象或随机匹配一个互动对象,例如随机匹配了第二主播。第一主播获取互动主播对象的方式本申请不做限定。
在真人视频主播和虚拟人物主播建立连麦的过程中,发起连麦请求的主播(可以是真人视频主播也可以是虚拟人物主播)可以开播也可以不开播。在开播的情况下,发起连麦请求的主播向服务器端发送连麦请求;在不开播的情况下,发起连麦请求的主播首先以观众的身份进入连麦对象主播的直播间,然后通过服务器端向对象主播终端发送连麦请求。
在确定第一主播的互动对象(第二主播)后,服务器端将互动请求 发送至第二主播终端。第二主播终端在收到该请求后,向第二主播呈现该互动请求信息,若第二主播不同意进行互动,则第二主播终端向服务器端发送拒绝请求消息,在第一主播指定互动对象的模式下,服务器端向第一主播终端反馈互动请求被拒绝的消息;在随机匹配互动对象的模式下,服务器端继续匹配下一个互动对象,直至匹配到接受互动请求的对象或收到所有待匹配对象的拒绝消息。若第二主播同意进行互动,则第二主播终端向服务器端发送互动确认消息。
服务器端在收到第二主播的互动确认后,将虚拟人物主播终端的虚拟人物形象指令发送至真人视频主播终端和真人视频主播直播间的观众终端,并将真人视频主播终端的视频流发送至虚拟人物主播终端和虚拟人物主播直播间的观众终端,其中,虚拟人物形象指令用于调用虚拟人物主播终端、真人视频主播终端和观众终端中预置的虚拟元素集中的元素。
真人视频主播终端在收到虚拟人物主播终端发送的虚拟人物形象指令后,调用预置的虚拟元素集中的虚拟元素,显示虚拟人物主播的虚拟人物形象。虚拟人物主播在收到真人视频主播终端发送的视频流后,根据视频流,显示真人视频主播的连麦形象。观众终端在收到虚拟人物主播终端发送的虚拟人物形象指令后,调用预置的虚拟元素集中的虚拟元素,显示虚拟人物主播的虚拟人物形象以及主播之间的互动过程,例如显示主播之间虚拟摸头、虚拟敲打等动画。观众终端在收到真人视频主播终端发送的视频流后,根据视频流,显示真人视频主播的连麦形象。
这样,真人视频主播可以看到虚拟人物主播的虚拟人物形象,虚拟人物主播可以看到真人视频主播的真人连麦形象,而观众可以看到虚拟人物主播的虚拟人物形象以及真人视频主播的真人连麦形象。在直播互动的过程中,主播终端和观众终端可以采用各种形式显示主播画面,例如,观众将看到两个窗体,一个是虚拟形象主播的窗体,一个是真人视频主播的窗体,两个窗体一个全屏,一个小屏,可以通过点击小屏的方式,实现两种形象展示的切换,即大屏(虚拟形象),小屏(真人视频),或者大屏(真人视频),小屏(虚拟形象),本申请对此不做限制。
观众在观看主播互动的过程中可以参与进来,例如通过渲染气氛、 打赏等形式。例如,观众向喜欢的主播打赏,选择打赏效果。观众与真人视频主播和虚拟人物主播之间的互动可供选择的互动效果可以是不同的。例如,对于赠送礼物来说,对虚拟人物主播可以赠送特有的虚拟礼物及虚拟特效,而对于真人视频主播可以赠送普通连发礼物及特效礼物。这样,当观众希望与主播进行互动时,互动对象的不同,供观众选择的互动形式也不同。例如,当观众选择与主播进行互动时,互动对象为真人视频主播和虚拟人物主播,观众终端为观众呈现的互动效果选择列表的内容可以是不同的。
这样,对于观众与真人视频主播的互动,观众终端将代表所选互动效果(例如打赏效果)的普通效果指令发送至服务器端,服务器端将该指令发送至主播终端和其他观众终端,主播终端和其他观众终端在收到该指令后,根据该指令显示互动效果。对于观众与虚拟人物主播的互动,观众终端将代表所选互动效果(例如打赏效果)的虚拟特效指令发送至服务器端,服务器端将该指令发送至主播终端和其他观众终端,主播终端和其他观众终端在收到该指令后,根据该指令调用预置的虚拟元素集中的虚拟元素,显示互动效果。
或者,观众与真人视频主播和虚拟人物主播之间的互动可供选择的互动效果不受主播是真人视频主播还是虚拟人物主播的限制,即,对虚拟人物主播可以发送普通效果指令,虚拟人物主播终端呈现普通效果;对真人视频主播可以发送虚拟特效,例如基于人脸识别的特效,真人视频主播终端呈现虚拟效果。
在混合连麦过程中,真人视频主播和虚拟人物主播之间可以进行互动。例如,真人视频主播对虚拟人物主播拍一拍动作等。其中,真人视频主播向虚拟人物主播的互动,真人视频主播选择互动的内容和效果,生成虚拟互动指令并向服务器端发送的虚拟互动指令,服务器端将虚拟互动指令发送至虚拟人物主播终端和观众终端。虚拟人物主播终端和观众终端根据虚拟互动指令调用预置的虚拟元素集中的虚拟元素,显示真人视频主播与虚拟人物主播之间的互动过效果。
上述实现混合连麦的方案中,为了便于说明,采用两个主播的混合连麦进行描述。然而,本领域技术人员可以理解的是,本申请混合连麦 的方案的主播的数量可以是任意的,而且,只要存在真人视频主播和虚拟人物主播就可以,真人视频主播和虚拟人物主播的数量不做限制。
采用本申请的混合连麦互动方式,不限制连麦双方主播必须是相同的主播人物呈现形式,即可以存在真人视频主播与虚拟人物主播的混合形式,这样使得连麦的形式更加灵活,为观众呈现新颖的直播观看体验。另外,在混合连麦的过程中,主播终端、观众终端和服务器端发送的调用虚拟元素集的指令所呈现的效果,相对于真人直播,与虚拟人物更匹配,例如,观众为主播打赏了一个帽子的虚拟效果,这个帽子能够刚好戴在主播的虚拟人物的头上,而如果主播是真人形象,匹配效果不是很好。这样,采用本申请的混合连麦互动方式,从视觉、互动、可玩性等各个维度比真人直播更加丰富。
根据图1所示的实现虚拟人物和真人视频混合连麦的系统以及组成该系统的服务器端和终端的运行过程,本申请还提供一种实现虚拟人物和真人视频混合连麦的方法,分别如图2、图3(a)、图3(b)、图4(a)、图4(b)和图5所示。
图2是根据本申请实施例的由服务器端执行的实现虚拟人物和真人视频混合连麦的方法的流程图。如图2所示,该方法包括如下步骤。
步骤S201,获取来自虚拟人物主播终端和真人视频主播终端中任一者的连麦请求。
在主播直播的过程中,如果需要与另一个主播互动,则需要向服务器端发送互动请求。例如,第一主播(可以是真人视频主播也可以是虚拟人物主播)通过其终端向服务器端发送互动请求。该互动请求可以指定进行互动的对象,例如指定第二主播;该互动请求还可以不指定进行互动的对象,而指示服务器端指定一个互动对象或随机匹配一个互动对象,例如随机匹配了第二主播。第一主播获取互动主播对象的方式本申请不做限定。
步骤S202,将所述连麦请求发送至所述虚拟人物主播终端和真人视频主播终端中的对方主播终端。
在确定第一主播的互动对象(第二主播)后,服务器端将互动请求发送至第二主播终端。第二主播终端在收到该请求后,向第二主播呈现 该互动请求信息,若第二主播不同意进行互动,则第二主播终端向服务器端发送拒绝请求消息,在第一主播指定互动对象的模式下,服务器端向第一主播终端反馈互动请求被拒绝的消息;在随机匹配互动对象的模式下,服务器端继续匹配下一个互动对象,直至匹配到接受互动请求的对象或收到所有待匹配对象的拒绝消息。若第二主播同意进行互动,则第二主播终端向服务器端发送互动确认消息。
步骤S203,响应于所述对方主播终端对所述连麦请求的确认,将所述虚拟人物主播终端的虚拟人物形象指令发送至所述真人视频主播终端和真人视频主播直播间的观众终端,以及将所述真人视频主播终端的视频流发送至所述虚拟人物主播终端和虚拟人物主播直播间的观众终端,其中,所述虚拟人物形象指令用于调用所述虚拟人物主播终端、所述真人视频主播终端和观众终端中预置的虚拟元素集中的元素。
服务器端在收到第二主播的互动确认后,将虚拟人物主播终端的虚拟人物形象指令发送至真人视频主播终端和真人视频主播直播间的观众终端,并将真人视频主播终端的视频流发送至虚拟人物主播终端和虚拟人物主播直播间的观众终端,其中,虚拟人物形象指令用于调用虚拟人物主播终端、真人视频主播终端和观众终端中预置的虚拟元素集中的元素。
真人视频主播终端在收到虚拟人物主播终端发送的虚拟人物形象指令后,调用预置的虚拟元素集中的虚拟元素,显示虚拟人物主播的虚拟人物形象。虚拟人物主播在收到真人视频主播终端发送的视频流后,根据视频流,显示真人视频主播的连麦形象。观众终端在收到虚拟人物主播终端发送的虚拟人物形象指令后,调用预置的虚拟元素集中的虚拟元素,显示虚拟人物主播的虚拟人物形象以及主播之间的互动过程,例如显示主播之间虚拟摸头、虚拟敲打等动画。观众终端在收到真人视频主播终端发送的视频流后,根据视频流,显示真人视频主播的连麦形象。
这样,真人视频主播可以看到虚拟人物主播的虚拟人物形象,虚拟人物主播可以看到真人视频主播的真人连麦形象,而观众可以看到虚拟人物主播的虚拟人物形象以及真人视频主播的真人连麦形象。在直播互动的过程中,主播终端和观众终端可以采用各种形式显示主播画面,例 如,观众将看到两个窗体,一个是虚拟形象主播的窗体,一个是真人视频主播的窗体,两个窗体一个全屏,一个小屏,可以通过点击小屏的方式,实现两种形象展示的切换,即大屏(虚拟形象),小屏(真人视频),或者大屏(真人视频),小屏(虚拟形象),本申请对此不做限制。
在混合连麦过程中,服务器端除了收到来自主播终端的虚拟人物形象指令和视频流,还会收到来自观众终端的指令。这样,由服务器端执行的实现虚拟人物和真人视频混合连麦的方法还包括:
步骤S204,获取来自所述观众终端中的第一观众终端的虚拟特效指令。
步骤S205,将所述虚拟特效指令发送至所述虚拟人物主播终端、所述真人视频主播终端和除所述第一观众终端的其他观众终端。
观众在观看主播互动的过程中可以参与进来,例如通过渲染气氛、打赏等形式。例如,观众向喜欢的主播打赏,选择打赏效果。观众与真人视频主播和虚拟人物主播之间的互动可供选择的互动效果可以是不同的。例如,对于赠送礼物来说,对虚拟人物主播可以赠送特有的虚拟礼物及虚拟特效,而对于真人视频主播可以赠送普通连发礼物及特效礼物。这样,当观众希望与主播进行互动时,互动对象的不同,供观众选择的互动形式也不同。例如,当观众选择与主播进行互动时,互动对象为真人视频主播和虚拟人物主播,观众终端为观众呈现的互动效果选择列表的内容可以是不同的。
这样,对于观众与真人视频主播的互动,观众终端将代表所选互动效果(例如打赏效果)的普通效果指令发送至服务器端,服务器端将该指令发送至主播终端和其他观众终端,主播终端和其他观众终端在收到该指令后,根据该指令显示互动效果。对于观众与虚拟人物主播的互动,观众终端将代表所选互动效果(例如打赏效果)的虚拟特效指令发送至服务器端,服务器端将该指令发送至主播终端和其他观众终端,主播终端和其他观众终端在收到该指令后,根据该指令调用预置的虚拟元素集中的虚拟元素,显示互动效果。
或者,观众与真人视频主播和虚拟人物主播之间的互动可供选择的互动效果不受主播是真人视频主播还是虚拟人物主播的限制,即,对虚 拟人物主播可以发送普通效果指令,虚拟人物主播终端呈现普通效果;对真人视频主播可以发送虚拟特效指令,真人视频主播终端呈现虚拟效果。
在混合连麦过程中,真人视频主播和虚拟人物主播之间可以进行互动。这样,由服务器端执行的实现虚拟人物和真人视频混合连麦的方法还包括:
步骤S206,获取来自所述真人视频主播终端的虚拟互动指令。
步骤S207,将所述虚拟互动指令发送至所述虚拟人物主播终端和观众终端。
真人视频主播向虚拟人物主播的互动,真人视频主播选择互动的内容和效果,生成虚拟互动指令并向服务器端发送的虚拟互动指令,服务器端将虚拟互动指令发送至虚拟人物主播终端和观众终端。虚拟人物主播终端和观众终端根据虚拟互动指令调用预置的虚拟元素集中的虚拟元素,显示真人视频主播与虚拟人物主播之间的互动过效果。
在混合连麦结束时,服务器端会停止两个主播直播间之间的互动。这样,由服务器端执行的实现虚拟人物和真人视频混合连麦的方法还包括:
步骤S208,响应于连麦的结束,停止将所述虚拟人物主播终端的虚拟人物形象指令发送至所述真人视频主播终端和真人视频主播直播间的观众终端,停止将所述真人视频主播终端的视频流发送至所述虚拟人物主播终端和虚拟人物主播直播间的观众终端。
这样,观众终端恢复混合连麦之前的主播的画面,即虚拟人物主播直播间的观众恢复虚拟人物形象画面,真人视频主播直播间的观众恢复真人视频画面。
图3(a)是根据本申请一个实施例的由虚拟人物主播终端执行的实现虚拟人物和真人视频混合连麦的方法的流程图。如图3(a)所示,该方法包括如下步骤。
步骤S301,生成虚拟人物形象指令,并将所述虚拟人物形象指令发送至服务器端。
为了实现虚拟人物直播,虚拟人物主播在开播前,需要预置虚拟元素集(例如,从服务器端下载),并需要设置自己的虚拟形象。例如,虚拟人物主播可以设置自己的眼睛、耳朵,还可以设置自己的穿戴等。虚拟人物主播在个性化设置自己的虚拟形象后,在开播时,就形成与该虚拟形象的设置对应的虚拟人物形象指令,并将该虚拟人物形象指令发送至服务器端,服务器端将该虚拟人物形象指令发送至主播直播间内的观众终端。
观众终端为了呈现虚拟人物直播,在进入直播间的同时或之前,需要预置虚拟元素集(例如,从服务器端下载)。在进入直播间后,观众终端可以从服务器端接收到主播终端发送的虚拟人物形象指令,根据该虚拟人物形象指令调用预置的虚拟元素集,呈现虚拟人物直播的效果。
虚拟人物形象指令包括静态元素指令和动态元素指令,其中,静态元素指令包括直播所呈现的虚拟形象,例如,主播所呈现的静态状态(身体部位是什么样的,穿戴的衣服帽子是什么样的),虚拟人物所处的背景以及主播发送的气氛或打赏是什么样的等;而虚拟人物主播在直播的过程中会存在各种表情和动作,比如眨眼、跳舞、微笑等,虚拟人物主播的终端(例如通过摄像头)会捕捉这些动作,所形成的指令为动态元素指令,动态元素指令用于指示主播的虚拟形象进行各种动作。例如,主播眨眼,主播的终端捕捉到眨眼动作,形成指示主播的虚拟人物形象的眨眼的指令,并将该指令发送到服务器端,服务器端再将该指令发送到观众终端,观众终端根据该指令呈现主播的虚拟人物形象眨眼的效果。
步骤S302,接收来自真人视频主播终端的视频流。
步骤S303,根据所述视频流,显示真人视频主播的连麦形象。
真人视频主播和虚拟人物主播确认连麦时,服务器端将虚拟人物主播终端的虚拟人物形象指令发送至真人视频主播终端和真人视频主播直播间的观众终端,并将真人视频主播终端的视频流发送至虚拟人物主播终端和虚拟人物主播直播间的观众终端。虚拟人物主播在收到真人视频主播终端发送的视频流后,根据视频流,显示真人视频主播的连麦形象。
在混合连麦过程中,虚拟人物主播终端除了收到来自真人视频主播终端的视频流,还会收到来自观众终端的指令。这样,由虚拟人物主播终端执行的实现虚拟人物和真人视频混合连麦的方法还包括:
步骤S304,接收来自观众终端的虚拟特效指令。
步骤S305,根据所述虚拟特效指令调用预置的虚拟元素集中的虚拟元素显示虚拟特效。
观众在观看主播互动的过程中可以参与进来,例如通过渲染气氛、打赏等形式。例如,观众向喜欢的主播打赏,选择打赏效果。观众与真人视频主播和虚拟人物主播之间的互动可供选择的互动效果可以是不同的。例如,对于赠送礼物来说,对虚拟人物主播可以赠送特有的虚拟礼物及虚拟特效,而对于真人视频主播可以赠送普通连发礼物及特效礼物。这样,当观众希望与主播进行互动时,互动对象的不同,供观众选择的互动形式也不同。例如,当观众选择与主播进行互动时,互动对象为真人视频主播和虚拟人物主播,观众终端为观众呈现的互动效果选择列表的内容可以是不同的。
这样,对于观众与真人视频主播的互动,观众终端将代表所选互动效果(例如打赏效果)的普通效果指令发送至服务器端,服务器端将该指令发送至主播终端和其他观众终端,主播终端和其他观众终端在收到该指令后,根据该指令显示互动效果。对于观众与虚拟人物主播的互动,观众终端将代表所选互动效果(例如打赏效果)的虚拟特效指令发送至服务器端,服务器端将该指令发送至主播终端和其他观众终端,主播终端和其他观众终端在收到该指令后,根据该指令调用预置的虚拟元素集中的虚拟元素,显示互动效果。
或者,观众与真人视频主播和虚拟人物主播之间的互动可供选择的互动效果不受主播是真人视频主播还是虚拟人物主播的限制,即,对虚拟人物主播可以发送普通效果指令,虚拟人物主播终端呈现普通效果;对真人视频主播可以发送虚拟特效指令,真人视频主播终端呈现虚拟效果。
在混合连麦过程中,真人视频主播和虚拟人物主播之间可以进行互动。这样,由虚拟人物主播终端执行的实现虚拟人物和真人视频混合连 麦的方法还包括:
步骤S306,接收来自所述真人视频主播终端的虚拟互动指令。
步骤S307,根据所述虚拟互动指令调用预置的虚拟元素集中的虚拟元素显示互动特效。
真人视频主播向虚拟人物主播的互动,真人视频主播选择互动的内容和效果,生成虚拟互动指令并向服务器端发送的虚拟互动指令,服务器端将虚拟互动指令发送至虚拟人物主播终端和观众终端。虚拟人物主播终端和观众终端根据虚拟互动指令调用预置的虚拟元素集中的虚拟元素,显示真人视频主播与虚拟人物主播之间的互动过效果。
在一个实施例中,在真人视频主播和虚拟人物主播建立连麦的过程中,虚拟人物主播可以开播也可以不开播。在开播的情况下,虚拟人物主播向服务器端发送连麦请求;在不开播的情况下,虚拟人物主播首先以观众的身份进入真人视频主播的直播间,然后通过服务器端向真人视频主播终端发送连麦请求。
从而,如图3(b)所示,除了图3(a)所示的步骤S301至步骤S307,由虚拟人物主播终端执行的实现虚拟人物和真人视频混合连麦的方法还可以包括步骤S308和步骤S309。
步骤S308,进入所述真人视频主播的直播间;
步骤S309,通过所述服务器端向所述真人视频主播终端发送连麦请求。
图4(a)是根据本申请一个实施例的由真人视频主播终端执行的实现虚拟人物和真人视频混合连麦的方法的流程图。如图4(a)所示,该方法包括如下步骤。
步骤S401,生成直播视频流,并将所述直播视频流发送至服务器端。
步骤S402,接收来自虚拟人物主播终端的虚拟人物形象指令。
步骤S403,根据所述虚拟人物形象指令调用预置的虚拟元素集中的虚拟元素显示对应主播的虚拟人物形象。
真人视频主播和虚拟人物主播确认连麦时,服务器端将虚拟人物主 播终端的虚拟人物形象指令发送至真人视频主播终端和真人视频主播直播间的观众终端,并将真人视频主播终端的视频流发送至虚拟人物主播终端和虚拟人物主播直播间的观众终端,其中,虚拟人物形象指令用于调用虚拟人物主播终端、真人视频主播终端和观众终端中预置的虚拟元素集中的元素。
真人视频主播终端在收到虚拟人物主播终端发送的虚拟人物形象指令后,调用预置的虚拟元素集中的虚拟元素,显示虚拟人物主播的虚拟人物形象。虚拟人物主播在收到真人视频主播终端发送的视频流后,根据视频流,显示真人视频主播的连麦形象。观众终端在收到虚拟人物主播终端发送的虚拟人物形象指令后,调用预置的虚拟元素集中的虚拟元素,显示虚拟人物主播的虚拟人物形象以及主播之间的互动过程,例如显示主播之间虚拟摸头、虚拟敲打等动画。观众终端在收到真人视频主播终端发送的视频流后,根据视频流,显示真人视频主播的连麦形象。
在混合连麦过程中,真人视频主播终端除了收到来自虚拟人物主播终端的虚拟人物形象指令,还会收到来自观众终端的指令。这样,由真人视频主播终端执行的实现虚拟人物和真人视频混合连麦的方法还包括:
步骤S404,接收来自观众终端的虚拟特效指令。
步骤S405,根据所述虚拟特效指令调用预置的虚拟元素集中的虚拟元素显示虚拟特效。
观众在观看主播互动的过程中可以参与进来,例如通过渲染气氛、打赏等形式。例如,观众向喜欢的主播打赏,选择打赏效果。观众与真人视频主播和虚拟人物主播之间的互动可供选择的互动效果可以是不同的。例如,对于赠送礼物来说,对虚拟人物主播可以赠送特有的虚拟礼物及虚拟特效,而对于真人视频主播可以赠送普通连发礼物及特效礼物。这样,当观众希望与主播进行互动时,互动对象的不同,供观众选择的互动形式也不同。例如,当观众选择与主播进行互动时,互动对象为真人视频主播和虚拟人物主播,观众终端为观众呈现的互动效果选择列表的内容可以是不同的。
这样,对于观众与真人视频主播的互动,观众终端将代表所选互动 效果(例如打赏效果)的普通效果指令发送至服务器端,服务器端将该指令发送至主播终端和其他观众终端,主播终端和其他观众终端在收到该指令后,根据该指令显示互动效果。对于观众与虚拟人物主播的互动,观众终端将代表所选互动效果(例如打赏效果)的虚拟特效指令发送至服务器端,服务器端将该指令发送至主播终端和其他观众终端,主播终端和其他观众终端在收到该指令后,根据该指令调用预置的虚拟元素集中的虚拟元素,显示互动效果。
或者,观众与真人视频主播和虚拟人物主播之间的互动可供选择的互动效果不受主播是真人视频主播还是虚拟人物主播的限制,即,对虚拟人物主播可以发送普通效果指令,虚拟人物主播终端呈现普通效果;对真人视频主播可以发送虚拟特效指令,真人视频主播终端呈现虚拟效果。
在混合连麦过程中,真人视频主播和虚拟人物主播之间可以进行互动。这样,由真人视频主播终端执行的实现虚拟人物和真人视频混合连麦的方法还包括:
步骤S406,向所述服务器端发送虚拟互动指令。
真人视频主播向虚拟人物主播的互动,真人视频主播选择互动的内容和效果,生成虚拟互动指令并向服务器端发送的虚拟互动指令,服务器端将虚拟互动指令发送至虚拟人物主播终端和观众终端。虚拟人物主播终端和观众终端根据虚拟互动指令调用预置的虚拟元素集中的虚拟元素,显示真人视频主播与虚拟人物主播之间的互动过效果。
在一个实施例中,在真人视频主播和虚拟人物主播建立连麦的过程中,真人视频主播可以开播也可以不开播。在开播的情况下,真人视频主播向服务器端发送连麦请求;在不开播的情况下,真人视频主播首先以观众的身份进入虚拟人物主播的直播间,然后通过服务器端向虚拟人物主播终端发送连麦请求。
从而,如图4(b)所示,除了图4(a)所示的步骤S401至步骤S406,由虚拟人物主播终端执行的实现虚拟人物和真人视频混合连麦的方法还可以包括步骤S407和步骤S408。
步骤S407,进入所述虚拟人物主播的直播间;
步骤S408,通过所述服务器端向所述虚拟人物主播终端发送连麦请求。
图5是根据本申请实施例的由观众终端执行的实现虚拟人物和真人视频混合连麦的方法的流程图。如图5所示,该方法包括如下步骤。
步骤S501,接收来自虚拟人物主播终端的虚拟人物形象指令。
步骤S502,接收来自真人视频主播终端的直播视频流。
步骤S503,根据所述虚拟人物形象指令调用预置的虚拟元素集中的虚拟元素显示所述虚拟人物主播的虚拟人物形象,并根据所述直播视频流显示真人视频主播的连麦形象。
真人视频主播和虚拟人物主播确认连麦时,服务器端将虚拟人物主播终端的虚拟人物形象指令发送至真人视频主播终端和真人视频主播直播间的观众终端,并将真人视频主播终端的视频流发送至虚拟人物主播终端和虚拟人物主播直播间的观众终端,其中,虚拟人物形象指令用于调用虚拟人物主播终端、真人视频主播终端和观众终端中预置的虚拟元素集中的元素。
真人视频主播终端在收到虚拟人物主播终端发送的虚拟人物形象指令后,调用预置的虚拟元素集中的虚拟元素,显示虚拟人物主播的虚拟人物形象。虚拟人物主播在收到真人视频主播终端发送的视频流后,根据视频流,显示真人视频主播的连麦形象。观众终端在收到虚拟人物主播终端发送的虚拟人物形象指令后,调用预置的虚拟元素集中的虚拟元素,显示虚拟人物主播的虚拟人物形象以及主播之间的互动过程,例如显示主播之间虚拟摸头、虚拟敲打等动画。观众终端在收到真人视频主播终端发送的视频流后,根据视频流,显示真人视频主播的连麦形象。
在混合连麦过程中,观众终端也可以参与互动,例如,对主播进行打赏。这样,由观众终端执行的实现虚拟人物和真人视频混合连麦的方法还包括:
步骤S504,生成第一虚拟特效指令,并将所述第一虚拟特效指令发送至服务器端。
步骤S505,接收来自所述服务器端的第二虚拟特效指令。
步骤S506,根据所述第二虚拟特效指令调用预置的虚拟元素集中的虚拟元素显示虚拟特效。
观众在观看主播互动的过程中可以参与进来,例如通过渲染气氛、打赏等形式。例如,观众向喜欢的主播打赏,选择打赏效果。观众与真人视频主播和虚拟人物主播之间的互动可供选择的互动效果可以是不同的。例如,对于赠送礼物来说,对虚拟人物主播可以赠送特有的虚拟礼物及虚拟特效,而对于真人视频主播可以赠送普通连发礼物及特效礼物。这样,当观众希望与主播进行互动时,互动对象的不同,供观众选择的互动形式也不同。例如,当观众选择与主播进行互动时,互动对象为真人视频主播和虚拟人物主播,观众终端为观众呈现的互动效果选择列表的内容可以是不同的。
这样,对于观众与真人视频主播的互动,观众终端将代表所选互动效果(例如打赏效果)的普通效果指令发送至服务器端,服务器端将该指令发送至主播终端和其他观众终端,主播终端和其他观众终端在收到该指令后,根据该指令显示互动效果。对于观众与虚拟人物主播的互动,观众终端将代表所选互动效果(例如打赏效果)的虚拟特效指令发送至服务器端,服务器端将该指令发送至主播终端和其他观众终端,主播终端和其他观众终端在收到该指令后,根据该指令调用预置的虚拟元素集中的虚拟元素,显示互动效果。
或者,观众与真人视频主播和虚拟人物主播之间的互动可供选择的互动效果不受主播是真人视频主播还是虚拟人物主播的限制,即,对虚拟人物主播可以发送普通效果指令,虚拟人物主播终端呈现普通效果;对真人视频主播可以发送虚拟特效指令,真人视频主播终端呈现虚拟效果。
在混合连麦过程中,真人视频主播和虚拟人物主播之间可以进行互动。这样,由观众终端执行的实现虚拟人物和真人视频混合连麦的方法还包括:
步骤S507,接收来自所述真人视频主播终端的虚拟互动指令。
步骤S508,根据所述虚拟互动指令调用预置的虚拟元素集中的虚拟元素显示互动特效。
真人视频主播向虚拟人物主播的互动,真人视频主播选择互动的内容和效果,生成虚拟互动指令并向服务器端发送的虚拟互动指令,服务器端将虚拟互动指令发送至虚拟人物主播终端和观众终端。虚拟人物主播终端和观众终端根据虚拟互动指令调用预置的虚拟元素集中的虚拟元素,显示真人视频主播与虚拟人物主播之间的互动过效果。
需要说明的是,对于前述的各方法实施例,为了简单描述,故将其都表述为一系列的动作组合,但是本领域技术人员应该知悉,本申请并不受所描述的动作顺序的限制,因为依据本申请,某些步骤可以采用其他顺序或者同时进行。其次,本领域技术人员也应该知悉,说明书中所描述的实施例均属于可选实施例,所涉及的动作和模块并不一定是本申请所必须的。
在上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述的部分,可以参见其他实施例的相关描述。
根据本申请提出的实现虚拟人物和真人视频混合连麦的系统和方法,在主播和观众的终端均分别预置与主播虚拟形象相关的虚拟形象元素,虚拟形象元素有对应的虚拟形象元素指令,主播终端或观众终端通过服务器端向其他终端发送虚拟元素指令,主播终端和观众终端根据虚拟形象元素指令,调用预置的虚拟形象元素,呈现虚拟效果。这样,首先,不限制连麦双方主播必须是相同的主播人物呈现形式,即可以存在真人视频主播与虚拟人物主播的混合形式,这样使得连麦的形式更加灵活,为观众呈现新颖的直播观看体验;其次,对于虚拟人物主播的直播互动,由于只需传输虚拟元素指令,无需传输真实人物直播过程中的真实画面,降低了对网络带宽的需求;再者,主播终端和用户终端发送的虚拟效果(例如气氛效果、打赏效果)更匹配主播的虚拟人物形象,从视觉、互动、可玩性等各个维度比真人直播更加丰富。
以上对本申请实施例进行了详细介绍,本文中应用了具体个例对本申请的原理及实施方式进行了阐述,以上实施例的说明仅用于帮助理解 本申请的方法及其核心思想。同时,本领域技术人员依据本申请的思想,基于本申请的具体实施方式及应用范围上做出的改变或变形之处,都属于本申请保护的范围。综上所述,本说明书内容不应理解为对本申请的限制。

Claims (17)

  1. 一种实现虚拟人物和真人视频混合连麦的方法,应用于服务器端,包括:
    获取来自虚拟人物主播终端和真人视频主播终端中任一者的连麦请求;
    将所述连麦请求发送至所述虚拟人物主播终端和真人视频主播终端中的对方主播终端;
    响应于所述对方主播终端对所述连麦请求的确认,将所述虚拟人物主播终端的虚拟人物形象指令发送至所述真人视频主播终端和真人视频主播直播间的观众终端,以及将所述真人视频主播终端的视频流发送至所述虚拟人物主播终端和虚拟人物主播直播间的观众终端,其中,所述虚拟人物形象指令用于调用所述虚拟人物主播终端、所述真人视频主播终端和观众终端中预置的虚拟元素集中的元素。
  2. 如权利要求1所述的方法,还包括:
    获取来自所述观众终端中的第一观众终端的虚拟特效指令;以及
    将所述虚拟特效指令发送至所述虚拟人物主播终端、所述真人视频主播终端和除所述第一观众终端的其他观众终端。
  3. 如权利要求1所述的方法,还包括:
    获取来自所述真人视频主播终端的虚拟互动指令;以及
    将所述虚拟互动指令发送至所述虚拟人物主播终端和观众终端。
  4. 如权利要求1至3任一者所述的方法,还包括:
    响应于连麦的结束,停止将所述虚拟人物主播终端的虚拟人物形象指令发送至所述真人视频主播终端和真人视频主播直播间的观众终端,停止将所述真人视频主播终端的视频流发送至所述虚拟人物主播终端和虚拟人物主播直播间的观众终端。
  5. 如权利要求1至3任一者所述的方法,其中,所述虚拟人物形象指令包括静态元素指令和动态元素指令。
  6. 一种实现虚拟人物和真人视频混合连麦的方法,应用于虚拟人物主播终端,包括:
    生成虚拟人物形象指令,并将所述虚拟人物形象指令发送至服务器端;
    接收来自真人视频主播终端的视频流;以及
    根据所述视频流,显示真人视频主播的连麦形象。
  7. 如权利要求6所述的方法,还包括:
    接收来自观众终端的虚拟特效指令;
    根据所述虚拟特效指令调用预置的虚拟元素集中的虚拟元素显示虚拟特效。
  8. 如权利要求6或7所述的方法,还包括:
    接收来自所述真人视频主播终端的虚拟互动指令;
    根据所述虚拟互动指令调用预置的虚拟元素集中的虚拟元素显示互动特效。
  9. 如权利要求6或7所述的方法,还包括:
    进入所述真人视频主播的直播间;
    通过所述服务器端向所述真人视频主播终端发送连麦请求。
  10. 一种实现虚拟人物和真人视频混合连麦的方法,应用于真人视频主播终端,包括:
    生成直播视频流,并将所述直播视频流发送至服务器端;
    接收来自虚拟人物主播终端的虚拟人物形象指令;以及
    根据所述虚拟人物形象指令调用预置的虚拟元素集中的虚拟元素显示对应主播的虚拟人物形象。
  11. 如权利要求10所述的方法,还包括:
    接收来自观众终端的虚拟特效指令;
    根据所述虚拟特效指令调用预置的虚拟元素集中的虚拟元素显示虚拟特效。
  12. 如权利要求10或11所述的方法,还包括:
    向所述服务器端发送虚拟互动指令。
  13. 如权利要求10或11所述的方法,还包括:
    进入所述虚拟人物主播的直播间;
    通过所述服务器端向所述虚拟人物主播终端发送连麦请求。
  14. 一种实现虚拟人物和真人视频混合连麦的方法,应用于观众终端,包括:
    接收来自虚拟人物主播终端的虚拟人物形象指令;
    接收来自真人视频主播终端的直播视频流;以及
    根据所述虚拟人物形象指令调用预置的虚拟元素集中的虚拟元素显示所述虚拟人物主播的虚拟人物形象,并根据所述直播视频流显示真人视频主播的连麦形象。
  15. 如权利要求14所述的方法,还包括:
    生成第一虚拟特效指令,并将所述第一虚拟特效指令发送至服务器端;
    接收来自所述服务器端的第二虚拟特效指令;
    根据所述第二虚拟特效指令调用预置的虚拟元素集中的虚拟元素显示虚拟特效。
  16. 如权利要求14或15所述的方法,还包括:
    接收来自所述真人视频主播终端的虚拟互动指令;
    根据所述虚拟互动指令调用预置的虚拟元素集中的虚拟元素显示互动特效。
  17. 一种实现虚拟人物和真人视频混合连麦的系统,其包括服务器端、虚拟人物主播终端、真人视频主播终端和观众终端,其中,所述服务器端执行如权利要求1至5任一者所述的方法,所述虚拟人物主播终端执行如权利要求6至9任一者所述的方法,所述真人视频主播终端执行如权利要求10至13任一者所述的方法,所述观众终端执行如权利要求14至16任一者所述的方法。
PCT/CN2022/097401 2022-06-07 2022-06-07 实现虚拟人物和真人视频混合连麦的系统和方法 WO2023236045A1 (zh)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2022/097401 WO2023236045A1 (zh) 2022-06-07 2022-06-07 实现虚拟人物和真人视频混合连麦的系统和方法

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2022/097401 WO2023236045A1 (zh) 2022-06-07 2022-06-07 实现虚拟人物和真人视频混合连麦的系统和方法

Publications (1)

Publication Number Publication Date
WO2023236045A1 true WO2023236045A1 (zh) 2023-12-14

Family

ID=89117364

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/097401 WO2023236045A1 (zh) 2022-06-07 2022-06-07 实现虚拟人物和真人视频混合连麦的系统和方法

Country Status (1)

Country Link
WO (1) WO2023236045A1 (zh)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107438183A (zh) * 2017-07-26 2017-12-05 北京暴风魔镜科技有限公司 一种虚拟人物直播方法、装置及系统
CN109874021A (zh) * 2017-12-04 2019-06-11 腾讯科技(深圳)有限公司 直播互动方法、装置及系统
CN113965812A (zh) * 2021-12-21 2022-01-21 广州虎牙信息科技有限公司 直播方法、系统及直播设备
CN114374880A (zh) * 2021-12-23 2022-04-19 北京达佳互联信息技术有限公司 联合直播方法、装置、电子设备及计算机可读存储介质

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107438183A (zh) * 2017-07-26 2017-12-05 北京暴风魔镜科技有限公司 一种虚拟人物直播方法、装置及系统
CN109874021A (zh) * 2017-12-04 2019-06-11 腾讯科技(深圳)有限公司 直播互动方法、装置及系统
CN113965812A (zh) * 2021-12-21 2022-01-21 广州虎牙信息科技有限公司 直播方法、系统及直播设备
CN114374880A (zh) * 2021-12-23 2022-04-19 北京达佳互联信息技术有限公司 联合直播方法、装置、电子设备及计算机可读存储介质

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
WUHAN BROADCASTING AND TELEVISION STATION: "Broadcast room virtual anchor appearance, her name is Xiao Fang", BAIJIAHAO.BAIDU, 6 May 2022 (2022-05-06), XP093113138, Retrieved from the Internet <URL:https://baijiahao.baidu.com/s?id=1732057633165175388&wfr=spider&for=pc> [retrieved on 20231218] *

Similar Documents

Publication Publication Date Title
CN113395533B (zh) 虚拟礼物特效显示方法、装置、计算机设备及存储介质
KR101326739B1 (ko) 정보처리 시스템 및 정보처리 장치
CN106161219B (zh) 消息处理方法及装置
CN113457123B (zh) 基于云游戏的互动方法、装置、电子设备及可读存储介质
CN111314720A (zh) 直播连麦控制方法、装置、电子设备及计算机可读介质
WO2020010819A1 (zh) 基于直播间的数据交互方法、装置、终端和存储介质
WO2017101320A1 (zh) 一种实现双人同时视频直播的方法及移动客户端
CN109195003B (zh) 基于直播进行游戏的互动方法、系统、终端及装置
JP2023538958A (ja) 撮影方法、装置、電子機器及びコンピュータ可読記憶媒体
US10289193B2 (en) Use of virtual-reality systems to provide an immersive on-demand content experience
CN111064919A (zh) 一种vr远程会议方法及装置
US9398260B2 (en) Teleconference system, storage medium storing program for server apparatus, and storage medium storing program for terminal apparatus
CN114430494B (zh) 界面显示方法、装置、设备及存储介质
WO2019107274A1 (ja) 情報処理装置およびゲーム画像配信方法
CN106686463A (zh) 一种视频中的角色替换方法及装置
CN114201095A (zh) 直播界面的控制方法、装置、存储介质及电子设备
WO2023098011A1 (zh) 视频播放方法及电子设备
CN112055252A (zh) 多屏互动方法、装置、计算机可读介质及电子设备
CN112383832A (zh) 直播间互动方法、电子设备和计算机可读存储介质
JP6688378B1 (ja) コンテンツ配信システム、配信装置、受信装置及びプログラム
CN110730340A (zh) 基于镜头变换的虚拟观众席展示方法、系统及存储介质
CN116057896A (zh) 持续共存群组视频会议系统
WO2023236045A1 (zh) 实现虚拟人物和真人视频混合连麦的系统和方法
WO2019107275A1 (ja) 情報処理装置およびゲーム画像配信方法
CN114760520A (zh) 直播中小视频拍摄互动方法、装置、设备及存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22945193

Country of ref document: EP

Kind code of ref document: A1