WO2022089192A1 - 一种互动处理方法、装置、电子设备和存储介质 - Google Patents

一种互动处理方法、装置、电子设备和存储介质 Download PDF

Info

Publication number
WO2022089192A1
WO2022089192A1 PCT/CN2021/123202 CN2021123202W WO2022089192A1 WO 2022089192 A1 WO2022089192 A1 WO 2022089192A1 CN 2021123202 W CN2021123202 W CN 2021123202W WO 2022089192 A1 WO2022089192 A1 WO 2022089192A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
information
target
terminal
avatar
Prior art date
Application number
PCT/CN2021/123202
Other languages
English (en)
French (fr)
Inventor
顾宇
Original Assignee
北京有竹居网络技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京有竹居网络技术有限公司 filed Critical 北京有竹居网络技术有限公司
Publication of WO2022089192A1 publication Critical patent/WO2022089192A1/zh

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/451Execution arrangements for user interfaces
    • G06F9/452Remote windowing, e.g. X-Window System, desktop virtualisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/04817Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance using icons
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/0486Drag-and-drop
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • G06Q50/20Education
    • G06Q50/205Education administration or guidance

Definitions

  • the embodiments of the present disclosure relate to the technical field of information processing, and in particular, to an interaction processing method, apparatus, electronic device, and storage medium.
  • teaching methods have also undergone great changes.
  • the teaching method was that teachers and students were in the same classroom, and teachers and students taught face-to-face.
  • the current teacher can teach through the Internet, so that students can also learn at home.
  • the teacher's terminal can send the video data of the teacher's teaching to the student's terminal in real time, so that the students can watch the teacher's teaching video in real time through their terminal.
  • teachers may need to interact with individual students in the process of online teaching. For example, the teacher may remind a certain classmate to pay attention to the lecture on the spot. However, in a scenario with a large number of students, if the teacher interacts with multiple students one by one, the teaching efficiency will be greatly reduced.
  • the embodiments of the present disclosure provide an interactive processing method, apparatus, electronic device and storage medium, so as to improve teaching efficiency and teaching quality.
  • An embodiment of the present disclosure provides an interaction processing method, including:
  • the user interface including identification information for each of the one or more first users
  • the avatar is sent to the target user's terminal.
  • Embodiments of the present disclosure also provide an interactive processing device, including:
  • a display module for displaying a user interface, the user interface including identification information of each of the one or more first users;
  • an obtaining module configured to obtain the interactive instruction of the second user for the target user in the one or more first users
  • a generating module configured to generate an avatar according to the interactive content associated with the interactive instruction
  • a sending module configured to send the virtual image to the terminal of the target user.
  • Embodiments of the present disclosure also provide an electronic device, the electronic device comprising:
  • processors one or more processors
  • a storage device for storing one or more programs
  • the one or more processors When the one or more programs are executed by the one or more processors, the one or more processors implement the above-mentioned interactive processing method.
  • Embodiments of the present disclosure also provide a computer-readable storage medium on which a computer program is stored, and when the program is executed by a processor, implements the above-mentioned interactive processing method.
  • the interactive processing method provided by the embodiments of the present disclosure, by displaying a user interface, and displaying the identifier of each first user among one or more first users in the user interface information, so that the second user can input an interactive instruction for the target user among one or more first users according to the identification information of each first user, and can further generate an avatar according to the interactive content associated with the interactive instruction, and use the The avatar is sent to the terminal of the target user, so that the avatar can interact with the target user.
  • the interaction time between the second user and individual first users can be saved, thereby It can improve the efficiency of interaction.
  • the first user is a student and the second user is a teacher
  • the avatar is generated for individual students, and the interactive content between the teacher and individual students can be carried in the avatar, there is no need for the teacher to stop teaching in the middle. Therefore, interacting with individual students through virtual images can not only save the interaction time between teachers and individual students, but also ensure that other students’ normal listening is not affected. In large online classes with a large number of students, it can greatly improve teaching. Efficiency and teaching quality.
  • FIG. 1 is a flowchart of an interaction processing method in an embodiment of the present disclosure
  • FIG. 2 is a schematic diagram of an application scenario in an embodiment of the present disclosure
  • FIG. 3 is a schematic diagram of a user interface in an embodiment of the present disclosure.
  • FIG. 4 is a flowchart of another interaction processing method in an embodiment of the present disclosure.
  • FIG. 5 is a schematic diagram of another user interface in an embodiment of the present disclosure.
  • FIG. 6 is a schematic diagram of yet another user interface in an embodiment of the disclosure.
  • FIG. 7 is a flowchart of still another interaction processing method in an embodiment of the disclosure.
  • FIG. 8 is a schematic diagram of still another user interface in an embodiment of the disclosure.
  • FIG. 9 is a flowchart of still another interaction processing method in an embodiment of the present disclosure.
  • FIG. 10 is a schematic diagram of another user interface in an embodiment of the disclosure.
  • FIG. 11 is a schematic diagram of another user interface in an embodiment of the disclosure.
  • FIG. 13 is a schematic diagram of another user interface in an embodiment of the disclosure.
  • FIG. 14 is a schematic structural diagram of an interactive processing apparatus according to an embodiment of the disclosure.
  • FIG. 15 is a schematic structural diagram of an electronic device according to an embodiment of the disclosure.
  • the term “including” and variations thereof are open-ended inclusions, ie, "including but not limited to”.
  • the term “based on” is “based at least in part on.”
  • the term “one embodiment” means “at least one embodiment”; the term “another embodiment” means “at least one additional embodiment”; the term “some embodiments” means “at least some embodiments”. Relevant definitions of other terms will be given in the description below.
  • FIG. 1 is a flowchart of an interaction processing method in an embodiment of the present disclosure.
  • This embodiment is applicable to the case of performing interaction processing in a client.
  • the method may be executed by an interaction processing apparatus, and the apparatus may use software and/or
  • the device can be configured in electronic devices, such as terminals, specifically including but not limited to smart phones, PDAs, tablet computers, wearable devices with display screens, desktop computers, notebook computers, all-in-one computers, and smart homes. equipment, etc.
  • this embodiment may be applicable to the case of interactive processing in a server, the method may be executed by an interactive processing apparatus, the apparatus may be implemented in software and/or hardware, and the apparatus may be configured in an electronic device, such as a server .
  • the interaction processing method described in the embodiment of the present disclosure may be applicable to the application scenario shown in FIG. 2 .
  • the application scenario may include one or more first terminals, for example, a first terminal 21, a first terminal 22, and a first terminal 23, where each first terminal may be a terminal of a first user.
  • the first user may be, for example, a student. It can be understood that the number of first terminals is not limited to the three shown in FIG. 2 , and in other embodiments, there may be more first terminals, and the embodiment of the present disclosure takes three as an example for schematic illustration .
  • the embodiment of the present disclosure does not limit the product form of the first terminal, for example, it can be a smart phone, a handheld computer, a tablet computer, a wearable device with a display screen, a desktop computer, a notebook computer, an all-in-one computer, a smart home device, etc. .
  • the application scenario may further include one or more second terminals, for example, the second terminal 30, and the second terminal may be a terminal of a second user.
  • the second user may be a teacher, for example.
  • the number of second terminals is not limited to the one shown in FIG. 2 , and in other embodiments, there may be more second terminals, and the embodiment of the present disclosure takes one as an example for schematic illustration .
  • the embodiment of the present disclosure does not limit the product form of the second terminal, for example, it may be a smart phone, a handheld computer, a tablet computer, a wearable device with a display screen, a desktop computer, a notebook computer, an all-in-one computer, a smart home device, etc. .
  • the first user is not limited to students, and the second user is not limited to teachers.
  • the second user may also be a live broadcaster, and the first user may be a user watching the live broadcast. That is to say, the interaction processing method described in the embodiments of the present disclosure may be applicable to one-to-many or one-to-one interaction scenarios.
  • the second terminal 30 may communicate with the first terminal 21 , the first terminal 22 and the first terminal 23 respectively.
  • the second terminal 30 , the first terminal 21 , the first terminal 22 and the first terminal 23 can be connected to the server 40 , so that the second terminal 30 can communicate with the first terminal 21 , A terminal 22 communicates with the first terminal 23 .
  • the first terminal 21 , the first terminal 22 and the first terminal 23 may also communicate with each other through the server 40 .
  • server 40 may be a server cluster or one or more servers in a server cluster.
  • the server 40 may also be a cloud server.
  • the second terminal 30 , the first terminal 21 , the first terminal 22 and the first terminal 23 are respectively installed with the same application program (Application, APP for short), and the server 40 may provide the APP with Serve.
  • the interaction processing method described in this embodiment may be executed by the second terminal 30 or the server 40 as shown in FIG. 2 , and the second terminal 30 is taken as an example for schematic illustration below.
  • the method may specifically include:
  • S101 Display a user interface, wherein the user interface includes identification information of each of the one or more first users.
  • the second terminal 30 shown in FIG. 2 is a desktop computer, and the second terminal 30 includes a host and a display component.
  • the display component may be a display screen.
  • the second terminal 30 may display a user interface on the display screen, and in a scenario where the second user of the second terminal is a teacher, the user interface may be the user interface of the teacher.
  • the user interface may include identification information for each of the one or more first users.
  • the identification information of the first user includes at least one of the following: image information, name, identification number of the first user, and terminal identification of the first user.
  • the teacher's user interface may display identification information of one or more students who are taking a class.
  • the identification information of each student includes at least one of the student's image information, name, identification number, and a terminal identification of the student's first terminal.
  • the image information may be the student's avatar or other images.
  • the name can be the student's name or nickname.
  • the identification number may be, for example, the student's student number, account number, and the like.
  • the terminal identification of the first terminal may be a port number, a hardware identification number, and the like of the first terminal.
  • the display form of the identification information of the multiple students is not limited.
  • the identification information of the multiple students may be displayed in the user interface of the teacher in the form of a list.
  • Figure 3 is a schematic diagram of the teacher's user interface.
  • the user interface displays the name of student Zhang San and the avatar of Zhang San 31, the name of student Li Si and the avatar of Li Si 32, the name of student Wang Wu and the avatar of Wang Wu. Five's avatar 33.
  • the user interface shown in FIG. 3 is only a schematic illustration, and not specifically limited.
  • the user interface may also include other icons or buttons.
  • S102 Acquire an interaction instruction of the second user with respect to the target user among the one or more first users.
  • the user interface of the teacher may also display an electronic teaching material 34 used by the teacher in the teaching process, and the electronic teaching material 34 may be at least one of text data, video data, audio data, and image data.
  • the electronic textbook 34 and the identification information of multiple students may be displayed on the same display screen, or may be displayed on different display screens, which are not specifically limited here.
  • the second terminal 30 can send the electronic teaching material 34 to the first terminal 21, the first terminal 22 and the first terminal 23 in real time through the server 40, so that students can view the electronic teaching material 34 through the first terminal in real time.
  • the second terminal 30 may further include a photographing apparatus, or the second terminal 30 may be connected with the photographing apparatus.
  • the photographing device may be, for example, a camera, and the camera may be integrated into the display component of the second terminal 30 , or the camera may be connected to the host of the second terminal 30 .
  • the camera can collect the teacher's image or video information in real time.
  • the second terminal 30 may send the teacher's image or video information to the first terminal 21 , the first terminal 22 and the first terminal 23 in real time through the server 40 .
  • the teacher's image or video information is displayed in real time on the first terminal 21, the first terminal 22 and the first terminal 22.
  • students can view the teacher's image or video information in real time through the first terminal.
  • a teacher may need to interact with individual students in the process of teaching, such as reminding individual students to pay attention to lectures or asking individual students to answer questions.
  • An individual student here can be a single student or multiple students.
  • individual students can be recorded as target users here.
  • the teacher can input the interactive instruction for the individual student to the second terminal 30 .
  • the second terminal 30 may acquire the teacher's interactive instruction for the individual student.
  • This embodiment does not limit the input method of the interactive instruction.
  • the teacher can operate the avatar of a certain student in the user interface, and the operation can be, for example, clicking, double-clicking, long-pressing, voice control operation, or gesture control operations, etc.
  • the second terminal 30 may acquire the teacher's interactive instruction for the student according to the teacher's operation on the student's avatar.
  • the second terminal 30 may display preset interactive content, and further determine the interactive content selected by the teacher according to the teacher's operation on the interactive content, wherein the interactive content selected by the teacher may be as the interactive content associated with the interactive instruction.
  • the second terminal 30 can display an input box after acquiring the interactive instruction from the teacher for a certain student, and further receive the interactive content input by the teacher in the input box, and the interactive content input by the teacher in the input box can be used as the interactive instruction.
  • the interactive content may be at least one of text data, video data, audio data, and image data.
  • the second terminal 30 may generate an avatar according to the interactive content associated with the interactive instruction. It can be understood that the avatar is not limited to be generated on the side of the second terminal 30 , for example, the avatar can also be generated on the side of the server 40 .
  • S104 Send the virtual image to the terminal of the target user.
  • Zhang San is the target user that the teacher needs to interact with.
  • the avatar can also be sent to the first terminal of Zhang San, such as the first terminal 21 .
  • the first terminal 21 can display the avatar on the screen of the first terminal 21 , so that Zhang San can see the avatar in the first terminal 21 .
  • sending the avatar to the terminal of the target user may include: first adding a mark corresponding to the target user to the avatar; and then sending the avatar with the mark added to the terminal of the target user.
  • each port may correspond to a different label
  • the label corresponding to the port 1 of the first terminal 21 of student Zhang San is the language "Zhang San”
  • the mark corresponding to port 2 of the first terminal 22 of student Li Si is the language "Li Si”.
  • the second terminal 30 determines that the teacher interacts with Zhang San, the avatar "Zhang San” can be added to the avatar, and the avatar with the huashu "Zhang San” added can be sent to the first terminal 21 through port 1 . That is to say, the vocabulary of the avatar presented by the client corresponding to the first terminal of each student may be different. Among them, the words can be understood as the voice content spoken by the avatar.
  • the avatar received by the first terminal 21 through port 1 is that the facial expression of the avatar is moving, and the voice heard by Zhang San may be, for example, "Zhang San, please pay attention to the lecture". That is to say, in the process of playing the avatar on the first terminal, the playing time may be several seconds. During these few seconds, what the student hears or sees from the first terminal is the avatar talking, not the real one. The teacher is talking so that all the students are listening.
  • the avatar can actually be a short video, and the short video includes image information and audio information.
  • the image information can be the facial image of the avatar
  • the audio information can be the audio information corresponding to the interactive content initiated by the teacher.
  • the facial image of the virtual character may be the facial image of the second user of the second terminal 30 (ie, the real teacher), or the facial image of the virtual character may also be the facial image of other real people or other virtual characters.
  • the attributes of the audio information in the avatar such as timbre, pitch, or volume, may be the same as those of the real teacher, such as timbre, pitch, or volume.
  • the attributes of the audio information in the avatar may also be audio attributes of other real characters or other virtual characters.
  • the video information of the teacher collected by the shooting device of the second terminal 30 may be the real video of the teacher, and the real video includes the real image of the teacher and real audio information.
  • the server 40 distributes the video, the real video of the teacher needs to be distributed to each student, and the avatar for the individual student only needs to be sent to the first terminal of the individual student. That is, the teacher's real video is the information shared by each student, while the avatar is the student's personalized information.
  • the server 40 may send the real video of the teacher together with the avatar for the student to the first terminal of the student, or the server 40 may only send the virtual image for the student The image is sent to the first terminal of the student, and after the server 40 completes sending the virtual image, the real video of the teacher continues to be sent to the first terminal of the student.
  • the second user by displaying a user interface and displaying identification information of each first user among one or more first users in the user interface, the second user can The user's identification information is input to the interactive instruction for the target user among the one or more first users, and further, an avatar can be generated according to the interactive content associated with the interactive instruction, and the avatar can be sent to the terminal of the target user.
  • an avatar can be generated according to the interactive content associated with the interactive instruction, and the avatar can be sent to the terminal of the target user.
  • the first user is a student and the second user is a teacher
  • the avatar is generated for individual students, and the interactive content between the teacher and individual students can be carried in the avatar, there is no need for the teacher to stop teaching in the middle. Therefore, interacting with individual students through virtual images can not only save the interaction time between teachers and individual students, but also ensure that other students’ normal listening is not affected. In large online classes with a large number of students, it can greatly improve teaching. Efficiency and teaching quality.
  • FIG. 4 is a flowchart of another interaction processing method in an embodiment of the present disclosure.
  • the interaction processing method may further include the following steps as shown in FIG. 4 :
  • each of the first terminal 21 , the first terminal 22 and the first terminal 23 may include a photographing device, such as a camera.
  • each of the first terminal 21 , the first terminal 22 and the first terminal 23 may be respectively connected in communication with the camera.
  • the camera can collect images or video information of students in real time.
  • each first terminal may send the image or video information of the corresponding student to the second terminal 30 .
  • each first terminal sends the image or video information of the corresponding student to the second terminal 30 through the server 40.
  • the second terminal 30 receives the image or video information of each student, the image or video information of each student may be displayed on the teacher's user interface.
  • FIG. 5 is another schematic diagram of a teacher's user interface.
  • the user interface may include a plurality of video players, each of which is used to play video information of a student. For example, as shown in FIG. 5 , the user interface displays the name of the student Zhang San and the video information of Zhang San, the name of Li Si and the video information of Li Si, the name of Wang Wu and the video information of Wang Wu.
  • the video information of Li Si is played in the video player 51
  • the video information of Li Si is played in the video player 52
  • the video information of Wang Wu is played in the video player 53 .
  • the user interface shown in FIG. 5 is only a schematic illustration, and not specifically limited.
  • the user interface may also include other icons or buttons.
  • the teacher can operate the video information of the target user (ie, individual students) in the user interface, and this operation is recorded as the first operation, and the first operation can be, for example, clicking, double-clicking, long-pressing , voice control operation, or gesture control operation, etc.
  • the first operation may further include the teacher moving the mouse arrow over the video information.
  • the second terminal 30 may display the first operation interface associated with the target user in the user interface according to the first operation of the teacher.
  • the teacher can move the mouse to move the mouse arrow to Zhang San's video information.
  • the first operation interface 511 associated with Zhang San may be displayed in the user interface. Further, the second terminal 30 may acquire the teacher's interactive instruction for Zhang San according to the teacher's operation on the first operation interface 511 .
  • the second terminal 30 can also control the first operation interface 511 to hide and display the first operation interface 512 associated with Li Si.
  • the user interface further includes: a first operation interface respectively associated with the identification information of each first user.
  • the user interface may further include a first operation interface respectively associated with the identification information of each student.
  • the first operation interface 511 is associated with the identification information of Zhang San
  • the first operation interface 512 is associated with the identification information of Li Si
  • the first operation interface 513 is associated with the identification information of Wang Wu. That is to say, the first operation interface may not need to be dynamically displayed according to the first operation of the teacher on the video information, but may be presented in the user interface at the same time as the user interface is displayed.
  • the second terminal 30 may acquire the teacher's interactive instruction for Zhang San according to the teacher's operation on the first operation interface 511 .
  • the second terminal 30 generates an avatar according to the interactive content associated with the interactive instruction, and sends the avatar to Zhang San's terminal such as the first terminal 21 . That is, the teacher can initiate an avatar for a single student, and the avatar for a single student can be sent to the terminal of the single student through the port of the single student, so as to be presented in the client used by the single student.
  • the display manner of the first operation interface is not limited to the display manner of FIG. 5 or FIG. 6 , and may also include other display manners, which are only schematically illustrated here.
  • the user interface shown in FIG. 6 is also a schematic illustration, which is not specifically limited.
  • the relative positional relationship between the identification information of the same student and the first operation interface is not limited.
  • the embodiments of the present disclosure do not specifically limit the first operation interface.
  • the first operation interface may include icons, buttons, keys, input boxes, or preset text information.
  • the user interface further includes: a second operation interface, and the second operation interface is associated with identification information of some or all of the one or more first users, and the target users include some or all of the first users.
  • the teacher's user interface may also display a second operation interface, and the difference between the second operation interface and the above-mentioned first operation interface is that a first operation interface is associated with a student, But the second operating interface can be associated with multiple students.
  • the user interface displays the identification information of three students in total, and the second operation interface may be associated with the identification information of some or all of the three students, for example, the number of some students is greater than one.
  • the above-mentioned target users may be the part or all of the students.
  • the interaction processing method further includes the following steps as shown in FIG. 7 step:
  • the user interface includes a second operation interface 80 .
  • the teacher may perform an associated operation on the second operation interface 80 and the identification information of the multiple students.
  • Zhang San and Li Si are the students who want to initiate avatars.
  • the teacher can select Zhang San's identification information and Li Si's identification information, and drag the Zhang San's identification information and Li Si's identification information after the selection. Drag or drag to the second operation interface 80 or a preset area around the second operation interface 80 .
  • the second terminal 30 may determine the multi-selection, dragging or dragging operation as an association operation in which the teacher associates the second operation interface 80 , the identification information of Zhang San and the identification information of Li Si.
  • the association operation is not specifically limited here, and the multi-selection, dragging or dragging operation is only a schematic illustration.
  • the second terminal 30 may establish an association relationship between the second operation interface 80 , the identification information of Zhang San and the identification information of Li Si according to the association operation.
  • the second terminal 30 may display Zhang San's name and Li Si's name in a preset area around the second operation interface 80 . It can be understood that this embodiment does not limit the presentation form of the association relationship, and the presentation form shown in FIG. 8 is only a schematic illustration.
  • the teacher when the teacher needs to interact with all the students, he can select all the identification information of all the students, and further associate the identification information of all the students with the second operation interface 80.
  • the specific association operation and association relationship please refer to the aforementioned method, which will not be repeated here.
  • the second terminal 30 can obtain the teacher's interactive instructions for Zhang San and Li Si according to the operation of the teacher on the second operation interface 80, and respectively generate virtual images for Zhang San and Li Si according to the interactive instructions, and target Zhang San and Li Si respectively.
  • the avatar of San is sent to the first terminal 21 of Zhang San through the port corresponding to Zhang San
  • the avatar for Li Si is sent to the first terminal 22 of Li Si through the port corresponding to Li Si, so that the first terminal 21 can
  • the avatar for Zhang San is displayed on the client used by Zhang San
  • the first terminal 22 may display the avatar for Li Si on the client used by Li Si.
  • the avatar for Zhang San and the avatar for Li Si may be the same or different.
  • an avatar can be initiated for multiple students or all students, and the avatar can be sent to the first terminal of each student through the port of each student, and the avatar can be sent to the first terminal of each student through the port of each student.
  • the used client presents an avatar, and the avatar presented on each student's client may be different.
  • the second user by displaying second operation interfaces associated with multiple first users on the user interface, the second user can interact with multiple first users at the same time by operating the second operation interface. , which further improves the interaction efficiency.
  • the teacher can input interactive instructions for individual students by operating the identification information of individual students, or the teacher can also operate the above-mentioned first operation interface or second operation interface to Inputting interactive instructions for individual students, the following describes the situation in which the second terminal 30 obtains the interactive instructions through the first operation interface or the second operation interface with reference to specific embodiments.
  • acquiring the interactive instruction of the second user for the target user among the one or more first users may include the following steps as shown in FIG. 9 :
  • the second terminal 30 can detect the teacher's operation on the first operation interface or the second operation interface, for example, the teacher's operation on the first operation interface or the second operation interface
  • the operation can be recorded as the second operation.
  • the second operation is not limited to click, double-click, long-press, voice control operation, or gesture control operation.
  • one first operation interface is associated with one student
  • one second operation interface is associated with multiple students.
  • the second terminal 30 may acquire the teacher's interactive instruction for the target user (ie, individual student) according to the teacher's second operation on the first operation interface or the second operation interface. For example, as shown in FIG. 6 , when the second terminal 30 detects the teacher's second operation on the first operation interface 512 , it can obtain the teacher's interactive instruction for Li Si. For example, as shown in FIG. 8 , when the second terminal 30 detects the teacher's second operation on the second operation interface 80 , the teacher's interactive instructions for Zhang San and Li Si can be obtained.
  • each first operation interface may be associated with one or more preset text information
  • the second operation interface may be associated with one or more preset text information.
  • the preset text information may be commonly used communication content between the teacher and the students, for example, “You are excellent”, “Pay attention to listening to the lecture”, “Please look at the screen”, “Pay attention to your sitting posture” and so on.
  • acquiring the interactive instruction of the second user for the target user according to the second operation may include: first, according to the second operation, displaying one or more preset text information associated with the target user in the user interface ; and then detect the selection operation of the second user on the target text information in the one or more preset text information; and then obtain the interactive instruction of the second user for the target user according to the selection operation.
  • the second terminal 30 may display one or more presets associated with Li Si in the user interface Text information, here is a schematic illustration by taking a plurality of preset text information as an example.
  • the teacher clicks on the first operation interface 512 multiple preset text messages related to Li Si are displayed, such as "You are awesome”, “Pay attention to listening”, “Please look at the screen”, "Pay attention to sitting posture” Schematic diagram of the user interface.
  • FIG. 10 is only a schematic illustration, and not specifically limited. For example, the display position, color, brightness, etc.
  • the teacher can select one or more of the plurality of preset text information according to his interactive requirements, and the preset text information selected by the teacher can be recorded as the target text information. Further, the second terminal 30 can detect the teacher's selection operation on the target text information, and obtain the teacher's interactive instruction for Li Si according to the selection operation. It can be understood that, one or more preset text information associated with different students may be different or the same. Alternatively, the teacher can also set one or more preset text messages in the user interface according to the characteristics of each student in the user interface.
  • the interactive content associated with the interactive instruction may include target text information. For example, in the user interface shown in FIG.
  • the second terminal 30 may determine that the interactive content associated with the teacher's interactive instruction for Li Si includes "you Awesome". It can be understood that the interactive content may also include other content other than the target text information such as "you are awesome", for example, the interactive content may be "Li Si, you are awesome”. Further, the second terminal 30 may generate an avatar for Li Si according to the interactive content, and send the avatar to the first terminal 22 of Li Si through the server 40 .
  • the second terminal 30 may display the one associated with Zhang San and Li Si at the same time in the user interface. or multiple preset text messages. Further, the second terminal 30 may detect the teacher's selection operation of the target text information in the one or more preset text information, and obtain the teacher's interactive instructions for Zhang San and Li Si according to the selection operation.
  • the interactive content associated with the interactive instruction includes the target text information. Further, the second terminal 30 can respectively generate an avatar for Zhang San and an avatar for Li Si according to the interactive content, and add a mark corresponding to Zhang San in the avatar for Zhang San, such as the words "Zhang San".
  • the second terminal 30 can send the avatar with the language "Zhang San” added to the first terminal 21 of Zhang San through the server 40, and send the avatar with the language "Li Si” added to Li Si's avatar.
  • the first terminal 22 The first terminal 22 .
  • the first operation interface or the second operation interface associated with the target user may include one or more preset text information; then the first operation interface or the second operation interface associated with the target user by the second user is detected.
  • the second operation of the second operation interface may include: detecting a second operation of the second user on target text information in the one or more preset text information.
  • each first operation interface may include one or more preset text information. That is to say, one or more preset text information associated with each first operation interface is not displayed on the user interface after the teacher performs a second operation on the first operation interface as in FIG.
  • a preset area in or around the first operation interface is displayed. Therefore, when detecting the teacher's second operation on the first operation interface, the second terminal 30 can specifically detect the teacher's second operation on the target text information in the one or more preset text information included in the first operation interface.
  • the operation may be a selection operation on the target text information. For example, in FIG.
  • the second terminal 30 can determine according to the selection operation that the interactive content associated with the teacher's interactive instruction for Zhang San includes the target Text message "Pay attention to your sitting posture".
  • the interactive content may also include other content other than the target text information "Pay attention to sitting posture", for example, the interactive content may be "Little Zhang San, pay attention to sitting posture”.
  • the second terminal 30 may generate an avatar for Zhang San according to the interactive content, and send the avatar to Zhang San's first terminal 21 through the server 40 .
  • a plurality of preset text messages may be displayed in the area where the second operation interface 80 is located or in the preset area around the second operation interface 80, for example, "You are awesome”, “Attention to listening” , “Look at the screen”, “Pay attention to your sitting posture”, etc.
  • the teacher selects the target text information in the plurality of preset text information, for example, "Pay attention to sitting posture”
  • the second terminal 30 may determine that the interactive content associated with the teacher's interactive instructions for Zhang San and Li Si includes “Pay attention to sitting posture” ". Further, the second terminal 30 can respectively generate an avatar for Zhang San and an avatar for Li Si according to the interactive content, and add a mark corresponding to Zhang San in the avatar for Zhang San, such as the words "Zhang San”.
  • the second terminal 30 can send the avatar with the language "Zhang San” added to the first terminal 21 of Zhang San through the server 40, and send the avatar with the language "Li Si” added to Li Si's avatar.
  • the first terminal 22 The first terminal 22 .
  • the interactive processing method provided in this embodiment further includes: in response to a second user's modification instruction, addition instruction, or deletion instruction to the preset text information, updating the preset text information in the user interface.
  • the teacher can also modify, add or delete preset text information associated with the first operation interface or the second operation interface.
  • the second terminal 30 may update the preset text information associated with the first operation interface or the second operation interface in the teacher's user interface in response to the teacher's modification instruction, addition instruction, or deletion instruction to the preset text information.
  • the interaction processing method provided by the embodiment of the present disclosure is associated with one or more preset text information through the first operation interface and/or the second operation interface, so that when the second user interacts with the first user, the user only needs to
  • the preset text information may be selected as the interactive content or a part of the interactive content, which further improves the interaction efficiency between the second user and the first user.
  • the first operation interface and/or the second operation interface may be associated with one or more preset text information, wherein the preset text information is preset, and in some application scenarios, multiple preset text information There may be no preset text information that meets the teacher's needs in the text information. Therefore, in the following embodiments, the first operation interface and/or the second operation interface may also be associated with the input box.
  • acquiring the interactive instruction of the second user for the target user according to the second operation may include: first, according to the second operation, displaying the input associated with the first operation interface or the second operation interface in the user interface box; then detect the input operation of the second user inputting the target interaction information in the input box; and then obtain the interaction instruction of the second user for the target user according to the input operation.
  • the second terminal 30 may display an input box associated with the first operation interface 512 in the user interface.
  • FIG. 11 it is a schematic diagram of the user interface where the input box 110 is displayed after the teacher clicks the first operation interface 512 . It can be understood that the user interface shown in FIG. 11 is only a schematic illustration, and not specifically limited, for example, the relative positions of the first operation interface 512 and the input box 110 are not limited. In the user interface shown in FIG.
  • the teacher can input target interaction information in the input box 110 according to his interaction requirements, and the target interaction information can be at least one of text information, audio information, video information, and image information. .
  • the second terminal 30 can detect the input operation of the teacher inputting the target interaction information in the input box 110, and obtain the teacher's interaction instruction for Li Si according to the input operation.
  • the interactive content associated with the interactive instruction may include target interactive information.
  • the second terminal 30 can determine, according to the target interaction information input by the teacher in the input box 110 , that the interaction content associated with the teacher's interaction instruction for Li Si includes the target interaction information.
  • the interactive content may also include other content other than the target interactive information, such as Li Si's nickname or name.
  • the second terminal 30 may generate an avatar for Li Si according to the interactive content and send the avatar to the first terminal 22 of Li Si through the server 40 .
  • the second terminal 30 may display an input box in the user interface. Further, the second terminal 30 The input operation of the teacher entering the target interactive information in the input box can be detected, and the teacher's interactive instructions for Zhang San and Li Si can be obtained according to the input operation.
  • the interactive content associated with the interactive instruction includes the target interactive information.
  • the second terminal 30 can respectively generate an avatar for Zhang San and an avatar for Li Si according to the interactive content, and add a mark corresponding to Zhang San in the avatar for Zhang San, such as the words "Zhang San”. ”; add the mark corresponding to Li Si in the avatar for Li Si, such as the language “Li Si”. Further, the second terminal 30 can send the avatar with the language "Zhang San” added to the first terminal 21 of Zhang San through the server 40, and send the avatar with the language "Li Si” added to Li Si's avatar.
  • the first terminal 22 is an avatar for Zhang San and an avatar for Li Si according to the interactive content, and add a mark corresponding to Zhang San in the avatar for Zhang San, such as the words "
  • the first operation interface or the second operation interface may include an input box; then detecting the second operation of the second user for the first operation interface or the second operation interface associated with the target user may include: An input operation of the second user inputting the target interaction information in the input box is detected.
  • each first operation interface may include an input box. That is to say, the input box associated with each first operation interface is not displayed on the user interface after the teacher performs the second operation on the first operation interface as in FIG. At the same time, it is displayed in the first operation interface or a preset area around the first operation interface.
  • the teacher can input target interaction information in the input box.
  • the target interaction information input by the teacher in the input boxes included in the different first operation interfaces may be different. For example, in the input box included in the first operation interface associated with Zhang San, input "Attention to listening", and in the first operation associated with Li Si Enter “see the screen" in the input box included in the interface.
  • the second terminal 30 may specifically detect the teacher's input operation of inputting the target interaction information in the input box.
  • the interactive content associated with the interactive instruction may include target interactive information.
  • target interactive information such as "pay attention to sitting posture” in the input box included in the first operation interface associated with Zhang San
  • the second terminal 30 may determine that the interactive content associated with the teacher's interactive instruction for Zhang San includes "attention to sitting posture” sitting position".
  • the interactive content may also include other content than the target text information "Pay attention to sitting posture", for example, the interactive content may be "Little Zhang San, pay attention to sitting posture”.
  • the second terminal 30 may generate an avatar for Zhang San according to the interactive content and send the avatar to Zhang San's first terminal 21 through the server 40 .
  • the second operation interface 80 may include an input box.
  • the teacher inputs target interaction information in the input box, such as “pay attention to your sitting posture”
  • the second terminal 30 may determine that the teacher is targeting Zhang San and Li
  • the interactive content associated with the fourth interactive instruction includes "pay attention to sitting posture”.
  • the second terminal 30 can respectively generate an avatar for Zhang San and an avatar for Li Si according to the interactive content, and add a mark corresponding to Zhang San in the avatar for Zhang San, such as the words "Zhang San”. ”; add the mark corresponding to Li Si in the avatar for Li Si, such as the language “Li Si”.
  • the second terminal 30 can send the avatar with the language "Zhang San” added to the first terminal 21 of Zhang San through the server 40, and send the avatar with the language "Li Si” added to Li Si's avatar.
  • the first terminal 22 is a mark corresponding to Zhang San in the avatar for Zhang San, such as the words "Zhang San”.
  • the interaction processing method provided by the embodiment of the present disclosure is associated with the input box through the first operation interface and/or the second operation interface, so that when the second user interacts with the first user, he only needs to input interactive content or Only a part of the interactive content is required, which further improves the interaction efficiency between the second user and the first user, and also improves the flexibility of the interactive content.
  • "generating an avatar according to the interaction content associated with the interaction instruction” may include the following steps as shown in FIG. 12 :
  • This embodiment is illustratively described by taking the interactive content as text information as an example.
  • a teacher who is teaching a large class online needs to interact with one or more designated students without affecting the normal class of other students, he can trigger the second terminal 30 to create an avatar, and use the created virtual image.
  • the image interacts with the designated students, while other students still use their real image (ie, real video) for online teaching.
  • real image ie, real video
  • the duration corresponding to the text information can be determined.
  • the method of determining the duration can be implemented by using a pre-trained duration prediction model. For example, the text information "please pay attention to listening" is input into the duration prediction model, and the obtained duration is 3 seconds.
  • the text information can be divided into a plurality of subsequences, the duration of each subsequence is determined by a duration prediction model, and the duration of each subsequence is further added to obtain the duration of the text information.
  • the method for determining the duration of the text information is not limited to the duration prediction model, for example, it can also be implemented by looking up a table. This embodiment does not limit the method for determining the duration of text information.
  • the text information and its corresponding duration can be input into the pre-trained speech synthesis model to obtain the audio corresponding to the text information, for example, the text information "Please pay attention to listening" and the duration of 3 seconds are input into the speech synthesis model. Afterwards, the speech synthesis model can convert the text message "Please pay attention to listening" into audio according to the duration of 3 seconds.
  • the facial feature point sequence may include multiple groups of facial feature points; then obtaining the facial feature point sequence corresponding to the audio may include: first obtaining the audio frame sequence according to the audio; then inputting the audio frame sequence into the pre-trained feature point prediction.
  • the model obtains a facial feature point sequence, wherein the facial feature point sequence includes multiple sets of facial feature points, and the feature point prediction model is used to determine a set of facial feature points corresponding to each audio frame.
  • an audio frame sequence can be acquired according to the audio.
  • the audio "please listen carefully" with a duration of 3 seconds includes 100 audio frames.
  • the 100 audio frames can be input into the pre-trained feature
  • the point prediction model obtains a sequence of facial feature points, which includes multiple groups of facial feature points, and the number of groups of facial feature points output by the feature point prediction model is the same as the number of audio frames, that is, each audio frame.
  • the number of feature points in each group of facial feature points may not be specifically limited, for example, it may be 68.
  • the 100 groups of facial feature points are arranged in sequence according to the frame sequence to constitute a facial feature point sequence.
  • the method before inputting the audio frame sequence into the pre-trained feature point prediction model and acquiring the facial feature point sequence, the method further includes: first acquiring a media file, wherein the media file includes sample audio and synchronously played facial video; Extract the sample audio frame from the sample audio, and extract the sample video frame synchronized with the sample audio frame from the face video; then extract the sample facial feature points included in the sample video frame; and then according to the sample audio frame and the sample facial feature point pair The feature point prediction model is trained.
  • a set of facial feature points to ensure the accuracy of facial feature point acquisition.
  • the way to obtain the sample is to obtain a media file, which can be a screen recording when the teacher conducts an online class or before the online class.
  • the media file includes sample audio and synchronously played face video. Since the sample audio and face video are played synchronously, sample audio frames are extracted from the sample audio, and sample video that is synchronized with the sample audio frames is extracted from the face video. After the frame, the frame numbers of the sample audio frames and the sample video frames are the same.
  • the number of sample video frames is also 50.
  • extract the sample facial feature points included in the first sample video frame One sample audio frame and the sample facial feature points included in the first sample video frame are taken as a group of samples, and correspondingly, 50 groups of samples can be sequentially extracted.
  • the feature point prediction model is trained according to the 50 sets of samples, when it is determined that the training error reaches a preset threshold, it is determined that the training is completed, and the trained feature point prediction model is used to predict subsequent facial feature points.
  • acquiring the video according to the sequence of facial feature points may include: first acquiring the facial standard template corresponding to the target user; then adding each group of facial feature points to the facial standard template to obtain each frame of facial image; then adding each frame of The facial images are sequentially concatenated in frame order to obtain the video.
  • the face standard template in this embodiment may be a frontal photo of the teacher's face randomly captured during the teacher's online teaching process.
  • each group of facial feature points in the above-mentioned sequence of facial feature points is added to the face standard template to obtain each frame of facial image. It can be understood that since each group of facial feature points in the facial feature point sequence may be different, different facial images can be obtained after adding different facial feature points to the face standard template.
  • each frame of facial image also has May differ from this face standard template.
  • the audio "please listen carefully" includes 100 audio frames
  • 100 audio frames correspond to 100 sets of facial feature points
  • 100 frames of facial images can be acquired, and the acquired 100 frames of facial images can be sorted in frame order Connect in turn to get the video.
  • the facial standard templates used when generating avatars for different students may be different.
  • the interaction processing method provided in this embodiment further includes: in response to a first setting instruction by the second user for the first user, displaying a plurality of preset facial standard templates; For the selection operation of the target template among the plurality of preset face standard templates, the target template is set as the face standard template corresponding to the first user.
  • the teacher can set the facial standard template corresponding to a certain student or several students in his user interface, and the second terminal 30 can respond to the teacher's first setting instruction for a certain student or certain students, A plurality of preset face standard templates are displayed in the user interface. Further, according to the selection operation of the target template in the plurality of preset face standard templates by the teacher, the target template can be set as a certain student or a certain number of face templates. The face standard template corresponding to the student. The teacher may randomly select a face standard template from the plurality of preset face standard templates as the target template.
  • the interaction processing method provided in this embodiment further includes: receiving a second setting instruction sent by the terminal of the first user, and the second setting instruction is used to set the facial standard corresponding to the first user template.
  • the first terminal may display multiple preset facial standard templates on the student interface, and students may select one they like from the multiple preset facial standard templates as the facial standard template used by their corresponding avatars .
  • the first terminal may send the second setting instruction for setting the face standard template by the student to the second terminal 30, so that the second terminal 30 uses the face set by the student himself when generating the avatar for the student Standard template.
  • the video generated in S1203 may include 100 frames of facial images, and it is necessary to further combine the video and audio "please listen carefully” to obtain an avatar.
  • the audio "please listen to the lecture” includes 100 audio frames
  • the video generated in S1203 includes 100 frames of facial images, so the 100 audio frames and the 100 frames of facial images can be processed according to the same time axis.
  • the combination of images for example, combine the first audio frame of the 100 audio frames with the first facial image of the 100 facial images, and so on until the 100th audio frame of the 100
  • the audio frames are combined with the 100th frame of facial images in the 100 frames of facial images to obtain an avatar, so that the sound and image of the avatar presented by the first terminal are matched and aligned in time sequence.
  • the expressions, actions, etc. of the avatar for each student may be different. That is to say, the expressions, actions, etc. of the avatar presented by the first terminal of each student may be different.
  • the avatar can be displayed on the first terminal of each student among the multiple students or all students at the same time, but the avatar displayed on the first terminal of each student The expressions, movements, etc. can be different.
  • the interaction processing method provided in this embodiment further includes: according to the type information of the interactive content, determining expression information and/or action information corresponding to the type information; added to the avatar.
  • preset text information displayed in the teacher's user interface is divided by type. As shown in FIG. 13 , when the second terminal 30 detects that the teacher operates the first operation interface 512, several types of preset text information are displayed in the user interface. For example, the preset text information of the praise category includes “you Awesome”, the preset text messages for reminders include “pay attention to listening”, “please look at the screen”, and "pay attention to sitting posture”.
  • the second terminal 30 can determine the type information corresponding to the preset text information, and determine the expression information and/or action information corresponding to the type information, in the process of generating the avatar Or after the avatar is generated, the expression information and/or action information is added to the avatar, so that the content spoken by the avatar matches the expression, action, etc. of the avatar. For example, when the avatar says, "Zhang San, you are awesome," the avatar can display a thumbs-up gesture. For another example, when the avatar says "pay attention to the lecture", the avatar can display a slightly angry expression in coordination.
  • the type information corresponding to the preset text information, and the expression information and/or action information corresponding to the type information may be pre-stored in the second terminal 30 or the server 40 .
  • the user interface shown in FIG. 13 is only a schematic illustration, and is not specifically limited.
  • type information such as a "compliment” icon and a "reminder” icon, may be displayed in the user interface.
  • the teacher can click on the "compliment” icon or the "reminder” icon, and after clicking, the second terminal 30 further displays the preset text information of the praise category or the preset text information of the reminder category.
  • the interaction processing method provided in this embodiment further includes: performing emotion analysis on the interaction content to obtain emotion information; and adding expression information and/or action information corresponding to the emotion information to the avatar .
  • the teacher randomly selects a preset text message among "You are awesome”, “Attention to listening”, “Please look at the screen”, “Attention to sitting posture”, or as shown in Figure 11, the teacher is typing Target interaction information is entered in box 110 .
  • the second terminal 30 can perform sentiment analysis on the preset text information selected by the teacher or the input target interaction information to obtain the sentiment information, and in the process of generating the avatar or after the avatar is generated, the sentiment information corresponds to the sentiment information.
  • the expression information and/or action information of the avatar is added to the avatar, so that the content spoken by the avatar matches the expression, action, etc. of the avatar.
  • S1201-S1204 can be executed by the second terminal 30, or can also be executed by the server 40, for example, the second terminal 30 can send the interactive content set by the teacher to the server 40, and the server 40 S1201-S1204 may be further performed to obtain an avatar.
  • the interaction processing method according to the type information of the interactive content, determine the expression information and/or the action information corresponding to the type of information; and add the expression information and/or the action information to the avatar; Or perform emotional analysis on the interactive content to obtain emotional information; and add the expression information and/or action information corresponding to the emotional information to the avatar, so that the content of the avatar's speech can be matched with the avatar's expression, action, etc. are matched, so that the vividness of the avatar can be improved.
  • the avatars corresponding to different students can be made different, which further improves the flexibility of the avatars.
  • FIG. 14 is a schematic structural diagram of an interaction processing apparatus according to an embodiment of the disclosure.
  • the interaction processing apparatus provided by the embodiment of the present disclosure may be configured in a client or may be configured in a server, and the interaction processing apparatus 140 specifically includes:
  • a display module 1401, configured to display a user interface, where the user interface includes identification information of each of the one or more first users;
  • an obtaining module 1402 configured to obtain the interactive instruction of the second user for the target user in the one or more first users
  • a generating module 1403, configured to generate an avatar according to the interactive content associated with the interactive instruction
  • the sending module 1404 is configured to send the virtual image to the terminal of the target user.
  • the identification information of the first user includes at least one of the following: image information, name, identification number of the first user, and terminal identification of the first user.
  • the display module 1401 is further configured to: display the video information of each first user in the user interface; The first operation interface associated with the target user.
  • the user interface further includes: a first operation interface respectively associated with the identification information of each first user.
  • the user interface further includes: a second operation interface, the second operation interface is associated with the identification information of some or all of the one or more first users, and the target users include some or all of the first users.
  • the interaction processing apparatus 140 further includes: a detection module 1405 and an association relationship establishment module 1406, wherein the detection module 1405 is used for the acquisition module 1402 to acquire the second user's information about the target user among the one or more first users.
  • the association relationship establishing module 1406 is configured to establish an association according to the association The operation establishes an association relationship between the second operation interface and the identification information of some or all of the first users.
  • the detection module 1405 is further configured to: detect the second operation performed by the second user on the first operation interface or the second operation interface associated with the target user; the acquisition module 1402 is further configured to: according to the second operation The operation acquires an interaction instruction of the second user with respect to the target user.
  • the display module 1401 is further configured to: display one or more preset text information associated with the target user in the user interface according to the second operation;
  • the detection module 1405 is further configured to: detect all selection operation of the second user on the target text information in the one or more preset text information;
  • the obtaining module 1402 is further configured to: obtain the interactive instruction of the second user for the target user according to the selection operation.
  • the first operation interface or the second operation interface associated with the target user includes one or more preset text information; the detection module 1405 is specifically configured to: detect that the second user has targeted the one or more preset text information. Set the second operation of the target text information in the text information.
  • the interactive content associated with the interactive instruction includes the target text information.
  • the display module 1401 is further configured to: display an input box associated with the first operation interface or the second operation interface in the user interface according to the second operation; the detection module 1405 is further configured to : detect the input operation of the second user inputting target interaction information in the input box; the acquiring module 1402 is further configured to: acquire the interaction instruction of the second user for the target user according to the input operation.
  • the first operation interface or the second operation interface includes an input box; the detection module 1405 is specifically configured to: detect an input operation of the second user inputting target interaction information in the input box.
  • the interactive content associated with the interactive instruction includes the target interactive information.
  • the interaction processing device 140 further includes: an update module 1407, configured to update all the information in the user interface in response to the second user's modification instruction, addition instruction or deletion instruction to the preset text information. Describe the preset text information.
  • the obtaining module 1402 is further configured to: obtain the audio corresponding to the interactive content; obtain the facial feature point sequence corresponding to the audio; obtain the video according to the facial feature point sequence; the generating module 1403 is specifically used for: The audio and the video are combined to obtain an avatar.
  • the interaction processing device 140 further includes: a determining module 1408 and an adding module 1409, wherein the determining module 1408 is configured to determine the expression information and/or action corresponding to the type information according to the type information of the interactive content information; the adding module 1409 is configured to add the expression information and/or the action information to the avatar.
  • the determining module 1408 is further configured to perform sentiment analysis on the interactive content to obtain sentiment information; the adding module 1409 is further configured to: add the expression information and/or action information corresponding to the sentiment information to the avatar middle.
  • the sequence of facial feature points includes multiple groups of facial feature points; when acquiring a video according to the sequence of facial feature points, the acquisition module 1402 is specifically configured to: acquire a standard face template corresponding to the target user; A group of facial feature points are added to the facial standard template, and each frame of facial image is obtained; the facial images of each frame are connected in sequence in a frame order to obtain a video.
  • the interaction processing device 140 further includes: a setting module 1410; the display module 1401 is further configured to: display a plurality of preset facial standard templates in response to the first setting instruction of the first user by the second user;
  • the setting module 1410 is configured to set the target template as the face standard template corresponding to the first user according to the second user's selection operation on the target template from the plurality of preset face standard templates.
  • the interaction processing device 140 further includes: a receiving module 1411, configured to receive a second setting instruction sent by the terminal of the first user, where the second setting instruction is used to set the face corresponding to the first user Standard template.
  • a receiving module 1411 configured to receive a second setting instruction sent by the terminal of the first user, where the second setting instruction is used to set the face corresponding to the first user Standard template.
  • the adding module 1409 is further configured to add the mark corresponding to the target user to the avatar; the sending module 1404 is specifically configured to: send the avatar with the mark added to the target user's avatar. terminal.
  • the interactive processing apparatus provided by the embodiment of the present disclosure can execute the steps executed by the client or the server in the interactive processing method provided by the method embodiment of the present disclosure, and the execution steps and beneficial effects are not repeated here.
  • FIG. 15 is a schematic structural diagram of an electronic device according to an embodiment of the disclosure. Referring specifically to FIG. 15 below, it shows a schematic structural diagram of an electronic device 1500 suitable for implementing an embodiment of the present disclosure.
  • the electronic device 1500 in the embodiment of the present disclosure may include, but is not limited to, such as a mobile phone, a notebook computer, a digital broadcast receiver, a PDA (personal digital assistant), a PAD (tablet computer), a PMP (portable multimedia player), a vehicle-mounted terminal ( For example, mobile terminals such as car navigation terminals) and the like, and stationary terminals such as digital TVs, desktop computers, and the like.
  • the electronic device shown in FIG. 15 is only an example, and should not impose any limitation on the function and scope of use of the embodiments of the present disclosure.
  • electronic device 1500 may include processing means (eg, central processing unit, graphics processing unit, etc.) 1501 that may be loaded into random access according to a program stored in read only memory (ROM) 1502 or from storage means 1508
  • the program in the memory (RAM) 1503 executes various appropriate actions and processes to implement the interactive processing method of the embodiment described in the present disclosure.
  • various programs and data necessary for the operation of the electronic device 1500 are also stored.
  • the processing device 1501, the ROM 1502, and the RAM 1503 are connected to each other through a bus 1504.
  • An input/output (I/O) interface 1505 is also connected to bus 1504 .
  • I/O interface 1505 the following devices can be connected to the I/O interface 1505: input devices 1506 including, for example, a touch screen, touchpad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, etc.; including, for example, a liquid crystal display (LCD), speakers, vibration An output device 1507 of a computer, etc.; a storage device 1508 including, for example, a magnetic tape, a hard disk, etc.; and a communication device 1509. Communication means 1509 may allow electronic device 1500 to communicate wirelessly or by wire with other devices to exchange data. While FIG. 15 shows an electronic device 1500 having various means, it should be understood that not all of the illustrated means are required to be implemented or provided. More or fewer devices may alternatively be implemented or provided.
  • input devices 1506 including, for example, a touch screen, touchpad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, etc.
  • LCD liquid crystal display
  • speakers vibration
  • embodiments of the present disclosure include a computer program product comprising a computer program carried on a non-transitory computer readable medium, the computer program containing program code for performing the method illustrated in the flowchart.
  • the computer program may be downloaded and installed from the network via the communication device 1509, or from the storage device 1508, or from the ROM 1502.
  • the processing apparatus 1501 the above-mentioned functions defined in the methods of the embodiments of the present disclosure are executed.
  • the computer-readable medium mentioned above in the present disclosure may be a computer-readable signal medium or a computer-readable storage medium, or any combination of the above two.
  • the computer-readable storage medium can be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus or device, or a combination of any of the above. More specific examples of computer readable storage media may include, but are not limited to, electrical connections with one or more wires, portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable Programmable read only memory (EPROM or flash memory), fiber optics, portable compact disk read only memory (CD-ROM), optical storage devices, magnetic storage devices, or any suitable combination of the foregoing.
  • a computer-readable storage medium can be any tangible medium that contains or stores a program that can be used by or in conjunction with an instruction execution system, apparatus, or device.
  • a computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave with computer-readable program code embodied thereon. Such propagated data signals may take a variety of forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • a computer-readable signal medium can also be any computer-readable medium other than a computer-readable storage medium that can transmit, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device .
  • Program code embodied on a computer readable medium may be transmitted using any suitable medium including, but not limited to, electrical wire, optical fiber cable, RF (radio frequency), etc., or any suitable combination of the foregoing.
  • clients and servers can communicate using any currently known or future developed network protocols such as HTTP (HyperText Transfer Protocol), and can communicate with digital data in any form or medium.
  • Communication eg, a communication network
  • Examples of communication networks include local area networks (“LAN”), wide area networks (“WAN”), the Internet (eg, the Internet), and peer-to-peer networks (eg, ad hoc peer-to-peer networks), as well as any currently known or future development network of.
  • the above-mentioned computer-readable medium may be included in the above-mentioned electronic device; or may exist alone without being assembled into the electronic device.
  • the above-mentioned computer-readable medium carries one or more programs, and when the above-mentioned one or more programs are executed by the electronic device, the electronic device:
  • the user interface including identification information for each of the one or more first users
  • the avatar is sent to the target user's terminal.
  • the electronic device may also perform other steps described in the above embodiments.
  • Computer program code for performing operations of the present disclosure may be written in one or more programming languages, including but not limited to object-oriented programming languages—such as Java, Smalltalk, C++, and This includes conventional procedural programming languages - such as the "C" language or similar programming languages.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer, or entirely on the remote computer or server.
  • the remote computer may be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or may be connected to an external computer (eg, using an Internet service provider through Internet connection).
  • LAN local area network
  • WAN wide area network
  • each block in the flowchart or block diagrams may represent a module, segment, or portion of code that contains one or more logical functions for implementing the specified functions executable instructions.
  • the functions noted in the blocks may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
  • each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented in dedicated hardware-based systems that perform the specified functions or operations , or can be implemented in a combination of dedicated hardware and computer instructions.
  • the units involved in the embodiments of the present disclosure may be implemented in a software manner, and may also be implemented in a hardware manner. Among them, the name of the unit does not constitute a limitation of the unit itself under certain circumstances.
  • exemplary types of hardware logic components include: Field Programmable Gate Arrays (FPGAs), Application Specific Integrated Circuits (ASICs), Application Specific Standard Products (ASSPs), Systems on Chips (SOCs), Complex Programmable Logical Devices (CPLDs) and more.
  • FPGAs Field Programmable Gate Arrays
  • ASICs Application Specific Integrated Circuits
  • ASSPs Application Specific Standard Products
  • SOCs Systems on Chips
  • CPLDs Complex Programmable Logical Devices
  • a machine-readable medium may be a tangible medium that may contain or store a program for use by or in connection with the instruction execution system, apparatus or device.
  • the machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium.
  • Machine-readable media may include, but are not limited to, electronic, magnetic, optical, electromagnetic, infrared, or semiconductor systems, devices, or devices, or any suitable combination of the foregoing.
  • machine-readable storage media would include one or more wire-based electrical connections, portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM or flash memory), fiber optics, compact disk read only memory (CD-ROM), optical storage, magnetic storage, or any suitable combination of the foregoing.
  • RAM random access memory
  • ROM read only memory
  • EPROM or flash memory erasable programmable read only memory
  • CD-ROM compact disk read only memory
  • magnetic storage or any suitable combination of the foregoing.
  • the present disclosure provides an interaction processing method, including:
  • the user interface including identification information for each of the one or more first users
  • the avatar is sent to the target user's terminal.
  • the identification information of the first user includes at least one of the following:
  • the image information, name, identification number of the first user, and the terminal identification of the first user are the image information, name, identification number of the first user, and the terminal identification of the first user.
  • the method further includes:
  • a first operation interface associated with the target user is displayed.
  • the user interface further includes: a first operation interface respectively associated with the identification information of each first user.
  • the user interface further includes: a second operation interface, the second operation interface is connected with the one or more first users
  • the target users include the part or all of the first users.
  • the method before the acquiring the interaction instruction of the second user for the target user in the one or more first users, the method further include:
  • An association relationship between the second operation interface and the identification information of some or all of the first users is established according to the association operation.
  • the acquiring the interaction instruction of the second user for the target user among the one or more first users includes:
  • the interaction instruction of the second user with respect to the target user is acquired according to the second operation.
  • the acquiring, according to the second operation, the interaction instruction of the second user for the target user includes:
  • the interaction instruction of the second user with respect to the target user is acquired according to the selection operation.
  • the first operation interface or the second operation interface associated with the target user includes one or more preset text information
  • the detecting the second operation of the second user on the first operation interface or the second operation interface associated with the target user includes:
  • a second operation of the second user on target text information in the one or more preset text information is detected.
  • the interactive content associated with the interactive instruction includes the target text information.
  • the acquiring, according to the second operation, the interaction instruction of the second user for the target user includes:
  • the interactive instruction of the second user for the target user is acquired according to the input operation.
  • the first operation interface or the second operation interface includes an input box
  • the detecting the second operation of the second user on the first operation interface or the second operation interface associated with the target user includes:
  • An input operation of the second user inputting target interaction information in the input box is detected.
  • the interaction content associated with the interaction instruction includes the target interaction information.
  • the method further includes:
  • the preset text information in the user interface is updated in response to a modification instruction, an addition instruction, or a deletion instruction of the preset text information by the second user.
  • generating an avatar according to the interactive content associated with the interactive instruction includes:
  • the audio and the video are combined to obtain an avatar.
  • the method further includes:
  • the type information of the interactive content determine the expression information and/or action information corresponding to the type information
  • the facial expression information and/or the action information are added to the avatar.
  • the method further includes:
  • the expression information and/or action information corresponding to the emotion information is added to the avatar.
  • the facial feature point sequence includes multiple groups of facial feature points
  • Obtain a video according to the facial feature point sequence including:
  • the method further includes:
  • the target template is set as the face standard template corresponding to the first user.
  • the method further includes:
  • a second setting instruction sent by the terminal of the first user is received, where the second setting instruction is used to set a face standard template corresponding to the first user.
  • sending the avatar to the terminal of the target user includes:
  • the present disclosure provides an interaction processing apparatus, including:
  • a display module for displaying a user interface, the user interface including identification information of each of the one or more first users;
  • an obtaining module configured to obtain the interactive instruction of the second user for the target user in the one or more first users
  • a generating module configured to generate an avatar according to the interactive content associated with the interactive instruction
  • a sending module configured to send the virtual image to the terminal of the target user.
  • the identification information of the first user includes at least one of the following: image information, name, identification number, name of the first user, Describe the terminal identifier of the first user.
  • the display module is further configured to: display the video information of each first user in the user interface;
  • the first operation of the target user's video information displays a first operation interface associated with the target user.
  • the user interface further includes: a first operation interface respectively associated with the identification information of each first user.
  • the user interface further includes: a second operation interface, the second operation interface is connected to the one or more first users
  • the target users include the part or all of the first users.
  • the interactive processing apparatus further includes: a detection module and an association relationship establishment module, wherein the detection module is used for the acquisition module to acquire the second user Before the interactive instruction of the target user among the one or more first users, detecting the associated operation of the second user on the second operation interface and the identification information of some or all of the first users in the user interface ; an association relationship establishing module is configured to establish an association relationship between the second operation interface and the identification information of some or all of the first users according to the association operation.
  • the detection module is further configured to: detect whether the second user is associated with the first operation interface or the second operation interface for the target user.
  • the second operation the obtaining module is further configured to: obtain the interactive instruction of the second user for the target user according to the second operation.
  • the display module is further configured to: display one or more associated with the target user in the user interface according to the second operation. a plurality of preset text information; the detection module is further used for: detecting the selection operation of the second user on the target text information in the one or more preset text information; the obtaining module is further used for: obtaining according to the selection operation An interactive instruction of the second user for the target user.
  • the first operation interface or the second operation interface associated with the target user includes one or more preset text information; the detection module specifically uses In: detecting a second operation of the second user on target text information in the one or more preset text information.
  • the interactive content associated with the interactive instruction includes the target text information.
  • the display module is further configured to: display in the user interface, according to the second operation, the interface related to the first operation or the all the input box associated with the second operation interface; the detection module is further configured to: detect the input operation of the second user inputting the target interaction information in the input box; the acquisition module is further configured to: obtain the target interaction information according to the input operation An interactive instruction of the second user for the target user.
  • the first operation interface or the second operation interface includes an input box; the detection module is specifically configured to: detect the second user Input operation of inputting target interaction information in the input box.
  • the interactive content associated with the interactive instruction includes the target interactive information.
  • the interactive processing apparatus further includes: an update module, configured to respond to a modification instruction of the preset text information by the second user , adding an instruction or deleting an instruction, and updating the preset text information in the user interface.
  • the acquiring module is further configured to: acquire audio corresponding to the interactive content; acquire a facial feature point sequence corresponding to the audio; The facial feature point sequence obtains a video; the generating module is specifically configured to: combine the audio and the video to obtain a virtual image.
  • the interactive processing apparatus further includes: a determining module and an adding module, wherein the determining module is configured to determine, according to the type information of the interactive content, Expression information and/or action information corresponding to the type information; the adding module is configured to add the expression information and/or the action information to the avatar.
  • the determining module is further configured to perform sentiment analysis on the interactive content to obtain sentiment information; the adding module is further configured to: convert the sentiment information Corresponding expression information and/or action information is added to the avatar.
  • the facial feature point sequence includes multiple groups of facial feature points; when acquiring the video according to the facial feature point sequence, the acquisition module specifically uses In: obtaining the facial standard template corresponding to the target user; adding each group of facial feature points to the facial standard template to obtain each frame of facial image; connecting the each frame of facial images in frame order to obtain video.
  • the interactive processing apparatus further includes: a setting module; the display module is further configured to: in response to the second user's first user's first a setting instruction, displaying a plurality of preset face standard templates; the setting module is configured to set the target template to the target template according to the second user's selection operation on the target template in the plurality of preset face standard templates Describe the facial standard template corresponding to the first user.
  • the interactive processing apparatus further includes: a receiving module, configured to receive a second setting instruction sent by the terminal of the first user, the The second setting instruction is used to set the facial standard template corresponding to the first user.
  • the adding module is further configured to add the mark corresponding to the target user to the avatar; the sending module is specifically configured to: add the added The marked avatar is sent to the target user's terminal.
  • the present disclosure provides an electronic device, comprising:
  • processors one or more processors
  • memory for storing one or more programs
  • the one or more processors When the one or more programs are executed by the one or more processors, the one or more processors implement the interactive processing method as provided in any one of the present disclosure.
  • the present disclosure provides a computer-readable storage medium on which a computer program is stored, and when the program is executed by a processor, implements any one of the interactive processing provided by the present disclosure method.

Abstract

一种互动处理方法、装置、电子设备和存储介质。该方法通过显示用户界面,并在该用户界面中显示一个或多个第一用户中每个第一用户的标识信息(S101),从而使得第二用户可以根据每个第一用户的标识信息输入针对一个或多个第一用户中目标用户的互动指令(S102),进一步可以根据该互动指令所关联的互动内容生成虚拟形象(S103),并将该虚拟形象发送到该目标用户的终端(S104),从而可以通过该虚拟形象与该目标用户进行互动,在第一用户数量较多的情况下,可以节省第二用户与个别第一用户的互动时间,进而可以提高互动效率。

Description

一种互动处理方法、装置、电子设备和存储介质
本申请要求于2020年10月28日提交的申请号为202011175034.6申请名称为“一种互动处理方法、装置、电子设备和存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本公开实施例涉及信息处理技术领域,尤其涉及一种互动处理方法、装置、电子设备和存储介质。
背景技术
随着信息技术的发展,教学方式也发生了很大的变化。例如,以往的教学方式是老师与学生位于同一教室,老师与学生面对面进行教学。而当前老师可以通过网络进行教学,使得学生在家里也可以进行学习。
老师在进行网络教学的过程中,老师的终端可以将老师教学的视频数据实时的发送到学生的终端,从而使得学生可以通过其终端实时观看老师的教学视频。
由于学生的个体差异,使得老师在进行网络教学的过程中可能需要与个别学生进行互动,例如,老师可能会当场提醒某位同学注意听讲等。但是,在学生数量较多的场景下,如果老师逐一与多位学生互动,将大大的降低教学效率。
发明内容
为了解决上述技术问题或者至少部分地解决上述技术问题,本公开实施例提供了一种互动处理方法、装置、电子设备和存储介质,以提高教学效率和教学质量。
本公开实施例提供了一种互动处理方法,包括:
显示用户界面,所述用户界面包括一个或多个第一用户中每个第一用户的标识信息;
获取第二用户针对所述一个或多个第一用户中目标用户的互动指令;
根据所述互动指令所关联的互动内容生成虚拟形象;
将所述虚拟形象发送到所述目标用户的终端。
本公开实施例还提供了一种互动处理装置,包括:
显示模块,用于显示用户界面,所述用户界面包括一个或多个第一用户中每个第一用户的标识信息;
获取模块,用于获取第二用户针对所述一个或多个第一用户中目标用户的互动指令;
生成模块,用于根据所述互动指令所关联的互动内容生成虚拟形象;
发送模块,用于将所述虚拟形象发送到所述目标用户的终端。
本公开实施例还提供了一种电子设备,所述电子设备包括:
一个或多个处理器;
存储装置,用于存储一个或多个程序;
当所述一个或多个程序被所述一个或多个处理器执行,使得所述一个或多个处理器实 现如上所述的互动处理方法。
本公开实施例还提供了一种计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时实现如上所述的互动处理方法。
本公开实施例提供的技术方案至少具有如下优点:本公开实施例提供的互动处理方法,通过显示用户界面,并在该用户界面中显示一个或多个第一用户中每个第一用户的标识信息,从而使得第二用户可以根据每个第一用户的标识信息输入针对一个或多个第一用户中目标用户的互动指令,进一步可以根据该互动指令所关联的互动内容生成虚拟形象,并将该虚拟形象发送到该目标用户的终端,从而可以通过该虚拟形象与该目标用户进行互动,在第一用户数量较多的情况下,可以节省第二用户与个别第一用户的互动时间,从而可以提高互动效率。例如,在第一用户为学生、第二用户为老师的情况下,由于虚拟形象是针对个别学生生成的,并且老师与个别学生的互动内容可以承载在该虚拟形象中,不需要老师中途停止教学,因此,通过虚拟形象与个别学生进行互动,不仅可以节省老师与个别学生的互动时间,同时还可以保证其他学生正常听讲不受影响,在学生人数较多的大班网络课中,可以大大提高教学效率和教学质量。
附图说明
结合附图并参考以下具体实施方式,本公开各实施例的上述和其他特征、优点及方面将变得更加明显。贯穿附图中,相同或相似的附图标记表示相同或相似的元素。应当理解附图是示意性的,原件和元素不一定按照比例绘制。
图1为本公开实施例中的一种互动处理方法的流程图;
图2为本公开实施例中的一种应用场景的示意图;
图3为本公开实施例中的一种用户界面的示意图;
图4为本公开实施例中的另一种互动处理方法的流程图;
图5为本公开实施例中的另一种用户界面的示意图;
图6为本公开实施例中的又一种用户界面的示意图;
图7为本公开实施例中的又一种互动处理方法的流程图;
图8为本公开实施例中的再一种用户界面的示意图;
图9为本公开实施例中的再一种互动处理方法的流程图;
图10为本公开实施例中的另一种用户界面的示意图;
图11为本公开实施例中的另一种用户界面的示意图;
图12为本公开实施例中的另一种互动处理方法的流程图;
图13为本公开实施例中的另一种用户界面的示意图;
图14为本公开实施例中的一种互动处理装置的结构示意图;
图15为本公开实施例中的一种电子设备的结构示意图。
具体实施方式
下面将参照附图更详细地描述本公开的实施例。虽然附图中显示了本公开的某些实施 例,然而应当理解的是,本公开可以通过各种形式来实现,而且不应该被解释为限于这里阐述的实施例,相反提供这些实施例是为了更加透彻和完整地理解本公开。应当理解的是,本公开的附图及实施例仅用于示例性作用,并非用于限制本公开的保护范围。
应当理解,本公开的方法实施方式中记载的各个步骤可以按照不同的顺序执行,和/或并行执行。此外,方法实施方式可以包括附加的步骤和/或省略执行示出的步骤。本公开的范围在此方面不受限制。
本文使用的术语“包括”及其变形是开放性包括,即“包括但不限于”。术语“基于”是“至少部分地基于”。术语“一个实施例”表示“至少一个实施例”;术语“另一实施例”表示“至少一个另外的实施例”;术语“一些实施例”表示“至少一些实施例”。其他术语的相关定义将在下文描述中给出。
需要注意,本公开中提及的“第一”、“第二”等概念仅用于对不同的装置、模块或单元进行区分,并非用于限定这些装置、模块或单元所执行的功能的顺序或者相互依存关系。
需要注意,本公开中提及的“一个”、“多个”的修饰是示意性而非限制性的,本领域技术人员应当理解,除非在上下文另有明确指出,否则应该理解为“一个或多个”。
本公开实施方式中的多个装置之间所交互的消息或者信息的名称仅用于说明性的目的,而并不是用于对这些消息或信息的范围进行限制。
图1为本公开实施例中的一种互动处理方法的流程图,本实施例可适用于客户端中进行互动处理的情况,该方法可以由互动处理装置执行,该装置可以采用软件和/或硬件的方式实现,该装置可配置于电子设备中,例如终端,具体包括但不限于智能手机、掌上电脑、平板电脑、带显示屏的可穿戴设备、台式机、笔记本电脑、一体机、智能家居设备等。或者,本实施例可适用于服务端中进行互动处理的情况,该方法可以由互动处理装置执行,该装置可以采用软件和/或硬件的方式实现,该装置可配置于电子设备中,例如服务器。
本公开实施例所述的互动处理方法可以适用于如图2所示的应用场景。该应用场景中可以包括一个或多个第一终端,例如,第一终端21、第一终端22和第一终端23,其中,每个第一终端可以是第一用户的终端。第一用户例如可以是学生。可以理解的是,第一终端的个数并不限于图2所示的3个,在其他实施例中还可以有更多个第一终端,本公开实施例以3个为例进行示意性说明。另外,本公开实施例并不限定第一终端的产品形态,例如具体可以是智能手机、掌上电脑、平板电脑、带显示屏的可穿戴设备、台式机、笔记本电脑、一体机、智能家居设备等。
如图2所示,该应用场景中还可以包括一个或多个第二终端,例如,第二终端30,第二终端可以是第二用户的终端。第二用户例如可以是老师。可以理解的是,第二终端的个数并不限于图2所示的1个,在其他实施例中还可以有更多个第二终端,本公开实施例以1个为例进行示意性说明。另外,本公开实施例并不限定第二终端的产品形态,例如具体可以是智能手机、掌上电脑、平板电脑、带显示屏的可穿戴设备、台式机、笔记本电脑、一体机、智能家居设备等。
可以理解的是,第一用户并不限于学生,第二用户也并不限于老师。例如,第二用户还可以是直播播主,第一用户可以是观看直播的用户。也就是说,本公开实施例所述的互 动处理方法可以适用于一对多或一对一互动的场景中。
具体的,第二终端30可以分别与第一终端21、第一终端22和第一终端23进行通信。例如,第二终端30、第一终端21、第一终端22和第一终端23可以连接到服务器40,从而使得第二终端30可以通过服务器40以一对多的方式与第一终端21、第一终端22和第一终端23进行通信。另外,第一终端21、第一终端22和第一终端23之间也可以通过服务器40进行相互通信。在一些实施例中,服务器40可以是服务器集群或者是服务器集群中的一个或多个服务器。另外,服务器40还可以是云服务器。在一种可能的实现方式中,第二终端30、第一终端21、第一终端22和第一终端23中分别安装有同一款应用程序(Application,简称APP),服务器40可以给该APP提供服务。
具体的,本实施例所述的互动处理方法可以由如图2所示的第二终端30或服务器40执行,下面以第二终端30为例进行示意性说明。
如图1所示,该方法具体可以包括:
S101、显示用户界面,其中,用户界面包括一个或多个第一用户中每个第一用户的标识信息。
例如图2所示的第二终端30为台式机,第二终端30包括主机和显示组件。该显示组件具体可以是显示屏。具体的,第二终端30可以在显示屏中显示用户界面,在第二终端的第二用户为老师的场景中,该用户界面可以是老师的用户界面。该用户界面可以包括一个或多个第一用户中每个第一用户的标识信息。
可选的,第一用户的标识信息包括如下至少一种:第一用户的图像信息、名称、标识号、第一用户的终端标识。
例如,在第一用户为学生的场景中,老师的用户界面中可以显示正在听课的一个或多个学生的标识信息。其中,每个学生的标识信息包括该学生的图像信息、名称、标识号、该学生的第一终端的终端标识中的至少一个。其中,图像信息可以是学生的头像或其他图像。名称可以是学生的姓名或昵称。标识号例如可以是学生的学号、账号等。第一终端的终端标识可以是第一终端的端口号、硬件标识号等。另外,在老师的用户界面中,并不限定多个学生的标识信息的展现形式,例如,多个学生的标识信息可以以列表的形式展现在老师的用户界面中。如图3所示为老师的用户界面的示意图,该用户界面中显示有学生张三的姓名和张三的头像31、学生李四的姓名和李四的头像32、学生王五的姓名和王五的头像33。可以理解的是,如图3所示的用户界面只是一种示意性说明,并不做具体限定,例如,该用户界面中还可以包括其他图标或按钮等。
S102、获取第二用户针对一个或多个第一用户中目标用户的互动指令。
如图3所示,老师的用户界面中还可以显示有老师在教学过程中使用的电子教材34,该电子教材34可以是文本数据、视频数据、音频数据、图像数据中的至少一种。可以理解的是,电子教材34和多个学生的标识信息可以显示在同一个显示屏中,也可以显示在不同的显示屏中,此处不做具体限定。另外,第二终端30可以通过服务器40将电子教材34实时发送到第一终端21、第一终端22和第一终端23,从而使得学生可以实时通过第一终端观看到电子教材34。
此外,如图2所示,第二终端30还可以包括拍摄装置,或者第二终端30可以与拍摄装置连接。该拍摄装置例如可以是摄像头,该摄像头可以集成在第二终端30的显示组件中,或者该摄像头可以与第二终端30的主机连接。具体的,老师在教学的过程中,该摄像头可以实时采集老师的图像或视频信息。第二终端30可以通过服务器40实时的将老师的图像或视频信息发送到第一终端21、第一终端22和第一终端23。从而使得第一终端21、第一终端22和第一终端23在接收到老师的图像或视频信息的情况下,将老师的图像或视频信息实时显示在第一终端21、第一终端22和第一终端23各自的显示屏中,从而使得学生可以实时通过第一终端观看到老师的图像或视频信息。
例如,老师在教学的过程中可能会需要与个别学生进行互动,例如提醒个别学生注意听讲或让个别学生回答问题。此处的个别学生可以是一个学生,也可以是多个学生。另外,此处可以将个别学生记为目标用户。
此时,老师可以向第二终端30输入针对该个别学生的互动指令。相应的,第二终端30可以获取老师针对该个别学生的互动指令。本实施例并不限定互动指令的输入方式。例如,在一种可能的实现方式中,如图3所示,老师可以对该用户界面中某个学生的头像进行操作,该操作例如可以是点击、双击、长按、语音控制操作、或手势控制操作等。此时,第二终端30可以根据老师对该学生的头像的操作,获取老师针对该学生的互动指令。
S103、根据互动指令所关联的互动内容生成虚拟形象。
例如,第二终端30在获取到老师针对某个学生的互动指令后,可以显示预设的互动内容,进一步根据老师对互动内容的操作确定老师选择的互动内容,其中,老师选择的互动内容可以作为该互动指令所关联的互动内容。或者第二终端30在获取到老师针对某个学生的互动指令后可以显示输入框,进一步接收老师在该输入框中输入的互动内容,老师在该输入框中输入的互动内容可以作为该互动指令所关联的互动内容。该互动内容可以是文本数据、视频数据、音频数据、图像数据中的至少一种。进一步的,第二终端30可以根据该互动指令所关联的互动内容生成虚拟形象。可以理解的是,该虚拟形象不限于在第二终端30这一侧生成,例如,该虚拟形象还可以在服务器40这一侧生成。
S104、将虚拟形象发送到目标用户的终端。
例如,张三是老师需要互动的目标用户。第二终端30或服务器40在生成虚拟形象之后,还可以将该虚拟形象发送到该张三的第一终端,例如第一终端21。第一终端21在接收到该虚拟形象的情况下,可以将该虚拟形象显示在第一终端21的屏幕中,从而使得张三可以在第一终端21中看到该虚拟形象。
可选的,将虚拟形象发送到目标用户的终端可以包括:先将目标用户对应的标记添加到虚拟形象;再将添加有该标记的所述虚拟形象发送到所述目标用户的终端。
由于每个学生的第一终端分别对应不同的端口,并且每个端口可以分别对应不同的标记,例如,学生张三的第一终端21的端口1所对应的标记为话术“张三”,学生李四的第一终端22的端口2所对应的标记为话术“李四”。在第二终端30确定老师与张三进行互动时,可以将话术“张三”添加到虚拟形象中,并将添加有话术“张三”的虚拟形象通过端口1发送到第一终端21。也就是说,每个学生的第一终端对应的客户端所呈现的虚拟形象 的话术可以是不同的。其中,话术可以理解为虚拟形象说出的语音内容。第一终端21通过端口1所接收到的虚拟形象为虚拟人物的面部表情在动,张三听到的声音例如可以是“张三,请注意听讲”。也就是说,在第一终端播放虚拟形象的过程中,播放时长可能是几秒,在这几秒中,学生从第一终端中听到或看到的是虚拟人物在说话,而不是真实的老师在说话以便让所有的学生都在听。
可以理解的是,虚拟形象其实可以是一小段视频,该小段视频包括图像信息和音频信息,该图像信息可以是虚拟人物的面部图像,该音频信息可以是老师发起的互动内容所对应的音频信息。其中,该虚拟人物的面部图像可以是第二终端30的第二用户(即真实老师)的面部图像,或者该虚拟人物的面部图像还可以是其他真实人物或其他虚拟人物的面部图像。另外,虚拟形象中的音频信息的属性例如音色、音调或音量等可以与真实老师的音色、音调或音量等属性相同。或者,该虚拟形象中的音频信息的属性还可以是其他真实人物或其他虚拟人物的音频属性。
另外,在老师教学过程中,第二终端30的拍摄装置采集的老师的视频信息可以是老师的真实视频,该真实视频中包括的是老师的真实人物图像和真实音频信息。服务器40在进行视频分发的时候,老师的真实视频是需要分发给每个学生的,而针对个别学生的虚拟形象只需要发送到个别学生的第一终端即可。也就是说,老师的真实视频是每个学生共享的信息,而虚拟形象是学生个性化的信息。当第二终端30针对某一学生生成虚拟形象时,服务器40可以将老师的真实视频和针对该学生的虚拟形象一起发送到该学生的第一终端,或者服务器40可以只将针对该学生的虚拟形象发送到该学生的第一终端,在服务器40将该虚拟形象发送完成后继续向该学生的第一终端发送老师的真实视频。
本公开实施例提供的互动处理方法,通过显示用户界面,并在该用户界面中显示一个或多个第一用户中每个第一用户的标识信息,从而使得第二用户可以根据每个第一用户的标识信息输入针对一个或多个第一用户中目标用户的互动指令,进一步可以根据该互动指令所关联的互动内容生成虚拟形象,并将该虚拟形象发送到该目标用户的终端,从而可以通过该虚拟形象与该目标用户进行互动,在第一用户数量较多的情况下,可以节省第二用户与个别第一用户的互动时间,从而可以提高互动效率。例如,在第一用户为学生、第二用户为老师的情况下,由于虚拟形象是针对个别学生生成的,并且老师与个别学生的互动内容可以承载在该虚拟形象中,不需要老师中途停止教学,因此,通过虚拟形象与个别学生进行互动,不仅可以节省老师与个别学生的互动时间,同时还可以保证其他学生正常听讲不受影响,在学生人数较多的大班网络课中,可以大大提高教学效率和教学质量。
图4为本公开实施例中的另一种互动处理方法的流程图。例如,在上述实施例的基础上,互动处理方法还可以包括如图4所示的如下几个步骤:
S401、在用户界面中显示每个第一用户的视频信息。
如图2所示,第一终端21、第一终端22和第一终端23中的每个第一终端可以包括拍摄装置,例如摄像头。或者,第一终端21、第一终端22和第一终端23中的每个第一终端可以分别与摄像头通信连接。该摄像头可以实时采集学生的图像或视频信息。进一步的,每个第一终端可以将各自对应的学生的图像或视频信息发送到第二终端30。例如,每个第 一终端通过服务器40将各自对应的学生的图像或视频信息发送到第二终端30。在第二终端30接收到各个学生的图像或视频信息时,可以将各个学生的图像或视频信息显示在老师的用户界面中。如图5所示为老师的用户界面的另一示意图,该用户界面可以包括多个视频播放器,每个视频播放器用于播放一个学生的视频信息。例如图5所示,该用户界面中显示有学生张三的姓名和张三的视频信息、李四的姓名和李四的视频信息、王五的姓名和王五的视频信息,其中,张三的视频信息在视频播放器51中进行播放、李四的视频信息在视频播放器52中进行播放、王五的视频信息在视频播放器53中进行播放。可以理解的是,如图5所示的用户界面只是一种示意性说明,并不做具体限定,例如,该用户界面中还可以包括其他图标或按钮等。
S402、根据第二用户对目标用户的视频信息的第一操作,显示与目标用户关联的第一操作接口。
例如,在图5的基础上,老师可以在该用户界面中对目标用户(即个别学生)的视频信息进行操作,该操作记为第一操作,第一操作例如可以是点击、双击、长按、语音控制操作、或手势控制操作等。或者该第一操作还可以包括老师将鼠标箭头移动到该视频信息上。进一步的,第二终端30可以根据老师的第一操作在该用户界面中显示与该目标用户关联的第一操作接口。
例如,当老师需要与张三进行互动时,老师可以移动鼠标将鼠标箭头移动到张三的视频信息上。在第二终端30检测到鼠标箭头停留在视频播放器51上的时长大于或等于预设时长时,可以在该用户界面中显示与张三关联的第一操作接口511。进一步,第二终端30可以根据老师对第一操作接口511的操作获取老师针对张三的互动指令。或者,当老师将鼠标箭头从视频播放器51上移动到视频播放器52上之后,第二终端30还可以控制第一操作接口511隐藏并显示与李四关联的第一操作接口512。
作为另一种实现方式,用户界面还包括:与每个第一用户的标识信息分别关联的第一操作接口。
如图6所示,在图3的基础上,该用户界面中还可以包括与每个学生的标识信息分别关联的第一操作接口。例如,第一操作接口511与张三的标识信息关联、第一操作接口512与李四的标识信息关联、第一操作接口513与王五的标识信息关联。也就是说,第一操作接口可以不需要根据老师对视频信息的第一操作动态显示,而是在用户界面显示的同时即呈现在该用户界面中。进一步的,第二终端30可以根据老师对第一操作接口511的操作获取老师针对张三的互动指令。第二终端30根据该互动指令关联的互动内容生成虚拟形象,并将该虚拟形象发送到张三的终端例如第一终端21。也就是说,老师可以针对单个学生发起虚拟形象,针对单个学生的虚拟形象可以通过该单个学生的端口发送到该单个学生的终端,从而呈现在该单个学生所使用的客户端中。
可以理解的是,第一操作接口的显示方式不限于图5或图6的显示方式,还可以包括其他的显示方式,此处只是示意性说明。另外,如图6所示的用户界面也是一种示意性说明,并不做具体限定,例如,并不限定同一个学生的标识信息与第一操作接口之间的相对位置关系。此外,本公开实施例也不对第一操作接口进行具体限定,例如,第一操作接口 可以包括图标、按钮、按键、输入框、或预设文本信息等。
另外,用户界面还包括:第二操作接口,且该第二操作接口与一个或多个第一用户中的部分或全部第一用户的标识信息关联,目标用户包括部分或全部第一用户。
例如,在一些实施例中,老师的用户界面还可以显示有第二操作接口,第二操作接口与如上所述的第一操作接口的不同之处在于:一个第一操作接口与一个学生关联,但是第二操作接口可以与多个学生关联。例如,该用户界面中总共显示有3个学生的标识信息,该第二操作接口可以与该3个学生中的部分学生或全部学生的标识信息关联,例如,部分学生的数量大于1。相应的,如上所述的目标用户可以是该部分学生或全部学生。
另外,在用户界面包括第二操作接口的情况下,在获取第二用户针对一个或多个第一用户中目标用户的互动指令之前,该互动处理方法还包括如图7所示的如下几个步骤:
S701、检测第二用户在所述用户界面中对第二操作接口和部分或全部第一用户的标识信息的关联操作。
如图8所示,该用户界面包括第二操作接口80。在老师需要与多个学生互动时,老师可以对第二操作接口80和该多个学生的标识信息进行关联操作。例如,张三和李四是待发起虚拟形象的学生,老师可以对张三的标识信息和李四的标识信息进行多选,并在选中之后将张三的标识信息和李四的标识信息拖拽或拖动到第二操作接口80或第二操作接口80周围的预设区域内。相应的,第二终端30可以将该多选、拖拽或拖动操作确定为老师对第二操作接口80、张三的标识信息和李四的标识信息进行关联的关联操作。可以理解的是,此处并不对该关联操作进行具体限定,该多选、拖拽或拖动操作只是一种示意性说明。
S702、根据关联操作建立第二操作接口和部分或全部第一用户的标识信息的关联关系。
进一步的,第二终端30可以根据该关联操作,建立第二操作接口80、张三的标识信息和李四的标识信息的关联关系。例如,第二终端30可以在第二操作接口80周围的预设区域内显示张三的姓名和李四的姓名。可以理解的是,本实施例并不限定该关联关系的展现形式,图8所示的展现形式只是一种示意性说明。
另外,在老师需要与所有学生互动时,可以对所有学生的标识信息进行全选,并进一步将所有学生的标识信息与第二操作接口80进行关联,具体的关联操作、以及关联关系可以参照前述方法,此处不再赘述。
进一步的,第二终端30可以根据老师对第二操作接口80的操作获取老师针对张三和李四的互动指令,根据该互动指令分别生成针对张三和李四的虚拟形象,并将针对张三的虚拟形象通过张三对应的端口发送到张三的第一终端21,将针对李四的虚拟形象通过李四对应的端口发送到李四的第一终端22,从而使得第一终端21可以在张三所使用的客户端显示针对张三的虚拟形象,第一终端22可以在李四所使用的客户端显示针对李四的虚拟形象。其中,针对张三的虚拟形象和针对李四的虚拟形象可以相同,也可以不同。也就是说,通过老师对一个第二操作接口的操作,可以针对多个学生或所有学生发起虚拟形象,并且可以将虚拟形象通过各个学生的端口发送到各个学生的第一终端,在各个学生所使用的客户端呈现虚拟形象,并且在各个学生的客户端所呈现的虚拟形象可以不同。
本公开实施例提供的互动处理方法,通过在用户界面显示与多个第一用户关联的第二 操作接口,使得第二用户可以通过对第二操作接口操作时实现与多个第一用户同时互动,进一步提高了互动效率。
通过上述实施例可知,老师可以通过对个别学生的标识信息进行操作,以输入针对个别学生的互动指令,或者老师还可以通过对如上所述的第一操作接口或第二操作接口进行操作,以输入针对个别学生的互动指令,下面结合具体的实施例对第二终端30通过第一操作接口或第二操作接口获取互动指令的情况进行介绍。
具体的,获取第二用户针对一个或多个第一用户中目标用户的互动指令可以包括如图9所示的如下几个步骤:
S901、检测第二用户针对目标用户关联的第一操作接口或第二操作接口的第二操作。
例如在图5、图6或图8所示的用户界面中,第二终端30可以检测老师对第一操作接口或第二操作接口的操作,例如,老师对第一操作接口或第二操作接口的操作可以记为第二操作。该第二操作不限于点击、双击、长按、语音控制操作、或手势控制操作等。其中,一个第一操作接口关联一个学生,一个第二操作接口关联多个学生。
S902、根据第二操作获取第二用户针对目标用户的互动指令。
第二终端30可以根据老师对第一操作接口或第二操作接口的第二操作,获取老师针对目标用户(即个别学生)的互动指令。例如图6所示,在第二终端30检测到老师对第一操作接口512的第二操作时,可以获取老师针对李四的互动指令。例如图8所示,在第二终端30检测到老师对第二操作接口80的第二操作时,可以获取老师针对张三和李四的互动指令。
在本公开实施例中,每个第一操作接口可以与一个或多个预设文本信息关联,和/或第二操作接口可以与一个或多个预设文本信息关联。该预设文本信息可以是老师与学生之间常用的交流内容,例如,“你真棒”、“注意听讲”、“请看屏幕”、“注意坐姿”等。
在一种可能的实现方式中,根据第二操作获取第二用户针对目标用户的互动指令可以包括:首先根据第二操作,在用户界面中显示与目标用户关联的一个或多个预设文本信息;然后检测第二用户对一个或多个预设文本信息中目标文本信息的选择操作;接着根据该选择操作获取第二用户针对目标用户的互动指令。
例如,以图6为例,在第二终端30检测到老师对第一操作接口512的第二操作时,第二终端30可以在该用户界面中显示与李四关联的一个或多个预设文本信息,此处以多个预设文本信息为例进行示意性说明。如图10所示为老师点击第一操作接口512之后显示有与李四关联的多个预设文本信息,例如“你真棒”、“注意听讲”、“请看屏幕”、“注意坐姿”的用户界面示意图。可以理解的是,图10所示的用户界面只是一种示意性说明,并不做具体限定,例如,并不限定多个预设文本信息在该用户界面中的显示位置、颜色、亮度等。在如图10所示的用户界面中,老师可以根据其互动需求选择该多个预设文本信息中的一个或多个,并且老师选中的预设文本信息可记为目标文本信息。进一步的,第二终端30可以检测老师对目标文本信息的选择操作,并根据该选择操作获取老师针对李四的互动指令。可以理解的是,不同学生所关联的一个或多个预设文本信息可以不同,也可以相同。或者老师还可以根据每个学生的特征在该用户界面中设置与其特征相符的一个或多个预设文本 信息。可选的,互动指令所关联的互动内容可以包括目标文本信息。例如,在图10所示的用户界面中,第二终端30可以根据老师对目标文本信息例如“你真棒”的选择操作,确定出老师针对李四的互动指令所关联的互动内容包括“你真棒”。可以理解的是,该互动内容还可以包括除该目标文本信息例如“你真棒”之外的其他内容,例如,该互动内容可以为“李四同学你真棒”。进一步的,第二终端30可以根据该互动内容生成针对李四的虚拟形象,并通过服务器40将该虚拟形象发送到李四的第一终端22。
同理,以图8为例,在第二终端30检测到老师对第二操作接口80的第二操作时,第二终端30可以在该用户界面中显示与张三和李四同时关联的一个或多个预设文本信息。进一步的,第二终端30可以检测老师对该一个或多个预设文本信息中目标文本信息的选择操作,并根据该选择操作获取老师针对张三和李四的互动指令。该互动指令所关联的互动内容包括该目标文本信息。进一步的,第二终端30可以根据该互动内容分别生成针对张三的虚拟形象和针对李四的虚拟形象,并在针对张三的虚拟形象中加入张三对应的标记,例如话术“张三”;在针对李四的虚拟形象中加入李四对应的标记,例如话术“李四”。进一步的,第二终端30可以通过服务器40将添加有话术“张三”的虚拟形象发送到张三的第一终端21,将添加有话术“李四”的虚拟形象发送到李四的第一终端22。
在另一种可能的实现方式中,目标用户关联的第一操作接口或第二操作接口可以包括一个或多个预设文本信息;则检测第二用户针对目标用户关联的第一操作接口或第二操作接口的第二操作可以包括:检测第二用户针对一个或多个预设文本信息中目标文本信息的第二操作。
例如,在图5或图6中,每个第一操作接口可以包括一个或多个预设文本信息。也就是说,每个第一操作接口所关联的一个或多个预设文本信息并不是类似于图10一样在老师对第一操作接口进行第二操作之后才显示到用户界面上,而是在显示第一操作接口的同时即显示在第一操作接口中或第一操作接口周围的预设区域。因此,第二终端30在检测老师对第一操作接口的第二操作时具体可以检测老师对第一操作接口包括的一个或多个预设文本信息中目标文本信息的第二操作,该第二操作可以是对该目标文本信息的选择操作。例如在图5中,第一操作接口511所在的区域或第一操作接口511周围的预设区域内显示有“你真棒”、“注意听讲”、“请看屏幕”、“注意坐姿”这4个预设文本信息。当老师对该4个预设文本信息中的目标文本信息例如“注意坐姿”进行选择操作时,第二终端30可以根据该选择操作确定老师针对张三的互动指令所关联的互动内容包括该目标文本信息“注意坐姿”。例如,该互动内容还可以包括除目标文本信息“注意坐姿”之外的其他内容,例如,该互动内容可以为“张三小朋友注意坐姿”。进一步的,第二终端30可以根据该互动内容生成针对张三的虚拟形象,并通过服务器40将该虚拟形象发送到张三的第一终端21。
同理,在图8中,第二操作接口80所在的区域或第二操作接口80周围的预设区域内可显示有多个预设文本信息,例如,“你真棒”、“注意听讲”、“请看屏幕”、“注意坐姿”等。当老师对该多个预设文本信息中的目标文本信息例如“注意坐姿”进行选择操作时,第二终端30可以确定老师针对张三和李四的互动指令所关联的互动内容包括“注意坐姿”。进一步的,第二终端30可以根据该互动内容分别生成针对张三的虚拟形象和针对李四的虚拟 形象,并在针对张三的虚拟形象中加入张三对应的标记,例如话术“张三”;在针对李四的虚拟形象中加入李四对应的标记,例如话术“李四”。进一步的,第二终端30可以通过服务器40将添加有话术“张三”的虚拟形象发送到张三的第一终端21,将添加有话术“李四”的虚拟形象发送到李四的第一终端22。
可选的,本实施例提供的互动处理方法还包括:响应于第二用户对预设文本信息的修改指令、增加指令或删除指令,更新用户界面中的预设文本信息。
例如,在一些实施例中,老师还可以对第一操作接口或第二操作接口关联的预设文本信息进行修改、增加或删除。相应的,第二终端30可以响应于老师对预设文本信息的修改指令、增加指令或删除指令,更新老师用户界面中第一操作接口或第二操作接口关联的预设文本信息。
本公开实施例提供的互动处理方法,通过第一操作接口和/或第二操作接口与一个或多个预设文本信息关联,使得第二用户在与第一用户互动时,只需从用户界面中选择预设文本信息作为互动内容或互动内容的一部分即可,进一步提高了第二用户与第一用户之间的互动效率。
上述实施例介绍了第一操作接口和/或第二操作接口可以与一个或多个预设文本信息关联,其中,预设文本信息是预先设置的,而在一些应用场景中,多个预设文本信息中可能没有满足老师需求的预设文本信息。因此,在下面实施例中,第一操作接口和/或第二操作接口还可以与输入框关联。
在一种可能的实现方式中,根据第二操作获取第二用户针对目标用户的互动指令可以包括:首先根据第二操作,在用户界面中显示与第一操作接口或第二操作接口关联的输入框;然后检测第二用户在输入框中输入目标互动信息的输入操作;接着可以根据该输入操作获取第二用户针对目标用户的互动指令。
例如,以图6为例,在第二终端30检测到老师对第一操作接口512的第二操作时,第二终端30可以在该用户界面中显示与第一操作接口512关联的输入框。如图11所示为老师点击第一操作接口512之后显示有输入框110的用户界面示意图。可以理解的是,图11所示的用户界面只是一种示意性说明,并不做具体限定,例如,并不限定第一操作接口512和输入框110的相对位置等。在如图11所示的用户界面中,老师可以根据其互动需求在该输入框110中输入目标互动信息,该目标互动信息可以是文本信息、音频信息、视频信息、图像信息中的至少一种。进一步的,第二终端30可以检测老师在该输入框110中输入目标互动信息的输入操作,并根据该输入操作获取老师针对李四的互动指令。可选的,互动指令所关联的互动内容可以包括目标互动信息。例如,在图11所示的用户界面中,第二终端30可以根据老师在输入框110中输入的目标互动信息,确定出老师针对李四的互动指令所关联的互动内容包括该目标互动信息。例如,该互动内容还可以包括除该目标互动信息之外的其他内容,例如李四的昵称或姓名等。进一步的,第二终端30可以根据该互动内容生成针对李四的虚拟形象并通过服务器40将该虚拟形象发送到李四的第一终端22。
同理,以图8为例,在第二终端30检测到老师对第二操作接口80的第二操作时,第二终端30可以在该用户界面中显示输入框,进一步的,第二终端30可以检测老师在该输 入框中输入目标互动信息的输入操作,并根据该输入操作获取老师针对张三和李四的互动指令。该互动指令所关联的互动内容包括该目标互动信息。进一步的,第二终端30可以根据该互动内容分别生成针对张三的虚拟形象和针对李四的虚拟形象,并在针对张三的虚拟形象中加入张三对应的标记,例如话术“张三”;在针对李四的虚拟形象中加入李四对应的标记,例如话术“李四”。进一步的,第二终端30可以通过服务器40将添加有话术“张三”的虚拟形象发送到张三的第一终端21,将添加有话术“李四”的虚拟形象发送到李四的第一终端22。
在另一种可能的实现方式中,第一操作接口或第二操作接口可以包括输入框;则检测第二用户针对目标用户关联的第一操作接口或第二操作接口的第二操作可以包括:检测第二用户在输入框中输入目标互动信息的输入操作。
例如,在图5或图6中,每个第一操作接口可以包括一个输入框。也就是说,每个第一操作接口所关联的输入框并不是类似于图11一样在老师对第一操作接口进行第二操作之后才显示到用户界面上,而是在显示第一操作接口的同时即显示在第一操作接口中或第一操作接口周围的预设区域。进一步的,老师可以在该输入框中输入目标互动信息。老师在不同第一操作接口包括的输入框中输入的目标互动信息可以不同,例如,在张三关联的第一操作接口包括的输入框中输入“注意听讲”,在李四关联的第一操作接口包括的输入框中输入“请看屏幕”。第二终端30在检测老师对第一操作接口的第二操作时具体可以检测老师在输入框中输入目标互动信息的输入操作。
可选的,互动指令所关联的互动内容可以包括目标互动信息。例如,当老师在张三关联的第一操作接口包括的输入框中输入目标互动信息例如“注意坐姿”时,第二终端30可以确定老师针对张三的互动指令所关联的互动内容包括“注意坐姿”。并且该互动内容还可以包括除目标文本信息“注意坐姿”之外的其他内容,例如,该互动内容可以为“张三小朋友注意坐姿”。进一步的,第二终端30可以根据该互动内容生成针对张三的虚拟形象并通过服务器40将该虚拟形象发送到张三的第一终端21。
同理,在图8中,第二操作接口80可以包括一个输入框,当老师在该输入框中输入目标互动信息,例如“注意坐姿”时,第二终端30可以确定老师针对张三和李四的互动指令所关联的互动内容包括“注意坐姿”。进一步的,第二终端30可以根据该互动内容分别生成针对张三的虚拟形象和针对李四的虚拟形象,并在针对张三的虚拟形象中加入张三对应的标记,例如话术“张三”;在针对李四的虚拟形象中加入李四对应的标记,例如话术“李四”。进一步的,第二终端30可以通过服务器40将添加有话术“张三”的虚拟形象发送到张三的第一终端21,将添加有话术“李四”的虚拟形象发送到李四的第一终端22。
本公开实施例提供的互动处理方法,通过第一操作接口和/或第二操作接口与输入框关联,使得第二用户在与第一用户互动时,只需在该输入框中输入互动内容或互动内容的一部分即可,进一步提高了第二用户与第一用户之间的互动效率,同时还提高了互动内容的灵活性。
在上述实施例的基础上,本实施例提供的互动处理方法中“根据互动指令所关联的互动内容生成虚拟形象”可以包括如图12所示的如下几个步骤:
S1201、获取与互动内容对应的音频。
本实施例以互动内容为文本信息为例进行示意性说明。例如,正在进行网上大班授课的老师如果需要在不影响其他学生正常上课的情况下,与指定的一个或多个学生进行互动时,可以触发第二终端30创建虚拟形象,并采用该创建的虚拟形象与指定的学生进行互动,而针对其他学生则仍然采用自己的真实形象(即真实视频)进行网络授课。具体的,触发第二终端30创建虚拟形象的方式可以参照上述实施例所述的内容,此处不再赘述。
在第二终端30确定出互动内容例如文本信息“请注意听讲”时,可以确定该文本信息对应的时长。其中,确定时长的方式具体可以采用预先训练完成的时长预测模型来实现,例如,将文本信息“请注意听讲”输入该时长预测模型,获得时长为3秒。或者可以将文本信息划分为多个子序列,通过时长预测模型确定出每个子序列的时长,进一步将每个子序列的时长进行相加得到该文本信息的时长。此外,在其他实施例中,确定文本信息的时长的方法并不限于时长预测模型,例如还可以通过查表的方式实现。本实施例并不对确定文本信息的时长的方法进行限定。进一步的,可以将该文本信息和其对应的时长输入到预先训练完成的语音合成模型,获得该文本信息对应的音频,例如,将文本信息“请注意听讲”和时长3秒输入该语音合成模型后,该语音合成模型可以按照3秒时长将文本信息“请注意听讲”转换为音频。
S1202、获取与音频对应的面部特征点序列。
可选的,面部特征点序列可以包括多组面部特征点;则获取与音频对应的面部特征点序列可以包括:先根据音频获取音频帧序列;再将音频帧序列输入预先训练完成的特征点预测模型,获取面部特征点序列,其中,面部特征点序列中包括多组面部特征点,特征点预测模型用于确定每一个音频帧所对应的一组面部特征点。
例如,在获取到音频后可以根据音频获取音频帧序列,例如,时长为3秒的音频“请注意听讲”包括100个音频帧,进一步的,可以将该100个音频帧输入预先训练完成的特征点预测模型,获取面部特征点序列,该面部特征点序列中包括多组面部特征点,并且特征点预测模型所输出的面部特征点的组数与音频帧的个数相同,即每一个音频帧对应一组面部特征点。每一组面部特征点中特征点的数目可以不做具体限定,例如可以是68个。100组面部特征点按照帧顺序依次排列即构成面部特征点序列。
可选的,在将音频帧序列输入预先训练完成的特征点预测模型,获取面部特征点序列之前,还包括:首先获取媒体文件,其中,媒体文件包括样本音频和同步播放的面部视频;然后从样本音频中提取样本音频帧,以及从面部视频中提取与样本音频帧同步的样本视频帧;接着提取样本视频帧中所包括的样本面部特征点;进而可以根据样本音频帧和样本面部特征点对特征点预测模型进行训练。
也就是说,在将音频帧序列输入特征点预测模型之前,需要先采用样本对特征点预测模型进行训练,以使特征点预测模型在正式应用的过程中能够输出与每个音频帧所对应的一组面部特征点,从而保证面部特征点获取的准确性。样本获取的方式是获取一个媒体文件,该媒体文件可以是老师进行网上授课时或网上授课前的一段录屏。该媒体文件中包括样本音频和同步播放的面部视频,由于样本音频和面部视频是同步播放的,因此,从样本 音频中提取样本音频帧,以及从面部视频中提取与样本音频帧同步的样本视频帧后,样本音频帧和样本视频帧的帧数相同,例如,样本音频帧提取到50个,样本视频帧的个数也为50个。以50个样本音频帧中的第一个样本音频帧和50个样本视频帧中的第一个样本视频帧为例,提取该第一个样本视频帧中所包括的样本面部特征点,将第一个样本音频帧和第一个样本视频帧所包括的样本面部特征点作为一组样本,则相应的可以依次提取到50组样本。进一步的,根据该50组样本对特征点预测模型进行训练,当确定训练误差达到预设阈值时,确定训练完成,并采用训练完成的特征点预测模型进行后续的面部特征点的预测。
S1203、根据面部特征点序列获取视频。
可选的,根据面部特征点序列获取视频可以包括:首先获取目标用户对应的面部标准模板;然后将每一组面部特征点添加到面部标准模板,获取每一帧面部图像;接着将每一帧面部图像按帧顺序依次连接获取所述视频。
例如,本实施例中的面部标准模板可以是在老师网上授课过程中随机截取的老师的一张脸部正面照片。进一步再将上述面部特征点序列中的每一组面部特征点分别添加到该面部标准模板中,得到每一帧面部图像。可以理解的是,由于面部特征点序列中的每一组面部特征点可能不同,因此,将不同的面部特征点添加到该面部标准模板后可以得到不同的面部图像,另外每一帧面部图像也有可能与该面部标准模板不同。例如,在确定音频“请注意听讲”包括100个音频帧,100个音频帧对应100组面部特征点时,相应的,可以获取100帧面部图像,并将所获取的100帧面部图像按帧顺序依次连接获得视频。
在一些实施例中,针对不同学生生成虚拟形象时所采用的面部标准模板可以不同。
在一种可能的实现方式中,本实施例提供的互动处理方法还包括:响应于第二用户对第一用户的第一设置指令,显示多个预设的面部标准模板;再根据第二用户对多个预设的面部标准模板中目标模板的选择操作,将目标模板设置为第一用户对应的面部标准模板。
例如,老师可以在其用户界面中对某个学生或某几个学生对应的面部标准模板进行设置,第二终端30可以响应于老师对某个学生或某几个学生的第一设置指令,在该用户界面中显示多个预设的面部标准模板,进一步的,根据老师对该多个预设的面部标准模板中目标模板的选择操作,可以将该目标模板设置为某个学生或某几个学生对应的面部标准模板。其中,老师可以从该多个预设的面部标准模板中随机的选取一个面部标准模板作为目标模板。
在另一种可能的实现方式中,本实施例提供的互动处理方法还包括:接收第一用户的终端发送的第二设置指令,且该第二设置指令用于设置第一用户对应的面部标准模板。
例如,在一些应用场景中,学生可以根据自己的喜好设置其对应的虚拟形象所采用的面部标准模板。例如,第一终端可以在学生界面中显示多个预设的面部标准模板,学生可以从该多个预设的面部标准模板中选择自己喜欢的一个作为其对应的虚拟形象所采用的面部标准模板。在这种情况下,第一终端可以将由学生设置面部标准模板的第二设置指令发送给第二终端30,从而使得第二终端30在生成针对该学生的虚拟形象时采用该学生自己设置的面部标准模板。
可以理解的是,每个学生对应的面部标准模板并不限于上述的几种设置方式,还可以 有其他的设置方式,此处不做具体限定。
S1204、将音频和视频结合获得虚拟形象。
例如,在S1203中生成的视频中可以包括100帧面部图像,则还需要进一步将该视频和音频“请注意听讲”进行结合获得虚拟形象。具体的,音频“请注意听讲”包括100个音频帧,S1203中生成的视频包括100帧面部图像,因此可以将该100个音频帧和该100帧面部图像按照相同的时间轴进行音频帧和面部图像的组合,例如,将该100个音频帧中的第一个音频帧与该100帧面部图像中的第一帧面部图像进行组合,以此类推,直至将该100个音频帧中的第100个音频帧与该100帧面部图像中的第100帧面部图像进行组合,以得到虚拟形象,从而使得第一终端所呈现的虚拟形象的声音和图像是匹配的,时序上是对齐的。
在本实施例中,针对某个学生发起虚拟形象,或者同时针对多个学生或所有学生发起虚拟形象时,针对每个学生的虚拟形象的表情、动作等可以不同。也就是说,每个学生的第一终端所呈现的虚拟形象的表情、动作等可以不同。另外,同时针对多个学生或所有学生发起虚拟形象时,该多个学生或所有学生中每个学生的第一终端上可以同时呈现虚拟形象,但每个学生的第一终端所呈现的虚拟形象的表情、动作等可以不同。
在一种可能的实现方式中,本实施例提供的互动处理方法还包括:根据互动内容的类型信息,确定与类型信息对应的表情信息和/或动作信息;并将表情信息和/或动作信息添加到虚拟形象中。
例如,老师的用户界面中显示的预设文本信息是按照类型进行划分的。如图13所示,第二终端30检测到老师对第一操作接口512进行操作时,在该用户界面中显示几种类型的预设文本信息,例如,夸赞类的预设文本信息包括“你真棒”,提醒类的预设文本信息包括“注意听讲”、“请看屏幕”、“注意坐姿”。当老师选择某一个预设文本信息时,第二终端30可以确定该预设文本信息对应的类型信息,并确定与该类型信息对应的表情信息和/或动作信息,在生成虚拟形象的过程中或在生成虚拟形象之后,将该表情信息和/或动作信息添加到该虚拟形象中,使得该虚拟形象说话的内容与该虚拟形象的表情、动作等是匹配的。例如,虚拟形象说“张三你真棒”时,该虚拟形象可以配合显示竖起大拇指的动作。再例如,虚拟形象说“注意听讲”时,该虚拟形象可以配合显示略带生气的表情。其中,预设文本信息对应的类型信息、以及类型信息对应的表情信息和/或动作信息可以预先存储在第二终端30或服务器40中。
可以理解的是,图13所示的用户界面只是一种示意性说明,并不做具体限定。例如,第二终端30检测到老师对第一操作接口512进行操作时,可以在该用户界面中显示类型信息,例如“夸赞”图标、“提醒”图标。老师可以对“夸赞”图标或“提醒”图标进行点击,在点击之后,第二终端30进一步显示夸赞类的预设文本信息或提醒类的预设文本信息。
在另一种可能的实现方式中,本实施例提供的互动处理方法还包括:对互动内容进行情感分析,得到情感信息;并将情感信息对应的表情信息和/或动作信息添加到虚拟形象中。
例如图10所示,老师随机选择了“你真棒”、“注意听讲”、“请看屏幕”、“注意坐姿”中的某一个预设文本信息,或者如图11所示,老师在输入框110中输入了目标互动信息。 进一步的,第二终端30可以对老师选择的预设文本信息或输入的目标互动信息进行情感分析,得到情感信息,并在生成虚拟形象的过程中或在生成虚拟形象之后,将该情感信息对应的表情信息和/或动作信息添加到虚拟形象中,从而使得该虚拟形象说话的内容与该虚拟形象的表情、动作等是匹配的。
可以理解的是,如上所述的S1201-S1204可以由第二终端30来执行,或者还可以由服务器40来执行,例如,第二终端30可以将老师设置的互动内容发送给服务器40,服务器40可以进一步执行S1201-S1204,从而获得虚拟形象。
本公开实施例提供的互动处理方法,根据互动内容的类型信息,确定与该类型信息对应的表情信息和/或动作信息;并将该表情信息和/或该动作信息添加到该虚拟形象中;或者对该互动内容进行情感分析,得到情感信息;并将该情感信息对应的表情信息和/或动作信息添加到该虚拟形象中,可使得虚拟形象说话的内容与该虚拟形象的表情、动作等是匹配的,从而可以提高虚拟形象的生动性。另外,通过老师对不同学生对应的面部标准模板进行设置或学生自己对其对应的面部标准模板进行设置,可使得不同学生对应的虚拟形象不同,进一步提高了虚拟形象的灵活性。
图14为本公开实施例中的一种互动处理装置的结构示意图。本公开实施例所提供的互动处理装置可以配置于客户端中,或者可以配置于服务端中,该互动处理装置140具体包括:
显示模块1401,用于显示用户界面,所述用户界面包括一个或多个第一用户中每个第一用户的标识信息;
获取模块1402,用于获取第二用户针对所述一个或多个第一用户中目标用户的互动指令;
生成模块1403,用于根据所述互动指令所关联的互动内容生成虚拟形象;
发送模块1404,用于将所述虚拟形象发送到所述目标用户的终端。
可选的,所述第一用户的标识信息包括如下至少一种:所述第一用户的图像信息、名称、标识号、所述第一用户的终端标识。
可选的,显示模块1401还用于:在所述用户界面中显示每个第一用户的视频信息;根据所述第二用户对所述目标用户的视频信息的第一操作,显示与所述目标用户关联的第一操作接口。
可选的,所述用户界面还包括:与每个第一用户的标识信息分别关联的第一操作接口。
可选的,所述用户界面还包括:第二操作接口,所述第二操作接口与所述一个或多个第一用户中的部分或全部第一用户的标识信息关联,所述目标用户包括所述部分或全部第一用户。
可选的,该互动处理装置140还包括:检测模块1405和关联关系建立模块1406,其中,检测模块1405用于获取模块1402获取第二用户针对所述一个或多个第一用户中目标用户的互动指令之前,检测所述第二用户在所述用户界面中对所述第二操作接口和所述部分或全部第一用户的标识信息的关联操作;关联关系建立模块1406用于根据所述关联操作建立所述第二操作接口和所述部分或全部第一用户的标识信息的关联关系。
可选的,检测模块1405还用于:检测所述第二用户针对所述目标用户关联的第一操作接口或第二操作接口的第二操作;获取模块1402还用于:根据所述第二操作获取所述第二用户针对所述目标用户的互动指令。
可选的,显示模块1401还用于:根据所述第二操作,在所述用户界面中显示与所述目标用户关联的一个或多个预设文本信息;检测模块1405还用于:检测所述第二用户对所述一个或多个预设文本信息中目标文本信息的选择操作;获取模块1402还用于:根据所述选择操作获取所述第二用户针对所述目标用户的互动指令。
可选的,所述目标用户关联的第一操作接口或第二操作接口包括一个或多个预设文本信息;检测模块1405具体用于:检测所述第二用户针对所述一个或多个预设文本信息中目标文本信息的第二操作。
可选的,所述互动指令所关联的互动内容包括所述目标文本信息。
可选的,显示模块1401还用于:根据所述第二操作,在所述用户界面中显示与所述第一操作接口或所述第二操作接口关联的输入框;检测模块1405还用于:检测所述第二用户在所述输入框中输入目标互动信息的输入操作;获取模块1402还用于:根据所述输入操作获取所述第二用户针对所述目标用户的互动指令。
可选的,所述第一操作接口或所述第二操作接口包括输入框;检测模块1405具体用于:检测所述第二用户在所述输入框中输入目标互动信息的输入操作。
可选的,所述互动指令所关联的互动内容包括所述目标互动信息。
可选的,该互动处理装置140还包括:更新模块1407,用于响应于所述第二用户对所述预设文本信息的修改指令、增加指令或删除指令,更新所述用户界面中的所述预设文本信息。
可选的,获取模块1402还用于:获取与所述互动内容对应的音频;获取与所述音频对应的面部特征点序列;根据所述面部特征点序列获取视频;生成模块1403具体用于:将所述音频和所述视频结合获得虚拟形象。
可选的,该互动处理装置140还包括:确定模块1408和添加模块1409,其中,确定模块1408用于根据所述互动内容的类型信息,确定与所述类型信息对应的表情信息和/或动作信息;添加模块1409用于将所述表情信息和/或所述动作信息添加到所述虚拟形象中。
可选的,确定模块1408还用于对所述互动内容进行情感分析,得到情感信息;添加模块1409还用于:将所述情感信息对应的表情信息和/或动作信息添加到所述虚拟形象中。
可选的,所述面部特征点序列包括多组面部特征点;获取模块1402在根据所述面部特征点序列获取视频时,具体用于:获取所述目标用户对应的面部标准模板;将每一组面部特征点添加到所述面部标准模板,获取每一帧面部图像;将所述每一帧面部图像按帧顺序依次连接,获取视频。
可选的,该互动处理装置140还包括:设置模块1410;显示模块1401还用于:响应于所述第二用户对第一用户的第一设置指令,显示多个预设的面部标准模板;设置模块1410用于根据所述第二用户对所述多个预设的面部标准模板中目标模板的选择操作,将所述目标模板设置为所述第一用户对应的面部标准模板。
可选的,该互动处理装置140还包括:接收模块1411,用于接收所述第一用户的终端发送的第二设置指令,所述第二设置指令用于设置所述第一用户对应的面部标准模板。
可选的,添加模块1409还用于将所述目标用户对应的标记添加到所述虚拟形象;发送模块1404具体用于:将添加有所述标记的所述虚拟形象发送到所述目标用户的终端。
本公开实施例提供的互动处理装置,可执行本公开方法实施例所提供的互动处理方法中客户端或服务端所执行的步骤,具备执行步骤和有益效果此处不再赘述。
图15为本公开实施例中的一种电子设备的结构示意图。下面具体参考图15,其示出了适于用来实现本公开实施例中的电子设备1500的结构示意图。本公开实施例中的电子设备1500可以包括但不限于诸如移动电话、笔记本电脑、数字广播接收器、PDA(个人数字助理)、PAD(平板电脑)、PMP(便携式多媒体播放器)、车载终端(例如车载导航终端)等等的移动终端以及诸如数字TV、台式计算机等等的固定终端。图15示出的电子设备仅仅是一个示例,不应对本公开实施例的功能和使用范围带来任何限制。
如图15所示,电子设备1500可以包括处理装置(例如中央处理器、图形处理器等)1501,其可以根据存储在只读存储器(ROM)1502中的程序或者从存储装置1508加载到随机访问存储器(RAM)1503中的程序而执行各种适当的动作和处理以实现如本公开所述的实施例的互动处理方法。在RAM 1503中,还存储有电子设备1500操作所需的各种程序和数据。处理装置1501、ROM 1502以及RAM 1503通过总线1504彼此相连。输入/输出(I/O)接口1505也连接至总线1504。
通常,以下装置可以连接至I/O接口1505:包括例如触摸屏、触摸板、键盘、鼠标、摄像头、麦克风、加速度计、陀螺仪等的输入装置1506;包括例如液晶显示器(LCD)、扬声器、振动器等的输出装置1507;包括例如磁带、硬盘等的存储装置1508;以及通信装置1509。通信装置1509可以允许电子设备1500与其他设备进行无线或有线通信以交换数据。虽然图15示出了具有各种装置的电子设备1500,但是应理解的是,并不要求实施或具备所有示出的装置。可以替代地实施或具备更多或更少的装置。
特别地,根据本公开的实施例,上文参考流程图描述的过程可以被实现为计算机软件程序。例如,本公开的实施例包括一种计算机程序产品,其包括承载在非暂态计算机可读介质上的计算机程序,该计算机程序包含用于执行流程图所示的方法的程序代码。在这样的实施例中,该计算机程序可以通过通信装置1509从网络上被下载和安装,或者从存储装置1508被安装,或者从ROM 1502被安装。在该计算机程序被处理装置1501执行时,执行本公开实施例的方法中限定的上述功能。
需要说明的是,本公开上述的计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质或者是上述两者的任意组合。计算机可读存储介质例如可以是——但不限于——电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。计算机可读存储介质的更具体的例子可以包括但不限于:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机访问存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、光纤、便携式紧凑磁盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。在本公开中,计算机可读存储介质可以是 任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。而在本公开中,计算机可读信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。计算机可读信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机可读信号介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。计算机可读介质上包含的程序代码可以用任何适当的介质传输,包括但不限于:电线、光缆、RF(射频)等等,或者上述的任意合适的组合。
在一些实施方式中,客户端、服务器可以利用诸如HTTP(HyperText Transfer Protocol,超文本传输协议)之类的任何当前已知或未来研发的网络协议进行通信,并且可以与任意形式或介质的数字数据通信(例如,通信网络)互连。通信网络的示例包括局域网(“LAN”),广域网(“WAN”),网际网(例如,互联网)以及端对端网络(例如,ad hoc端对端网络),以及任何当前已知或未来研发的网络。
上述计算机可读介质可以是上述电子设备中所包含的;也可以是单独存在,而未装配入该电子设备中。
上述计算机可读介质承载有一个或者多个程序,当上述一个或者多个程序被该电子设备执行时,使得该电子设备:
显示用户界面,所述用户界面包括一个或多个第一用户中每个第一用户的标识信息;
获取第二用户针对所述一个或多个第一用户中目标用户的互动指令;
根据所述互动指令所关联的互动内容生成虚拟形象;
将所述虚拟形象发送到所述目标用户的终端。
可选的,当上述一个或者多个程序被该电子设备执行时,该电子设备还可以执行上述实施例所述的其他步骤。
可以以一种或多种程序设计语言或其组合来编写用于执行本公开的操作的计算机程序代码,上述程序设计语言包括但不限于面向对象的程序设计语言—诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络——包括局域网(LAN)或广域网(WAN)—连接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。
附图中的流程图和框图,图示了按照本公开各种实施例的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段、或代码的一部分,该模块、程序段、或代码的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。也应当注意,在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个接连地表示的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。 也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。
描述于本公开实施例中所涉及到的单元可以通过软件的方式实现,也可以通过硬件的方式来实现。其中,单元的名称在某种情况下并不构成对该单元本身的限定。
本文中以上描述的功能可以至少部分地由一个或多个硬件逻辑部件来执行。例如,非限制性地,可以使用的示范类型的硬件逻辑部件包括:现场可编程门阵列(FPGA)、专用集成电路(ASIC)、专用标准产品(ASSP)、片上系统(SOC)、复杂可编程逻辑设备(CPLD)等等。
在本公开的上下文中,机器可读介质可以是有形的介质,其可以包含或存储以供指令执行系统、装置或设备使用或与指令执行系统、装置或设备结合地使用的程序。机器可读介质可以是机器可读信号介质或机器可读储存介质。机器可读介质可以包括但不限于电子的、磁性的、光学的、电磁的、红外的、或半导体系统、装置或设备,或者上述内容的任何合适组合。机器可读存储介质的更具体示例会包括基于一个或多个线的电气连接、便携式计算机盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦除可编程只读存储器(EPROM或快闪存储器)、光纤、便捷式紧凑盘只读存储器(CD-ROM)、光学储存设备、磁储存设备、或上述内容的任何合适组合。
根据本公开的一个或多个实施例,本公开提供了一种互动处理方法,包括:
显示用户界面,所述用户界面包括一个或多个第一用户中每个第一用户的标识信息;
获取第二用户针对所述一个或多个第一用户中目标用户的互动指令;
根据所述互动指令所关联的互动内容生成虚拟形象;
将所述虚拟形象发送到所述目标用户的终端。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述第一用户的标识信息包括如下至少一种:
所述第一用户的图像信息、名称、标识号、所述第一用户的终端标识。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述方法还包括:
在所述用户界面中显示每个第一用户的视频信息;
根据所述第二用户对所述目标用户的视频信息的第一操作,显示与所述目标用户关联的第一操作接口。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述用户界面还包括:与每个第一用户的标识信息分别关联的第一操作接口。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述用户界面还包括:第二操作接口,所述第二操作接口与所述一个或多个第一用户中的部分或全部第一用户的标识信息关联,所述目标用户包括所述部分或全部第一用户。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,在所述获取第二用户针对所述一个或多个第一用户中目标用户的互动指令之前,所述方法还包括:
检测所述第二用户在所述用户界面中对所述第二操作接口和所述部分或全部第一用户 的标识信息的关联操作;
根据所述关联操作建立所述第二操作接口和所述部分或全部第一用户的标识信息的关联关系。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述获取第二用户针对所述一个或多个第一用户中目标用户的互动指令,包括:
检测所述第二用户针对所述目标用户关联的第一操作接口或第二操作接口的第二操作;
根据所述第二操作获取所述第二用户针对所述目标用户的互动指令。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述根据所述第二操作获取所述第二用户针对所述目标用户的互动指令,包括:
根据所述第二操作,在所述用户界面中显示与所述目标用户关联的一个或多个预设文本信息;
检测所述第二用户对所述一个或多个预设文本信息中目标文本信息的选择操作;
根据所述选择操作获取所述第二用户针对所述目标用户的互动指令。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述目标用户关联的第一操作接口或第二操作接口包括一个或多个预设文本信息;
所述检测所述第二用户针对所述目标用户关联的第一操作接口或第二操作接口的第二操作,包括:
检测所述第二用户针对所述一个或多个预设文本信息中目标文本信息的第二操作。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述互动指令所关联的互动内容包括所述目标文本信息。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述根据所述第二操作获取所述第二用户针对所述目标用户的互动指令,包括:
根据所述第二操作,在所述用户界面中显示与所述第一操作接口或所述第二操作接口关联的输入框;
检测所述第二用户在所述输入框中输入目标互动信息的输入操作;
根据所述输入操作获取所述第二用户针对所述目标用户的互动指令。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述第一操作接口或所述第二操作接口包括输入框;
所述检测所述第二用户针对所述目标用户关联的第一操作接口或第二操作接口的第二操作,包括:
检测所述第二用户在所述输入框中输入目标互动信息的输入操作。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述互动指令所关联的互动内容包括所述目标互动信息。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述方法还包括:
响应于所述第二用户对所述预设文本信息的修改指令、增加指令或删除指令,更新所述用户界面中的所述预设文本信息。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,根据所述互动指令所关联的互动内容生成虚拟形象,包括:
获取与所述互动内容对应的音频;
获取与所述音频对应的面部特征点序列;
根据所述面部特征点序列获取视频;
将所述音频和所述视频结合获得虚拟形象。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述方法还包括:
根据所述互动内容的类型信息,确定与所述类型信息对应的表情信息和/或动作信息;
将所述表情信息和/或所述动作信息添加到所述虚拟形象中。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述方法还包括:
对所述互动内容进行情感分析,得到情感信息;
将所述情感信息对应的表情信息和/或动作信息添加到所述虚拟形象中。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述面部特征点序列包括多组面部特征点;
根据所述面部特征点序列获取视频,包括:
获取所述目标用户对应的面部标准模板;
将每一组面部特征点添加到所述面部标准模板,获取每一帧面部图像;
将所述每一帧面部图像按帧顺序依次连接,获取视频。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述方法还包括:
响应于所述第二用户对第一用户的第一设置指令,显示多个预设的面部标准模板;
根据所述第二用户对所述多个预设的面部标准模板中目标模板的选择操作,将所述目标模板设置为所述第一用户对应的面部标准模板。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,所述方法还包括:
接收所述第一用户的终端发送的第二设置指令,所述第二设置指令用于设置所述第一用户对应的面部标准模板。
根据本公开的一个或多个实施例,在本公开提供的互动处理方法中,将所述虚拟形象发送到所述目标用户的终端,包括:
将所述目标用户对应的标记添加到所述虚拟形象;
将添加有所述标记的所述虚拟形象发送到所述目标用户的终端。
根据本公开的一个或多个实施例,本公开提供了一种互动处理装置,包括:
显示模块,用于显示用户界面,所述用户界面包括一个或多个第一用户中每个第一用户的标识信息;
获取模块,用于获取第二用户针对所述一个或多个第一用户中目标用户的互动指令;
生成模块,用于根据所述互动指令所关联的互动内容生成虚拟形象;
发送模块,用于将所述虚拟形象发送到所述目标用户的终端。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,所述第一用户的标识信息包括如下至少一种:所述第一用户的图像信息、名称、标识号、所述第一用户的 终端标识。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,显示模块还用于:在所述用户界面中显示每个第一用户的视频信息;根据所述第二用户对所述目标用户的视频信息的第一操作,显示与所述目标用户关联的第一操作接口。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,所述用户界面还包括:与每个第一用户的标识信息分别关联的第一操作接口。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,所述用户界面还包括:第二操作接口,所述第二操作接口与所述一个或多个第一用户中的部分或全部第一用户的标识信息关联,所述目标用户包括所述部分或全部第一用户。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,该互动处理装置还包括:检测模块和关联关系建立模块,其中,检测模块用于获取模块获取第二用户针对所述一个或多个第一用户中目标用户的互动指令之前,检测所述第二用户在所述用户界面中对所述第二操作接口和所述部分或全部第一用户的标识信息的关联操作;关联关系建立模块用于根据所述关联操作建立所述第二操作接口和所述部分或全部第一用户的标识信息的关联关系。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,检测模块还用于:检测所述第二用户针对所述目标用户关联的第一操作接口或第二操作接口的第二操作;获取模块还用于:根据所述第二操作获取所述第二用户针对所述目标用户的互动指令。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,显示模块还用于:根据所述第二操作,在所述用户界面中显示与所述目标用户关联的一个或多个预设文本信息;检测模块还用于:检测所述第二用户对所述一个或多个预设文本信息中目标文本信息的选择操作;获取模块还用于:根据所述选择操作获取所述第二用户针对所述目标用户的互动指令。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,所述目标用户关联的第一操作接口或第二操作接口包括一个或多个预设文本信息;检测模块具体用于:检测所述第二用户针对所述一个或多个预设文本信息中目标文本信息的第二操作。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,所述互动指令所关联的互动内容包括所述目标文本信息。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,显示模块还用于:根据所述第二操作,在所述用户界面中显示与所述第一操作接口或所述第二操作接口关联的输入框;检测模块还用于:检测所述第二用户在所述输入框中输入目标互动信息的输入操作;获取模块还用于:根据所述输入操作获取所述第二用户针对所述目标用户的互动指令。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,所述第一操作接口或所述第二操作接口包括输入框;检测模块具体用于:检测所述第二用户在所述输入框中输入目标互动信息的输入操作。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,所述互动指令所 关联的互动内容包括所述目标互动信息。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,该互动处理装置还包括:更新模块,用于响应于所述第二用户对所述预设文本信息的修改指令、增加指令或删除指令,更新所述用户界面中的所述预设文本信息。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,获取模块还用于:获取与所述互动内容对应的音频;获取与所述音频对应的面部特征点序列;根据所述面部特征点序列获取视频;生成模块具体用于:将所述音频和所述视频结合获得虚拟形象。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,该互动处理装置还包括:确定模块和添加模块,其中,确定模块用于根据所述互动内容的类型信息,确定与所述类型信息对应的表情信息和/或动作信息;添加模块用于将所述表情信息和/或所述动作信息添加到所述虚拟形象中。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,确定模块还用于对所述互动内容进行情感分析,得到情感信息;添加模块还用于:将所述情感信息对应的表情信息和/或动作信息添加到所述虚拟形象中。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,所述面部特征点序列包括多组面部特征点;获取模块在根据所述面部特征点序列获取视频时,具体用于:获取所述目标用户对应的面部标准模板;将每一组面部特征点添加到所述面部标准模板,获取每一帧面部图像;将所述每一帧面部图像按帧顺序依次连接,获取视频。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,该互动处理装置还包括:设置模块;显示模块还用于:响应于所述第二用户对第一用户的第一设置指令,显示多个预设的面部标准模板;设置模块用于根据所述第二用户对所述多个预设的面部标准模板中目标模板的选择操作,将所述目标模板设置为所述第一用户对应的面部标准模板。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,该互动处理装置还包括:接收模块,用于接收所述第一用户的终端发送的第二设置指令,所述第二设置指令用于设置所述第一用户对应的面部标准模板。
根据本公开的一个或多个实施例,在本公开提供的互动处理装置中,添加模块还用于将所述目标用户对应的标记添加到所述虚拟形象;发送模块具体用于:将添加有所述标记的所述虚拟形象发送到所述目标用户的终端。
根据本公开的一个或多个实施例,本公开提供了一种电子设备,包括:
一个或多个处理器;
存储器,用于存储一个或多个程序;
当所述一个或多个程序被所述一个或多个处理器执行,使得所述一个或多个处理器实现如本公开提供的任一所述的互动处理方法。
根据本公开的一个或多个实施例,本公开提供了一种计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时实现如本公开提供的任一所述的互动处理方法。
以上描述仅为本公开的较佳实施例以及对所运用技术原理的说明。本领域技术人员应当理解,本公开中所涉及的公开范围,并不限于上述技术特征的特定组合而成的技术方案, 同时也应涵盖在不脱离上述公开构思的情况下,由上述技术特征或其等同特征进行任意组合而形成的其它技术方案。例如上述特征与本公开中公开的(但不限于)具有类似功能的技术特征进行互相替换而形成的技术方案。
此外,虽然采用特定次序描绘了各操作,但是这不应当理解为要求这些操作以所示出的特定次序或以顺序次序执行来执行。在一定环境下,多任务和并行处理可能是有利的。同样地,虽然在上面论述中包含了若干具体实现细节,但是这些不应当被解释为对本公开的范围的限制。在单独的实施例的上下文中描述的某些特征还可以组合地实现在单个实施例中。相反地,在单个实施例的上下文中描述的各种特征也可以单独地或以任何合适的子组合的方式实现在多个实施例中。
尽管已经采用特定于结构特征和/或方法逻辑动作的语言描述了本主题,但是应当理解所附权利要求书中所限定的主题未必局限于上面描述的特定特征或动作。相反,上面所描述的特定特征和动作仅仅是实现权利要求书的示例形式。

Claims (24)

  1. 一种互动处理方法,其特征在于,所述方法包括:
    显示用户界面,所述用户界面包括一个或多个第一用户中每个第一用户的标识信息;
    获取第二用户针对所述一个或多个第一用户中目标用户的互动指令;
    根据所述互动指令所关联的互动内容生成虚拟形象;
    将所述虚拟形象发送到所述目标用户的终端。
  2. 根据权利要求1所述的方法,其特征在于,所述第一用户的标识信息包括如下至少一种:
    所述第一用户的图像信息、名称、标识号、所述第一用户的终端标识。
  3. 根据权利要求1所述的方法,其特征在于,所述方法还包括:
    在所述用户界面中显示每个第一用户的视频信息;
    根据所述第二用户对所述目标用户的视频信息的第一操作,显示与所述目标用户关联的第一操作接口。
  4. 根据权利要求1所述的方法,其特征在于,所述用户界面还包括:与每个第一用户的标识信息分别关联的第一操作接口。
  5. 根据权利要求1所述的方法,其特征在于,所述用户界面还包括:第二操作接口,所述第二操作接口与所述一个或多个第一用户中的部分或全部第一用户的标识信息关联,所述目标用户包括所述部分或全部第一用户。
  6. 根据权利要求5所述的方法,其特征在于,在所述获取第二用户针对所述一个或多个第一用户中目标用户的互动指令之前,所述方法还包括:
    检测所述第二用户在所述用户界面中对所述第二操作接口和所述部分或全部第一用户的标识信息的关联操作;
    根据所述关联操作建立所述第二操作接口和所述部分或全部第一用户的标识信息的关联关系。
  7. 根据权利要求3-5任一项所述的方法,其特征在于,所述获取第二用户针对所述一个或多个第一用户中目标用户的互动指令,包括:
    检测所述第二用户针对所述目标用户关联的第一操作接口或第二操作接口的第二操作;
    根据所述第二操作获取所述第二用户针对所述目标用户的互动指令。
  8. 根据权利要求7所述的方法,其特征在于,所述根据所述第二操作获取所述第二用户针对所述目标用户的互动指令,包括:
    根据所述第二操作,在所述用户界面中显示与所述目标用户关联的一个或多个预设文本信息;
    检测所述第二用户对所述一个或多个预设文本信息中目标文本信息的选择操作;
    根据所述选择操作获取所述第二用户针对所述目标用户的互动指令。
  9. 根据权利要求7所述的方法,其特征在于,所述目标用户关联的第一操作接口或第二操作接口包括一个或多个预设文本信息;
    所述检测所述第二用户针对所述目标用户关联的第一操作接口或第二操作接口的第二操作,包括:
    检测所述第二用户针对所述一个或多个预设文本信息中目标文本信息的第二操作。
  10. 根据权利要求8或9所述的方法,其特征在于,所述互动指令所关联的互动内容包括所述目标文本信息。
  11. 根据权利要求7所述的方法,其特征在于,所述根据所述第二操作获取所述第二用户针对所述目标用户的互动指令,包括:
    根据所述第二操作,在所述用户界面中显示与所述第一操作接口或所述第二操作接口关联的输入框;
    检测所述第二用户在所述输入框中输入目标互动信息的输入操作;
    根据所述输入操作获取所述第二用户针对所述目标用户的互动指令。
  12. 根据权利要求7所述的方法,其特征在于,所述第一操作接口或所述第二操作接口包括输入框;
    所述检测所述第二用户针对所述目标用户关联的第一操作接口或第二操作接口的第二操作,包括:
    检测所述第二用户在所述输入框中输入目标互动信息的输入操作。
  13. 根据权利要求11或12所述的方法,其特征在于,所述互动指令所关联的互动内容包括所述目标互动信息。
  14. 根据权利要求8或9所述的方法,其特征在于,所述方法还包括:
    响应于所述第二用户对所述预设文本信息的修改指令、增加指令或删除指令,更新所述用户界面中的所述预设文本信息。
  15. 根据权利要求1所述的方法,其特征在于,根据所述互动指令所关联的互动内容生成虚拟形象,包括:
    获取与所述互动内容对应的音频;
    获取与所述音频对应的面部特征点序列;
    根据所述面部特征点序列获取视频;
    将所述音频和所述视频结合获得虚拟形象。
  16. 根据权利要求15所述的方法,其特征在于,所述方法还包括:
    根据所述互动内容的类型信息,确定与所述类型信息对应的表情信息和/或动作信息;
    将所述表情信息和/或所述动作信息添加到所述虚拟形象中。
  17. 根据权利要求15所述的方法,其特征在于,所述方法还包括:
    对所述互动内容进行情感分析,得到情感信息;
    将所述情感信息对应的表情信息和/或动作信息添加到所述虚拟形象中。
  18. 根据权利要求15所述的方法,其特征在于,所述面部特征点序列包括多组面部特征点;
    根据所述面部特征点序列获取视频,包括:
    获取所述目标用户对应的面部标准模板;
    将每一组面部特征点添加到所述面部标准模板,获取每一帧面部图像;
    将所述每一帧面部图像按帧顺序依次连接,获取视频。
  19. 根据权利要求18所述的方法,其特征在于,所述方法还包括:
    响应于所述第二用户对第一用户的第一设置指令,显示多个预设的面部标准模板;
    根据所述第二用户对所述多个预设的面部标准模板中目标模板的选择操作,将所述目标模板设置为所述第一用户对应的面部标准模板。
  20. 根据权利要求18所述的方法,其特征在于,所述方法还包括:
    接收所述第一用户的终端发送的第二设置指令,所述第二设置指令用于设置所述第一用户对应的面部标准模板。
  21. 根据权利要求1所述的方法,其特征在于,将所述虚拟形象发送到所述目标用户的终端,包括:
    将所述目标用户对应的标记添加到所述虚拟形象;
    将添加有所述标记的所述虚拟形象发送到所述目标用户的终端。
  22. 一种互动处理装置,其特征在于,包括:
    显示模块,用于显示用户界面,所述用户界面包括一个或多个第一用户中每个第一用户的标识信息;
    获取模块,用于获取第二用户针对所述一个或多个第一用户中目标用户的互动指令;
    生成模块,用于根据所述互动指令所关联的互动内容生成虚拟形象;
    发送模块,用于将所述虚拟形象发送到所述目标用户的终端。
  23. 一种电子设备,其特征在于,所述电子设备包括:
    一个或多个处理器;
    存储装置,用于存储一个或多个程序;
    当所述一个或多个程序被所述一个或多个处理器执行,使得所述一个或多个处理器实现如权利要求1-21中任一所述的方法。
  24. 一种计算机可读存储介质,其上存储有计算机程序,其特征在于,该程序被处理器执行时实现如权利要求1-21中任一所述的方法。
PCT/CN2021/123202 2020-10-28 2021-10-12 一种互动处理方法、装置、电子设备和存储介质 WO2022089192A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011175034.6 2020-10-28
CN202011175034.6A CN112286610A (zh) 2020-10-28 2020-10-28 一种互动处理方法、装置、电子设备和存储介质

Publications (1)

Publication Number Publication Date
WO2022089192A1 true WO2022089192A1 (zh) 2022-05-05

Family

ID=74372429

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/123202 WO2022089192A1 (zh) 2020-10-28 2021-10-12 一种互动处理方法、装置、电子设备和存储介质

Country Status (2)

Country Link
CN (1) CN112286610A (zh)
WO (1) WO2022089192A1 (zh)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115113781A (zh) * 2022-06-28 2022-09-27 广州博冠信息科技有限公司 互动图标显示方法、装置、介质与电子设备
CN115412518A (zh) * 2022-08-19 2022-11-29 网易传媒科技(北京)有限公司 表情发送方法、装置、存储介质及电子设备
CN116248812A (zh) * 2023-05-11 2023-06-09 广州佰锐网络科技有限公司 一种基于数字人交互视频的业务办理方法、存储介质及系统

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112286610A (zh) * 2020-10-28 2021-01-29 北京有竹居网络技术有限公司 一种互动处理方法、装置、电子设备和存储介质
KR20230071943A (ko) * 2021-11-17 2023-05-24 주식회사 제네시스랩 인터랙티브 콘텐츠를 제공하기 위한 방법, 시스템 및 비일시성의 컴퓨터 판독 가능 기록 매체
KR20230071944A (ko) * 2021-11-17 2023-05-24 주식회사 제네시스랩 인터랙티브 콘텐츠를 제공하기 위한 방법, 시스템 및 비일시성의 컴퓨터 판독 가능 기록 매체
CN115097984A (zh) * 2022-06-22 2022-09-23 北京字跳网络技术有限公司 交互方法、装置、电子设备和存储介质

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101968932A (zh) * 2010-09-19 2011-02-09 深圳市摩拓触摸科技有限公司 多媒体智能互动教学系统及其实现方法
US20150279222A1 (en) * 2014-03-31 2015-10-01 Konica Minolta Laboratory U.S.A., Inc. Method and system for enhancing interactions between teachers and students
CN107316520A (zh) * 2017-08-17 2017-11-03 广州视源电子科技股份有限公司 视频教学互动方法、装置、设备及存储介质
CN110850983A (zh) * 2019-11-13 2020-02-28 腾讯科技(深圳)有限公司 视频直播中的虚拟对象控制方法、装置和存储介质
CN111290568A (zh) * 2018-12-06 2020-06-16 阿里巴巴集团控股有限公司 交互方法、装置及计算机设备
CN112286610A (zh) * 2020-10-28 2021-01-29 北京有竹居网络技术有限公司 一种互动处理方法、装置、电子设备和存储介质

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110874557B (zh) * 2018-09-03 2023-06-16 阿里巴巴集团控股有限公司 一种语音驱动虚拟人脸的视频生成方法以及装置
CN109615961A (zh) * 2019-01-31 2019-04-12 华中师范大学 一种课堂教学师生互动网络系统与方法
CN110766777B (zh) * 2019-10-31 2023-09-29 北京字节跳动网络技术有限公司 虚拟形象的生成方法、装置、电子设备及存储介质
CN111126980B (zh) * 2019-12-30 2023-09-12 腾讯科技(深圳)有限公司 虚拟物品发送方法、处理方法、装置、设备及介质
CN111369967B (zh) * 2020-03-11 2021-03-05 北京字节跳动网络技术有限公司 基于虚拟人物的语音合成方法、装置、介质及设备

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101968932A (zh) * 2010-09-19 2011-02-09 深圳市摩拓触摸科技有限公司 多媒体智能互动教学系统及其实现方法
US20150279222A1 (en) * 2014-03-31 2015-10-01 Konica Minolta Laboratory U.S.A., Inc. Method and system for enhancing interactions between teachers and students
CN107316520A (zh) * 2017-08-17 2017-11-03 广州视源电子科技股份有限公司 视频教学互动方法、装置、设备及存储介质
CN111290568A (zh) * 2018-12-06 2020-06-16 阿里巴巴集团控股有限公司 交互方法、装置及计算机设备
CN110850983A (zh) * 2019-11-13 2020-02-28 腾讯科技(深圳)有限公司 视频直播中的虚拟对象控制方法、装置和存储介质
CN112286610A (zh) * 2020-10-28 2021-01-29 北京有竹居网络技术有限公司 一种互动处理方法、装置、电子设备和存储介质

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115113781A (zh) * 2022-06-28 2022-09-27 广州博冠信息科技有限公司 互动图标显示方法、装置、介质与电子设备
CN115412518A (zh) * 2022-08-19 2022-11-29 网易传媒科技(北京)有限公司 表情发送方法、装置、存储介质及电子设备
CN116248812A (zh) * 2023-05-11 2023-06-09 广州佰锐网络科技有限公司 一种基于数字人交互视频的业务办理方法、存储介质及系统
CN116248812B (zh) * 2023-05-11 2023-08-08 广州佰锐网络科技有限公司 一种基于数字人交互视频的业务办理方法、存储介质及系统

Also Published As

Publication number Publication date
CN112286610A (zh) 2021-01-29

Similar Documents

Publication Publication Date Title
WO2022089192A1 (zh) 一种互动处理方法、装置、电子设备和存储介质
CN110570698B (zh) 一种在线教学控制方法、装置、存储介质以及终端
US20200219216A1 (en) A Digital Classroom With a Breakout Feature
WO2021098571A1 (zh) 基于在线文档评论的反馈方法、装置、设备及存储介质
US11715386B1 (en) Queuing for a video conference session
EP4047490A1 (en) Video-based interaction realization method and apparatus, device and medium
WO2019033663A1 (zh) 视频教学互动方法、装置、设备及存储介质
CN113741765B (zh) 页面跳转方法、装置、设备、存储介质和程序产品
WO2022017203A1 (zh) 听写交互方法、装置和电子设备
US11886484B2 (en) Music playing method and apparatus based on user interaction, and device and storage medium
WO2020220782A1 (zh) 信息分享方法、装置、设备和介质
CN111260975B (zh) 用于多媒体黑板教学互动的方法、装置、介质和电子设备
CN111862705A (zh) 一种提示直播教学目标的方法、装置、介质和电子设备
KR102101817B1 (ko) 인터넷 상에 공개된 동영상을 이용한 어학 학습 지원 장치
WO2024037480A1 (zh) 交互方法、装置、电子设备和存储介质
WO2023134558A1 (zh) 交互方法、装置、电子设备、存储介质和程序产品
JP2019215502A (ja) サーバ、音データ評価方法、プログラム、通信システム
US20220215482A1 (en) User information processing method, apparatus, electronic device, and computer storage medium
CN112165627A (zh) 信息处理方法、装置、存储介质、终端及系统
JP2022534345A (ja) データ処理方法及び装置、電子機器並びに記憶媒体
CN112863277B (zh) 一种直播教学的互动方法、装置、介质和电子设备
CN111369848B (zh) 基于课件内容互动的方法、装置、存储介质和电子设备
US20220150290A1 (en) Adaptive collaborative real-time remote remediation
US20240094980A1 (en) Information processing apparatus, information processing system, non-transitory computer readable medium, and information processing method
WO2023109671A1 (zh) 直播信息处理方法、装置、设备和存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21884921

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21884921

Country of ref document: EP

Kind code of ref document: A1