CN111601064B - Information interaction method and information interaction device - Google Patents

Information interaction method and information interaction device Download PDF

Info

Publication number
CN111601064B
CN111601064B CN202010421274.3A CN202010421274A CN111601064B CN 111601064 B CN111601064 B CN 111601064B CN 202010421274 A CN202010421274 A CN 202010421274A CN 111601064 B CN111601064 B CN 111601064B
Authority
CN
China
Prior art keywords
information
user
screen
identifier
sight
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010421274.3A
Other languages
Chinese (zh)
Other versions
CN111601064A (en
Inventor
马全迎
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vivo Mobile Communication Co Ltd
Original Assignee
Vivo Mobile Communication Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vivo Mobile Communication Co Ltd filed Critical Vivo Mobile Communication Co Ltd
Priority to CN202010421274.3A priority Critical patent/CN111601064B/en
Publication of CN111601064A publication Critical patent/CN111601064A/en
Application granted granted Critical
Publication of CN111601064B publication Critical patent/CN111601064B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/141Systems for two-way working between two video terminals, e.g. videophone
    • H04N7/142Constructional details of the terminal equipment, e.g. arrangements of the camera and the display
    • H04N7/144Constructional details of the terminal equipment, e.g. arrangements of the camera and the display camera and display on the same optical axis, e.g. optically multiplexing the camera and display for eye to eye contact
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • General Engineering & Computer Science (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The application discloses an information interaction method and an information interaction device, belongs to the technical field of communication, and can solve the problem that the interaction effect of video call is poor. The method comprises the following steps: under the condition that the first device and the second device are in a video call state, acquiring a first time length, wherein the first time length is the time length of the deviation of the sight of the first user from the screen of the first device; and outputting the first information and sending the second information to the second equipment under the condition that the first duration is greater than or equal to a first preset duration. The first information is used for prompting the first user to adjust the sight line direction, the second information is used for indicating the second equipment to update the target identifier from the first display mode to the second display mode, and the target identifier is an identifier corresponding to the first user.

Description

Information interaction method and information interaction device
Technical Field
The application belongs to the technical field of communication, and particularly relates to an information interaction method and an information interaction device.
Background
With the development of communication technology, electronic devices are more widely used.
In general, different users may use different devices for video calls, such as online teaching, etc. Taking online teaching as an example, a teacher uses the teacher-side device to attend a class for a student using the student-side device. However, the size of the screen of the teacher-side device is limited, and the number of students participating in online teaching is large, so that when a teacher uses the teacher-side device to give a lecture, the teacher-side device not only needs to display teaching contents, but also needs to display the head portraits of some students, so that when the students do not pay attention to the teaching contents, the teacher cannot timely remind the students to listen to the lectures seriously, the teaching quality is reduced, and the interactive effect of video call is poor.
Disclosure of Invention
The embodiment of the application aims to provide an information interaction method and an information interaction device, which can solve the problem of poor interaction effect of video call.
In order to solve the technical problem, the present application is implemented as follows:
in a first aspect, an embodiment of the present application provides an information interaction method, where the method includes: under the condition that the first device and the second device are in a video call state, acquiring a first time length, wherein the first time length is the time length of the deviation of the sight of the first user from the screen of the first device; and outputting the first information and sending the second information to the second equipment under the condition that the first duration is greater than or equal to a first preset duration. The first information is used for prompting the first user to adjust the sight line direction, the second information is used for indicating the second equipment to update the target identifier from the first display mode to the second display mode, and the target identifier is an identifier corresponding to the first user.
In a second aspect, an embodiment of the present application provides an information interaction method, where the method includes: under the condition that the first device and the second device are in a video call state, acquiring a second time length, wherein the second time length is the time length when the sight of a second user points to a target identifier in a screen of the second device, and the target identifier is an identifier corresponding to a first user using the first device; and under the condition that the second duration is greater than or equal to a second preset duration, updating the target identifier from the first display mode to a second display mode, and sending third information to the first equipment. And the third information is used for prompting the first user to adjust the sight line direction.
In a third aspect, an embodiment of the present application provides an information interaction apparatus. The information interaction device comprises: the device comprises an acquisition module, a processing module and a sending module. The acquisition module is used for acquiring a first duration under the condition that the first device and the second device are in a video call state, wherein the first duration is the duration that the sight of the first user deviates from the screen of the first device. And the processing module is used for outputting the first information and controlling the sending module to send the second information to the second equipment under the condition that the first duration acquired by the acquisition module is greater than or equal to a first preset duration. The first information is used for prompting the first user to adjust the sight line direction, the second information is used for indicating the second equipment to update the target identifier from the first display mode to the second display mode, and the target identifier is an identifier corresponding to the first user.
In a fourth aspect, an embodiment of the present application provides an information interaction apparatus. The information interaction device comprises an acquisition module, a display module and a sending module. The acquisition module is used for acquiring a second time length when the first device and the second device are in a video call state, wherein the second time length is the time length when the sight of the second user points to a target identifier in a screen of the second device, and the target identifier is an identifier corresponding to a first user using the first device. And the display module is used for updating the target identifier from the first display mode to the second display mode under the condition that the second duration acquired by the acquisition module is greater than or equal to a second preset duration. And the sending module is used for sending third information to the first equipment under the condition that the second duration is greater than or equal to a second preset duration. Wherein the third information is used for prompting the first user to adjust the sight line direction.
In a fifth aspect, embodiments of the present application provide an electronic device, which includes a processor, a memory, and a program or instructions stored on the memory and executable on the processor, where the program or instructions, when executed by the processor, implement the steps of the method as in the first aspect or the second aspect.
In a sixth aspect, embodiments of the present application provide a readable storage medium on which a program or instructions are stored, which when executed by a processor implement the steps of the method as in the first or second aspect.
In a seventh aspect, an embodiment of the present application provides a chip, where the chip includes a processor and a communication interface, where the communication interface is coupled to the processor, and the processor is configured to execute a program or instructions to implement the method in the first aspect or the second aspect.
For the information interaction method provided by the first aspect, in the process of performing a video call between the first device and the second device, the first device may detect that the line of sight of a first user using the first device deviates from the screen of the first device, and acquire a first duration for which the line of sight of the first user deviates from the screen of the first device. Therefore, in the case that the first duration is greater than or equal to the first preset duration, on the one hand, the first device may output the first information, so that the first user may adjust the gaze direction according to the first information, i.e., the first user refocuses the attention on the screen of the first device; on the other hand, the first device may send the second information to the second device, so that the second device may highlight the target identifier corresponding to the first user, and further, the second user using the second device may remind the first user to focus on the screen of the first device again. In this way, the interactive performance of the first device and the second device is enhanced, so that the quality of the video call is improved (for example, the live teaching quality is improved).
For the information interaction method provided in the second aspect, when the first device and the second device are in a video call state, the second device may detect that the gaze of the second user using the second device is directed to the target identifier in the screen of the second device, and obtain a second duration that the gaze of the second user is directed to the target identifier in the screen of the second device. Therefore, in the case that the second duration is greater than or equal to the second preset duration, on one hand, the second device may highlight the target identifier corresponding to the first user, so that the second user may pay more attention to whether the first user corresponding to the target identifier focuses on; on the other hand, the second device may send third information to the first device, so that the first device may remind the first user of adjusting the sight line in time according to the third information. Therefore, the interactive performance of the first device and the second device is improved, and the video call quality is improved.
Drawings
Fig. 1 is a schematic diagram of an information interaction system according to an embodiment of the present application;
fig. 2 is a schematic diagram of an information interaction method according to an embodiment of the present application;
fig. 3 is a schematic view of a human face sight line provided in an embodiment of the present application;
fig. 4 is a second schematic diagram of an information interaction method according to an embodiment of the present application;
fig. 5 is a third schematic diagram of an information interaction method according to an embodiment of the present application;
fig. 6 is a schematic structural diagram of an information interaction device according to an embodiment of the present application;
fig. 7 is a second schematic structural diagram of an information interaction device according to an embodiment of the present application;
fig. 8 is a hardware schematic diagram of an electronic device according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some, but not all, embodiments of the present application. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
The terms first, second and the like in the description and in the claims of the present application are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the application are capable of operation in sequences other than those illustrated or described herein. In addition, "and/or" in the specification and claims means at least one of connected objects, a character "/" generally means that a preceding and succeeding related objects are in an "or" relationship.
In the embodiments of the present application, words such as "exemplary" or "for example" are used to mean serving as an example, instance, or illustration. Any embodiment or design described herein as "exemplary" or "e.g.," should not be construed as advantageous over other embodiments or designs. Rather, use of the word "exemplary" or "such as" is intended to present concepts related in a concrete fashion.
In the description of the embodiments of the present application, unless otherwise specified, "a plurality" means two or more, for example, a plurality of elements means two or more elements, and the like.
Some terms/nouns referred to in the embodiments of the present application are explained below.
open source computer vision library, OpenCV: is a cross-platform computer vision library. The method is used for realizing the artificial intelligence portrait recognition technology.
socket link: the transport layer implements the network links for end-to-end communication. The method and the device can be used for the second equipment to send information to the first equipment.
java script object notification: also known as JSON, is used to represent a lightweight data exchange format. In this embodiment of the application, the format of the information sent by the second device to the first device may be a JSON format.
As shown in fig. 1, the information interaction system provided in the embodiment of the present application may include at least a cloud server a, a first device B, and a second device C. It should be noted that fig. 1 illustrates an example that the information interaction system includes one first device, and in actual implementation, the number of the first devices may be multiple; the connection mode between the devices in fig. 1 may be a wired connection or a wireless connection; in addition, the first device and the second device in the embodiment of the present application may be the same type of electronic device, or different types of electronic devices. The method can be determined according to actual conditions, and is not limited in the embodiment of the application.
In the case where the first device and the second device establish a video call, the data stream transmitted by the first device B or the second device C may be transmitted to the counterpart device after being processed by the cloud server 10. Wherein, the data stream transmitted by the first device B may include at least one of the following: the live data stream and the video data stream collected by the camera of the first device B, and the data stream sent by the second device C may include at least one of the following: live broadcast data stream and video data stream collected by the camera of the second device C. Cloud server a may be used to process and compute live and video data streams.
Specifically, in order to more clearly illustrate the information interaction method provided in the embodiment of the present application, the following provides three possible application scenarios of the information interaction method:
(1) the method is applied to an online teaching scene. Taking the first device shown in fig. 1 as a student-side device and the second device as a teacher-side device as an example, the first user using the first device is a student, and the second user using the second device is a teacher. One way is that when the attention of the student is dispersed, the student end device can remind the student to concentrate the attention, and the teacher can learn that the student does not hear the talk seriously by sending information to the teacher end device; in another mode, when a teacher gazes that the sight line of a certain student in the teacher-side equipment is not on the screen of the student-side equipment for a long time, the student-side equipment can receive information sent by the teacher-side equipment and remind the student to concentrate on attention and listen to the speaker. Therefore, live broadcast teaching quality is improved.
(2) The method is applied to online live broadcast scenes. Taking the first device as the seller device and the second device as the buyer device as shown in fig. 1 as an example, the first user using the seller device is the seller and the second user using the buyer device is the buyer. When the buyer watches the goods which are sold in a live broadcast mode through the buyer device on the buyer device, if the buyer happens not to be in the direct broadcast mode of the buyer device due to other things when the seller device is in the direct broadcast mode, the buyer device can output a message (ringing, voice prompt and the like) to remind the buyer to return to the direct broadcast mode of the buyer device, so that the buyer does not miss the opportunity of buying the direct broadcast sold goods of the buyer device. Thus, live broadcast quality is improved.
(3) The method is applied to online video conferences. Taking the first device as the main speaking device and the second device as the participant device as an example, the first user using the first device is the host, and the second user using the second device is the participant. When a host uses a talkback device and participants use a participant device to carry out a video conference, one way is that when the attention of the participants is dispersed, the participant device can remind the participants to concentrate on the attention, and the host knows that the participants do not participate in the conference seriously by sending information to the talkback device; the other mode is that when the host watches that the sight of a certain participant in the talkback equipment is not on the screen of the participant equipment for a long time, the participant equipment can receive the information sent by the talkback equipment and remind the participant to focus attention to listen to the host for speaking. Therefore, live broadcast teaching quality is improved.
The embodiment of the application provides two information interaction methods. The following describes in detail an information interaction method provided by the embodiments of the present application with reference to the accompanying drawings, in which the following embodiments a, b and their respective application scenarios are described.
Example one
As shown in fig. 2, an embodiment of the present application provides an information interaction method, which is applied to a first device and a second device shown in fig. 1. The information interaction method may include S201 to S203 described below.
S201, under the condition that the first device and the second device are in a video call state, the first device obtains a first duration.
The first duration may be a duration in which the line of sight of the first user deviates from the screen of the first device.
It should be noted that, in this embodiment of the present application, that "the first device and the second device are in a video call state" specifically means: the first application program in the first device and the second application program in the second device are both in a foreground running state, and the first device establishes a video call through the first application program and the second device through the second application program. The video pictures in the first application program in the first device may at least include video pictures acquired by a camera of at least one second device, and the video pictures in the second application program in the second device may at least include video pictures acquired by a camera of at least one second device. The first application and the second application may be the same type of application or different types of applications.
Optionally, the video frame in the first application of the first device may further include a first file. The first file may be a file provided by the second device or a file provided by a second application program in the second device.
Optionally, the video frame in the second application of the second device may further include a second file. The second file may be a file stored by the second device or a file provided by a second application program in the second device.
Optionally, the first application program and the second application program may be live broadcast application programs or other application programs with a video call function.
It should be noted that, when the first device starts the first application program, the second device starts the second application program, and the first device and the second device are in a video call state, the first device may collect a face image or an eye image of a first user using the first device through a camera of the first device, and calculate related parameter information of a sight line of the first user based on an artificial intelligence portrait recognition technology, a portrait motion tracking technology, an artificial intelligence sight line algorithm detection technology, and the like, so as to determine whether the sight line of the first user deviates from a screen of the first device.
The artificial intelligence portrait recognition technology in the embodiment of the application at least comprises a face detection algorithm, a face recognition algorithm and the like. Therefore, by applying a face detection algorithm and a face recognition algorithm and calculating a video image collected by a camera of the electronic equipment, information such as a face contour, facial feature points and the like is obtained. After the electronic device detects the information of key points of the face, such as the left canthus, the right canthus, the nose tip, the left mouth corner, the right mouth corner, the lower jaw and the like, the rotation vector can be solved by adopting a solvePnP function of OpenCV, and finally the rotation vector is converted into an euler angle to be used as a face angle reference. If the face angle is greater than the threshold angle, the user is deemed not to be on the screen of the electronic device, i.e., the user's gaze is offset from the screen of the electronic device.
The portrait motion tracking technology obtains information such as a face contour, facial feature points and the like based on an artificial intelligent portrait recognition technology, and determines whether the face is in motions such as turning, shaking, lowering and the like. The portrait motion tracking technology can be realized through a preset portrait motion model.
Optionally, in the process of performing a video call between the first device and the second device, the second interface of the screen of the second device may display not only the second file, but also the portrait identifier and the graph corresponding to the portrait motion model. The portrait identifier may be a real head portrait of a first user using the first device, and the graph corresponding to the portrait motion model may be added or subtracted according to a requirement of the user.
The artificial intelligence sight line algorithm is a method for determining the sight line angle of a person according to picture frames acquired by a camera. The artificial intelligence sight line algorithm is realized based on an artificial intelligence portrait recognition technology. Specifically, the artificial intelligence sight line algorithm comprises three steps of face detection, eye region interception and sight line angle calculation.
(1) Face detection:
in the embodiment of the application, the face detection is a face detection function realized based on the OpenCV.
(2) Intercepting an eye region:
firstly, after the whole face image is obtained, images of left and right eye regions can be intercepted from the whole face image. Then, the regions and positions of the left and right eyes are detected from the images of the left and right eye regions by the face detection function.
(3) And (3) calculating the sight angle:
the sight angle is an included angle between the sight of the user and a screen of the electronic equipment.
The pupil position is determined by using OpenCV, the positions of the pupil and the face are positioned according to real-time data frame analysis, and the sight line movement angle is calculated according to the positions of the face and the pupil.
As shown in fig. 3 (a), if the distances A1B1 and B1C1 from the center points of the left and right pupils of the user to the left and right corner center intersection point B1 are equal when the intermediate points A1 and C1 of the left and right pupils are near the center of the eye region, the user is considered to be directly sighted, that is, the user's gaze is not off the screen. The angle of the movement of the line of sight is calculated based on this.
As shown in (B) of fig. 3, when the side of the user points at the screen of the electronic device, the center points A2, C2 of the left and right pupils of the user are not equal to the distances A2B2 and B2C2 from the left and right eye corner center intersection point B2, respectively, so that an offset can be obtained and the gaze angle of the user can be calculated from parameters such as the offset. The electronic equipment judges the distance between the pupil center points of the left eyeball and the right eyeball of the user and the boundary of the facial feature, and if the boundary of the pupils deflected to a certain direction exceeds a set threshold, the electronic equipment determines that the sight of the user is not on the screen of the electronic equipment; if the boundary of the pupil deviation in a certain direction does not exceed the set threshold, the electronic device determines that the sight line of the user is still on the screen of the electronic device, namely the sight line of the user points to the screen of the electronic device. The set threshold value can be dynamically calculated according to different face information and the distance between the left and right canthus.
Optionally, for specific implementation of the artificial intelligence portrait identification technology, the portrait motion tracking technology, and the artificial intelligence sight line algorithm, reference may be made to the content of the related technology, which is not described herein again.
Optionally, the electronic device determining that the line of sight of the user is still on the screen of the electronic device includes the following two cases:
(1) the user's line of sight is looking directly at the screen of the electronic device, i.e. there is no offset in the user's line of sight.
(2) The user's gaze is offset from direct view, i.e., the user's side face is pointing towards the screen of the electronic device, but the user's gaze angle is not above a set threshold.
When the first device calculates that the sight angle of the first user is larger than a set threshold value of the first device (namely, exceeds the set threshold value) through an artificial intelligence sight algorithm, the first device judges that the sight of the first user deviates from a screen of the first device.
It will be appreciated that in the event that the first device determines that the video of the first user deviates from the screen of the first device, the first device may initiate a timer to count for a first duration, i.e., a duration that the first user's gaze deviates from the screen of the first device.
S202, under the condition that the first duration is greater than or equal to a first preset duration, the first equipment outputs first information and sends second information to the second equipment.
The first information may be used to prompt the first user to adjust the gaze direction, and the second information may be used to instruct the second device to update the target identifier from the first display mode to the second display mode, where the target identifier is an identifier corresponding to the first user.
Optionally, the target identifier may include a face image of the first user acquired by a camera of the first device in real time, that is, the target identifier may be updated and changed in real time.
Optionally, the target identifier may further include a name, a nickname, a gender, and the like.
Optionally, the first preset duration may be a preset duration of a first application program of the first device, or a duration set by a first user in a user-defined manner in the first application program.
Optionally, the "first device outputs the first information" may include any one of: the method comprises the steps that the first equipment vibrates for a preset number of times, the first equipment plays a preset voice or a preset ring, the first equipment flickers light with a preset color, the first equipment displays a preset message and the like. In this way, the first user may adjust the gaze direction according to the first information, i.e. the first user again focuses on the screen of the first device.
Optionally, the information interaction method provided in the embodiment of the present application may further include: and displaying the live broadcast picture in a target mode in a first interface of a first application program of the first equipment. Therefore, the first user can be reminded to adjust the sight direction by highlighting the live broadcast picture so as to achieve the purpose of reminding the first user to watch the live broadcast picture seriously. For example, in a first interface of a first application of a first device, a live view marked with a red box line is displayed.
Optionally, when the second device establishes a video call with multiple devices, the first device may send the second information to the second device, and may also send the second information to other devices.
And S203, the second device receives the second information.
Optionally, before the second device receives the second information, in a case where the second device establishes a video call with the multiple devices, multiple identifiers including the target identifier are displayed in a second interface of a second application program of the second device, where the multiple identifiers are identifiers corresponding to the multiple devices including the first device that establish the video call with the second device.
Optionally, the first display mode may be a mode of displaying the target identifier in a second interface of the second application by default. For example, in a second interface of a second application of the second device, a plurality of identifiers including the target identifier are displayed in default footage.
Optionally, the second display mode may be a mode in which the target identifier is highlighted in relation to the first display mode in a second interface of a second application program of the second device. For example, in the second interface of the second application program of the second device, the target identifier is displayed in an enlarged manner in a preset area and marked by a red frame line, and the preset area may be a front area in all the preset areas in the second interface. Namely, the target mark is fixedly displayed in the preset area. In this way, a second user using the second device may know that the first user's gaze is off the screen of the first device, so that the second user may prompt the first user to adjust the gaze direction by sending a voice message or a text message to the first device through the second device.
Optionally, in a second interface of a second application program of the second device, if the target identifier is fixedly displayed in one preset area in the second interface, the second device may continue to update the identifiers in the display identifier list except for the target identifier in other preset areas in the second interface. For example, the second interface includes an area 1 and an area 2, the area 1 may fixedly display the identifier of the student 1, and the area 2 may sequentially update and display the identifier of the student 2, the identifier of the student 3, and the identifier of the student 4 according to a preset period.
Optionally, in the case that the second device establishes a video call with multiple devices, if the other devices except the second device receive the second information, the other devices may add a display target identifier in an interface of another application program. As such, some users using other devices may know that the first user's attention corresponding to the target identification is inattentive and may alert those users to focus their attention on the screen.
For example, suppose that a first device used by student a (a first user) and a second device used by a teacher establish a video call to perform live online lectures, and the first preset time duration is 30 seconds. When the first device detects that the line of sight of student A deviates from the screen of the first device, the first device may start a timer to start timing. If the time counted by the timer (i.e., the time length of the sight line of the student A deviating from the screen of the first device) acquired by the first device is greater than or equal to 30 seconds, the first device may ring to prompt the student A to adjust the sight line direction and send the second information to the second device. After the second device receives the second information, the second device may update the target identifier from the first display mode to the second display mode, so that the teacher may determine that student a does not listen to the second information seriously through the second device. Thus, the teacher can remind student A to listen and speak seriously.
The embodiment of the application provides an information interaction method, wherein in the process of carrying out video call between first equipment and second equipment, the first equipment can detect that the sight of a first user using the first equipment deviates from the screen of the first equipment, and acquire first duration of the deviation of the sight of the first user from the screen of the first equipment. Therefore, in the case that the first duration is greater than or equal to the first preset duration, on the one hand, the first device may output the first information, so that the first user may adjust the gaze direction according to the first information, i.e., the first user refocuses the attention on the screen of the first device; on the other hand, the first device may send the second information to the second device, so that the second device may highlight the target identifier corresponding to the first user, and further, the second user using the second device may remind the first user to focus on the screen of the first device again. In this way, the interactive performance of the first device and the second device is enhanced, so that the quality of the video call is improved (for example, the live teaching quality is improved).
Optionally, after the first device outputs the first information, the first device may continue to acquire a face image of the first user through the device camera, so as to determine whether the line of sight of the first user points to the screen of the first device. For example, after the above S202, the information interaction method provided in the embodiment of the present application may further include the following S204 or S205, that is, the S204 and the S205 are alternatively executed.
S204, if the fact that the sight line of the first user points to the screen of the first device is detected, the first device stops outputting the first information.
Optionally, starting from the first device outputting the first information for prompting the first user to adjust the gaze direction, the first device may start a timer to start timing.
Optionally, the step S204 may specifically be: and if the fact that the sight line of the first user always points to the screen of the first equipment is detected within the target duration, the first equipment stops outputting the first information.
Optionally, the "stop outputting the first information" may include any one of: stopping shaking for a fixed number of times, stopping playing preset voice or preset ring, stopping flashing preset color light, canceling displaying preset messages, and the like.
S205, if the fact that the line of sight of the first user is always deviated from the screen of the first device is detected, the first device continuously outputs the first information.
Optionally, the step S205 may specifically be: if the first device detects that the line of sight of the first user is always deviated from the screen of the first device, the first device may determine that the line of sight direction of the first user is not adjusted all the time, and thus, may continuously output the first information at regular time intervals.
Optionally, the "continuously outputting the first information" may include any one of: continuously vibrating for a fixed time, continuously playing preset voice or preset ring, continuously flashing preset color light, continuously displaying preset messages at intervals and the like.
According to the information interaction method provided by the embodiment of the application, on one hand, if the first device detects that the sight of the first user points to the screen of the first device, the first device can stop outputting the first information in time, so that the electric quantity resource of the first device can be saved; on the other hand, if the first device detects that the line of sight of the first user always deviates from the screen of the first device, the first device may continuously output the first information until the line of sight of the first user points to the screen of the first device, thereby improving the video call quality of the first device and the second device. Therefore, the first equipment can control whether to output the first information or not according to the actual condition, the real scene is simulated, and the user experience is improved.
Optionally, after the first device outputs first information for prompting the first user to adjust the gaze direction, if the first device continues to acquire a facial image of the first user through the device camera and determines that the gaze of the first user is directed to the screen of the first device, the first device may send a message for releasing the highlighted target identifier to the second device. Exemplarily, with reference to fig. 2, as shown in fig. 4, the information interaction method provided in the embodiment of the present application may further include the following S206 and S207.
S206, if the fact that the sight line of the first user points to the screen of the first device is detected, the first device sends target information to the second device.
The target information may be used to instruct the second device to update the target identifier from the second display mode to the first display mode.
Optionally, when the first device detects that the gaze of the first user is directed to the screen of the first device, the first device may determine that the gaze of the first user is on the first device, and may send, to the second device, target information for controlling the second device to update the target identifier from the second display mode to the first display mode, that is, cancel the highlighted target identifier.
And S207, the second equipment receives the target information.
Optionally, in a second interface of a second application program of the second device, if the second device receives the target information, since the target information is used to control the second device to update the target identifier from the second display mode to the first display mode, the second device may update and display all identifiers including the target identifier in all preset areas.
It should be noted that S204 and S206 can be executed simultaneously; or, executing S204 first, and then executing S206; alternatively, S206 is executed first, and then S204 is executed. The method is determined according to actual conditions, and the embodiment of the application is not limited to the method.
According to the information interaction method provided by the embodiment of the application, if the first device detects that the sight line of the first user points to the screen of the first device, the first device can send target information for controlling the second device to update the target identifier from the second display mode to the first display mode to the second device. In this way, after the second device receives the target information, the second device cancels the highlighting of the target identifier to update the target identifier to the initial display mode, so that the second user using the second device can determine that the sight line of the first user corresponding to the target identifier is already directed to the screen of the first device, for example, a teacher can determine that the student continues to listen seriously.
The embodiment of the application provides an information interaction method, which is applied to first equipment shown in fig. 1. The information interaction method may include S301 and S302 described below.
S301, under the condition that the first device and the second device are in a video call state, the first device obtains a first duration.
The first duration is duration that the line of sight of the first user deviates from the screen of the first device.
S302, under the condition that the first duration is greater than or equal to a first preset duration, the first device outputs first information and sends second information to the second device.
The first information is used for prompting the first user to adjust the sight line direction, the second information is used for indicating the second equipment to update the target identifier from the first display mode to the second display mode, and the target identifier is an identifier corresponding to the first user.
Optionally, after the first device outputs the first information, the information interaction method provided in the embodiment of the present application may further include S303 or S304, where S303 or S304 is alternatively executed.
And S303, if the fact that the sight line of the first user points to the screen of the first device is detected, the first device stops outputting the first information.
S304, if the fact that the line of sight of the first user is always deviated from the screen of the first device is detected, the first device continuously outputs the first information.
Optionally, after the first device outputs the first information, the information interaction method provided in the embodiment of the present application may further include the following step S305.
S305, if the fact that the sight line of the first user points to the screen of the first device is detected, the first device sends target information to the second device.
The target information is used for instructing the second device to update the target identifier from the second display mode to the first display mode.
For the descriptions of S301 to S305, reference may be specifically made to the descriptions of S201 to S207 in the foregoing embodiments, and details are not repeated here.
The embodiment of the application provides an information interaction method, wherein in the process of carrying out video call between first equipment and second equipment, the first equipment can detect that the sight of a first user using the first equipment deviates from the screen of the first equipment, and acquire first duration of the deviation of the sight of the first user from the screen of the first equipment. Therefore, in the case that the first duration is greater than or equal to the first preset duration, on the one hand, the first device may output the first information, so that the first user may adjust the gaze direction according to the first information, i.e., the first user refocuses the attention on the screen of the first device; on the other hand, the first device may send the second information to the second device, so that the second device may highlight the target identifier corresponding to the first user, and further, the second user using the second device may remind the first user to focus on the screen of the first device again. In this way, the interactive performance of the first device and the second device is enhanced, so that the quality of the video call is improved (for example, the live teaching quality is improved).
The embodiment of the application provides an information interaction method, which is applied to a second device shown in fig. 1. The information interaction may include S401 described below.
S401, the second device receives second information sent by the first device.
For the description of S401, reference may be specifically made to the description of S203 in the foregoing embodiment, and details are not repeated here.
Optionally, the information interaction method provided in the embodiment of the present application may further include the following S402.
S402, the second device receives the target information sent by the first device.
For the descriptions of S401 and S402, reference may be specifically made to the descriptions of S201 to S207 in the foregoing embodiments, and details are not described here again.
The embodiment of the application provides an information interaction method, and as the second device can receive second information for controlling the second device to update the target identifier from the first display mode to the second display mode, the second device can highlight the target identifier on the screen of the second device in the second display mode, so that a second user using the second device can know that the sight line of a first user corresponding to the target identifier deviates from the screen of the first device, and therefore the video of the current first user of the second user using the second device deviates from the screen of the first device, and the second user can remind the first user of adjusting the sight line direction.
Example two
As shown in fig. 5, an embodiment of the present application provides an information interaction method, which is applied to a first device and a second device shown in fig. 1. The information interaction method may include S501, S502, and S503 described below.
S501, under the condition that the first device and the second device are in a video call state, the second device obtains a second duration.
The second duration may be a duration in which the gaze of the second user points to a target identifier in the screen of the second device, and the target identifier may be an identifier corresponding to the first user using the first device.
Specifically, the second device may detect the gaze angle of the second user in real time according to the gaze detection algorithm in the first embodiment, equally divide the screen of the second device into N regions, and map the face key point information of the second user. And if the sight line of the second user is kept in the same direction for more than S seconds, the second equipment considers that the second user focuses on other users indicated by the correspondingly displayed identification in the direction. The second device may query a screen area of the second device corresponding to the second user according to the gaze angle of the second user, and search for user information displayed in the screen area according to the screen area. At this time, the screen region may display the identifier fixedly, and the screen region is stopped from being refreshed to other identifiers by turns. Wherein S is a positive integer.
Optionally, in a case that the second device detects that the gaze of the second user is directed to the target identifier in the screen of the device, the second device may start a timer to start timing, that is, a duration of time that the gaze of the second user is directed to the target identifier in the screen of the second device.
And S502, under the condition that the second duration is greater than or equal to a second preset duration, the second device updates the target identifier from the first display mode to the second display mode and sends third information to the first device.
And the third information is used for prompting the first user to adjust the sight line direction.
Optionally, the second preset time duration may be a preset time duration of a second application program of the second device, or a time duration set by a second user in the second application program in a self-defined manner.
Optionally, the second preset time period may be the same as or different from the first preset time period.
Optionally, specific descriptions about the first display mode and the second display mode may refer to the specific descriptions in the foregoing embodiments, and are not repeated here.
Optionally, when the second duration is greater than or equal to a second preset duration, if the plurality of devices establish the video call with the second device, the second device may send the fourth information to the other devices, except the first device, of the plurality of devices. And the fourth information is used for controlling other equipment to display the target identification according to the second display mode respectively. In this way, the user using the other device may be made aware that the line of sight of the first user corresponding to the target identification is offset from the screen of the first device, thereby alerting the user using the other device that the line of sight can be directed towards the screen of the first device.
For example, during live teaching, if the line of sight of student a deviates from the screen of the first device, the second device may control the avatar of student a to be displayed on the devices of other students, and may alert the other students to listen to the speech seriously.
Optionally, the second device sends the third information to the first device when the second device receives the prompt information input by the user. Wherein, the prompt message may include any one of the following items: voice information, text information, etc., and the prompt information is used to indicate the first user.
For example, in the live teaching process, if the sight line of the student a deviates from the screen of the first device, the teacher may call the student a through the second device by inputting a voice message, and in the case where the voice message is received by the second device, since the voice message indicates the student a, the second device may transmit third information for prompting the student a to adjust the sight line direction to the first device used by the student a.
S503, the first device receives the third information.
Optionally, in a case where the first device receives the third information, the first device may output information for prompting the first user to adjust the gaze direction.
Optionally, the information interaction method provided by the present application may further include: in a case where the first device receives the third information, live content is displayed in a targeted manner at a first interface of a first application of the first device. For example, the targeted manner may be to display live content with red box line markers.
For example, suppose that a first device used by student a (first user) and a first device used by a teacher establish a video call to perform live online lecture, and the second preset time duration is 10 seconds. When the second device detects through the sight line algorithm that the teacher's sight line is directed to the avatar identifier (i.e., target identifier) indicating student a displayed in the screen of the first device, the second device may start a timer to start timing. If the time counted by the timer is greater than or equal to 10 seconds, the second device may update the target identifier from the first display mode to the second display mode, and send json-formatted third information through the Socket link. Thus, after the first device receives the third information, the first device may vibrate 5 to remind the student to adjust the sight line to the screen of the first device.
The embodiment of the application provides an information interaction method, and under the condition that a first device and a second device are in a video call state, the second device can detect that the sight of a second user using the second device points to a target identifier in a screen of the second device, and obtain a second duration that the sight of the second user points to the target identifier in the screen of the second device. Therefore, in the case that the second duration is greater than or equal to the second preset duration, on one hand, the second device may highlight the target identifier corresponding to the first user, so that the second user may pay more attention to whether the first user corresponding to the target identifier focuses on; on the other hand, the second device may send third information to the first device, so that the first device may remind the first user of adjusting the sight line in time according to the third information. Therefore, the interactive performance of the first device and the second device is improved, and the video call quality is improved.
Optionally, the second device and M devices are in a video call state, where the M devices include the first device. The screen of the second device includes N preset regions. M and N are positive integers, and M is greater than N. The information interaction method provided by the embodiment of the application further includes the following steps S504 and S505.
S504, the second device displays at least one mark in each preset area of the N preset areas.
Wherein one identifier corresponds one-to-one to a user using one of the M devices. For example, a picture identified as a face of a user captured by a camera of a device.
It can be understood that the N preset regions are located in regions other than the region for displaying the live content, respectively. For example, N is 4, 4 preset regions are respectively and evenly distributed at four corners of the screen of the second device, and the size of each preset region is smaller than or equal to the preset size. It will be appreciated that where the 4 pre-regions display at least one logo, occlusion of the main picture content may be reduced.
Optionally, the number of the identifiers displayed in each of the N preset regions may be the same or different.
And S505, updating and displaying the identifier in each preset area by the second equipment every third preset time.
Optionally, the third preset time period may be a preset time period of a second application program of the second device, or a time period set by a second user in the second application program in a self-defined manner.
It should be noted that the third preset time period is longer than the second preset time period, that is, after the second device updates and displays the identifier in each preset area, the second device can have enough time to determine whether the line of sight of the second user always faces the screen of the second device within the second preset time period.
Optionally, the "updating and displaying the identifier in each preset area" includes the following two cases:
(1) if all the M identifiers corresponding to the M devices can be displayed in the N preset regions, the second device may control each preset region in the N preset regions to display the identifier of each region in an exchange manner every third preset time.
(2) If the M identifiers corresponding to the M devices cannot be completely displayed in the N preset regions, that is, the N preset regions can only display part of the M identifiers, the N preset regions can refresh and display another part of the M identifiers every third preset time, and finally the M identifiers are sequentially displayed in the N preset regions.
According to the information interaction method provided by the embodiment of the application, under the condition that at least one identifier is displayed in each preset area of the N preset areas, the identifier in each preset area is updated and displayed every other third preset time length by the second equipment, so that the second user can monitor whether the sight of the user corresponding to the different identifiers points to the screen of the electronic equipment in sequence through observing the different identifiers updated and displayed in the preset areas in the screen of the second equipment, and the video call quality is ensured.
In the embodiment of the present application, the execution sequence of S501 to S503 and S504 to S505 is not limited. The above embodiment is described by taking the application of S504-S505 to the second embodiment as an example, and it is understood that S504-S505 may also be applied to the first embodiment, which is determined by actual situations, and the embodiment of the present application is not limited.
The embodiment of the application provides an information interaction method which is applied to second equipment. The information interaction method may include S601 and S602 described below.
S601, under the condition that the first device and the second device are in a video call state, the second device obtains a second duration.
And the second duration is the duration that the sight of the second user points to a target identifier in the screen of the second device, and the target identifier is an identifier corresponding to the first user using the first device.
And S602, under the condition that the second duration is greater than or equal to a second preset duration, the second device updates the target identifier from the first display mode to the second display mode and sends third information to the first device.
And the third information is used for prompting the first user to adjust the sight line direction.
Optionally, the information interaction method provided in the embodiment of the present application may further include the following S603 and S604.
S603, the second device displays at least one mark in each preset area of the N preset areas.
Wherein one identifier corresponds one-to-one to a user using one of the M devices.
And S604, updating and displaying the identifier in each preset area by the second equipment every third preset time.
For the descriptions of S601 to S604, reference may be specifically made to the descriptions of S501 to S505 in the foregoing embodiments, and details are not repeated here.
The embodiment of the application provides an information interaction method, and under the condition that a first device and a second device are in a video call state, the second device can detect that the sight of a second user using the second device points to a target identifier in a screen of the second device, and obtain a second duration that the sight of the second user points to the target identifier in the screen of the second device. Therefore, in the case that the second duration is greater than or equal to the second preset duration, on one hand, the second device may highlight the target identifier corresponding to the first user, so that the second user may pay more attention to whether the first user corresponding to the target identifier focuses on; on the other hand, the second device may send third information to the first device, so that the first device may remind the first user to adjust the sight line direction in time according to the third information. Therefore, the interactive performance of the first device and the second device is improved, and the video call quality is improved.
The embodiment of the application provides an information interaction method which is applied to first equipment. The information interaction method may include S701 described below.
S701, the first device receives third information sent by the second device.
For the description of the above S701, reference may be specifically made to the description of S503 in the above embodiment, and details are not repeated here.
According to the information interaction method provided by the embodiment of the application, the first device can receive the third information for prompting the first user to adjust the sight direction, so that after the first device receives the third information, the first device can remind the first user corresponding to the target identifier of adjusting the sight direction in time through the third information, and the video call quality between the first device and the second device is improved.
It should be noted that, in the information interaction methods provided in the first and second embodiments of the present application, the execution main body may be an information interaction device, or a control module used for executing the loaded information interaction method in the information interaction device.
As shown in fig. 6, an information interaction device 60 is provided in the embodiment of the present application. The information interaction device may include an acquisition module 61, a processing module 62, and a transmission module 63. The obtaining module 61 may be configured to obtain a first duration when the first device and the second device are in a video call state, where the first duration is a duration when the line of sight of the first user deviates from a screen of the first device. The processing module 62 may be configured to output the first information and control the sending module 63 to send the second information to the second device when the first duration obtained by the obtaining module 61 is greater than or equal to a first preset duration. The first information may be used to prompt the first user to adjust the gaze direction, and the second information may be used to instruct the second device to update the target identifier from the first display mode to the second display mode, where the target identifier is an identifier corresponding to the first user.
Optionally, the processing module 62 may be further configured to, after outputting the first information, stop outputting the first information if it is detected that the line of sight of the first user is directed to the screen of the first device; or if the fact that the line of sight of the first user is always deviated from the screen of the first device is detected, the first information is continuously output.
Optionally, the sending module 63 may be configured to send target information to the second device if it is detected that the line of sight of the first user points to the screen of the first device, where the target information is used to instruct the second device to update the target identifier from the second display mode to the first display mode.
The information interaction apparatus provided in this embodiment of the present application can implement each process implemented by the first device in the method embodiments of fig. 2 to fig. 4, and the information interaction apparatus may be the first device or a functional module in the first device, and is not described here again to avoid repetition.
The embodiment of the application provides an information interaction device, and in the process of carrying out video call between first equipment and second equipment, the first equipment can detect that the sight of a first user using the first equipment deviates from the screen of the first equipment, and acquire first duration of the deviation of the sight of the first user from the screen of the first equipment. Therefore, in the case that the first duration is greater than or equal to the first preset duration, on the one hand, the first device may output the first information, so that the first user may adjust the gaze direction according to the first information, i.e., the first user refocuses the attention on the screen of the first device; on the other hand, the first device may send the second information to the second device, so that the second device may highlight the target identifier corresponding to the first user, and further, the second user using the second device may remind the first user to focus on the screen of the first device again. In this way, the interactive performance of the first device and the second device is enhanced, so that the quality of the video call is improved (for example, the live teaching quality is improved).
As shown in fig. 7, an information interaction device 70 is provided in the present embodiment. The information interaction device may include an acquisition module 71, a display module 72, and a transmission module 73. The obtaining module 71 may be configured to, in a case that the first device and the second device are in a video call state, obtain a second duration, where the second duration is a duration in which the line of sight of the second user points to a target identifier in a screen of the second device, where the target identifier is an identifier corresponding to a first user using the first device. The display module 72 may be configured to update the target identifier from the first display mode to the second display mode when the second duration acquired by the acquisition module 71 is greater than or equal to a second preset duration. And the sending module 73 sends the third information to the first device when the second duration is greater than or equal to a second preset duration. Wherein the third information is used for prompting the first user to adjust the sight line direction.
Optionally, the second device is in a video call state with M devices, where the M devices include the first device. The screen of the second device comprises N preset areas; m and N are positive integers, and M is greater than N. The display module 72 may be further configured to display at least one identifier in each of the N preset areas, where one identifier corresponds to a user using one of the M devices; and updating and displaying the identifier in each preset area every third preset time.
The information interaction apparatus provided in the second embodiment of the present application can implement each process implemented by the second device in the method embodiment of fig. 5, and the information interaction apparatus may be the second device or a functional module in the second device, and is not described here again to avoid repetition.
The embodiment of the application provides an information interaction device, and under the condition that a first device and a second device are in a video call state, the second device can detect that a sight line of a second user using the second device points to a target identifier in a screen of the second device, and obtain a second duration that the sight line of the second user points to the target identifier in the screen of the second device. Therefore, in the case that the second duration is greater than or equal to the second preset duration, on one hand, the second device may highlight the target identifier corresponding to the first user, so that the second user may pay more attention to whether the first user corresponding to the target identifier focuses on; on the other hand, the second device may send third information to the first device, so that the first device may remind the first user to adjust the sight line direction in time according to the third information. Therefore, the interactive performance of the first device and the second device is improved, and the video call quality is improved.
The information interaction device in the first and second embodiments of the present application may be a device, or may be a component, an integrated circuit, or a chip in a terminal. The device can be mobile electronic equipment or non-mobile electronic equipment. By way of example, the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a palm top computer, a vehicle-mounted electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook or a Personal Digital Assistant (PDA), and the like, and the non-mobile electronic device may be a server, a Network Attached Storage (NAS), a Personal Computer (PC), a Television (TV), a teller machine or a self-service machine, and the like, and the embodiments of the present application are not particularly limited.
The information interaction device in the embodiment of the present application may be a device having an operating system. The operating system may be an Android (Android) operating system, an ios operating system, or other possible operating systems, and embodiments of the present application are not limited specifically.
Optionally, an electronic device is further provided in this embodiment of the present application, and includes the processor 100 shown in fig. 8, a memory 109, and a program or an instruction stored in the memory 109 and executable on the processor 110, where the program or the instruction is executed by the processor 110 to implement each process of the information interaction method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
It should be noted that the electronic devices in the embodiments of the present application include the mobile electronic devices and the non-mobile electronic devices described above.
Fig. 8 is a schematic diagram of a hardware structure of an electronic device implementing an embodiment of the present application.
The electronic device 100 includes, but is not limited to: a radio frequency unit 101, a network module 102, an audio output unit 103, an input unit 104, a sensor 105, a display unit 106, a user input unit 107, an interface unit 108, a memory 109, and a processor 110. Among them, the input unit 104 may include a graphic processor and a microphone, the user input unit 107 may include a touch panel and other input devices, the display unit 106 may include a display panel, and the memory 109 may include an application program and an operating system.
Those skilled in the art will appreciate that the electronic device 100 may further comprise a power source (e.g., a battery) for supplying power to various components, and the power source may be logically connected to the processor 110 through a power management system, so as to implement functions of managing charging, discharging, and power consumption through the power management system. The electronic device structure shown in fig. 8 does not constitute a limitation of the electronic device, and the electronic device may include more or less components than those shown, or combine some components, or arrange different components, and thus, the description is omitted here.
The first mode is
The processor 110 is configured to obtain a first duration when the first device and the second device are in a video call state, where the first duration is a duration when a line of sight of the first user deviates from a screen of the first device; and outputting the first information and sending the second information to the second equipment under the condition that the first duration is greater than or equal to a first preset duration. The first information is used for prompting the first user to adjust the sight line direction, the second information is used for indicating the second equipment to update the target identifier from the first display mode to the second display mode, and the target identifier is an identifier corresponding to the first user.
Optionally, the processor 110 is further configured to stop outputting the first information if it is detected that the line of sight of the first user is directed to the screen of the first device; or if the fact that the line of sight of the first user is always deviated from the screen of the first device is detected, the first information is continuously output.
Optionally, the processor 110 is further configured to send target information to the second device if it is detected that the line of sight of the first user is directed to the screen of the first device, where the target information is used to instruct the second device to update the target identifier from the second display mode to the first display mode.
The embodiment of the application provides electronic equipment, and in the process of carrying out video call between first equipment and second equipment, the first equipment can detect that the sight of a first user using the first equipment deviates from a screen of the first equipment, and acquire first duration of the deviation of the sight of the first user from the screen of the first equipment. Therefore, in the case that the first duration is greater than or equal to the first preset duration, on the one hand, the first device may output the first information, so that the first user may adjust the gaze direction according to the first information, i.e., the first user refocuses the attention on the screen of the first device; on the other hand, the first device may send the second information to the second device, so that the second device may highlight the target identifier corresponding to the first user, and further, the second user using the second device may remind the first user to focus on the screen of the first device again. In this way, the interactive performance of the first device and the second device is enhanced, so that the quality of the video call is improved (for example, the live teaching quality is improved).
The second mode is
The processor 110 is configured to, when the first device and the second device are in a video call state, obtain a second duration, where the second duration is a duration in which a sight line of a second user points to a target identifier in a screen of the second device, and the target identifier is an identifier corresponding to a first user using the first device; and under the condition that the second duration is greater than or equal to a second preset duration, updating the target identifier from the first display mode to a second display mode, and sending third information to the first equipment. And the third information is used for prompting the first user to adjust the sight line direction.
Optionally, the second device is in a video call state with M devices, where the M devices include the first device. The screen of the second device comprises N preset areas, M and N are positive integers, and M is larger than N. The processor 110 is further configured to display at least one identifier in each of the N preset areas, where one identifier corresponds to a user using one of the M devices; and updating and displaying the identifier in each preset area every third preset time.
The embodiment of the application provides an electronic device, and under the condition that a first device and a second device are in a video call state, the second device can detect that a sight line of a second user using the second device points to a target identifier in a screen of the second device, and obtain a second duration that the sight line of the second user points to the target identifier in the screen of the second device. Therefore, in the case that the second duration is greater than or equal to the second preset duration, on one hand, the second device may highlight the target identifier corresponding to the first user, so that the second user may pay more attention to whether the first user corresponding to the target identifier focuses on; on the other hand, the second device may send third information to the first device, so that the first device may remind the first user to adjust the sight line direction in time according to the third information. Therefore, the interactive performance of the first device and the second device is improved, and the video call quality is improved.
The embodiment of the present application further provides a readable storage medium, where a program or an instruction is stored on the readable storage medium, and when the program or the instruction is executed by a processor, the program or the instruction implements each process of the above-mentioned information interaction method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
The processor is the processor in the electronic device in the above embodiment. Readable storage media, including computer-readable storage media such as a computer-read-only memory (ROM), a Random Access Memory (RAM), a magnetic or optical disk, and so forth.
The embodiment of the present application further provides a chip, where the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is configured to execute a program or an instruction to implement each process of the information interaction method embodiment, and the same technical effect can be achieved.
It should be understood that the chips mentioned in the embodiments of the present application may also be referred to as system-on-chip, system-on-chip or system-on-chip, etc.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element. Further, it should be noted that the scope of the methods and apparatus of the embodiments of the present application is not limited to performing the functions in the order illustrated or discussed, but may include performing the functions in a substantially simultaneous manner or in a reverse order based on the functions involved, e.g., the methods described may be performed in an order different than that described, and various steps may be added, omitted, or combined. In addition, features described with reference to certain examples may be combined in other examples.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present application may be embodied in the form of a software product, which is stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal (such as a mobile phone, a computer, a server, an air conditioner, or a network device) to execute the method in the embodiments of the present application.
While the present embodiments have been described with reference to the accompanying drawings, it is to be understood that the invention is not limited to the precise embodiments described above, which are meant to be illustrative and not restrictive, and that various changes may be made therein by those skilled in the art without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (8)

1. An information interaction method, characterized in that the method comprises:
under the condition that a first device and a second device are in a video call state, acquiring a first time length, wherein the first time length is the time length of the deviation of the sight of a first user from a screen of the first device;
under the condition that the first duration is greater than or equal to a first preset duration, outputting first information and sending second information to the second equipment;
the first information is used for prompting the first user to adjust the sight line direction, the second information is used for indicating the second equipment to update a target identifier from a first display mode to a second display mode, and the target identifier is an identifier corresponding to the first user;
after the outputting the first information, the method further comprises:
if the fact that the sight line of the first user points to the screen of the first device is detected, target information is sent to the second device, and the target information is used for indicating the second device to update the target identifier from the second display mode to the first display mode;
the second device displays a plurality of identifiers including the target identifier.
2. The method of claim 1, wherein after outputting the first information, the method further comprises:
if the fact that the sight line of the first user points to the screen of the first device is detected, stopping outputting the first information; alternatively, the first and second electrodes may be,
and if the fact that the sight line of the first user is always deviated from the screen of the first device is detected, the first information is continuously output.
3. An information interaction method, characterized in that the method comprises:
under the condition that a first device and a second device are in a video call state, acquiring a second time length, wherein the second time length is the time length when the sight of a second user points to a target identifier in a screen of the second device, and the target identifier is an identifier corresponding to a first user using the first device;
under the condition that the second duration is greater than or equal to a second preset duration, updating the target identifier from a first display mode to a second display mode, and sending third information to the first equipment;
the third information is used for prompting the first user to adjust the sight line direction; the third information is further used for outputting information for prompting the first user to adjust the sight line direction by the first equipment under the condition that the first equipment receives the third information.
4. The method of claim 3, wherein the second device is in a video call state with M devices, the M devices including the first device; the screen of the second device comprises N preset areas; m and N are positive integers, and M is greater than N;
the method further comprises the following steps:
displaying at least one identifier in each of the N preset areas, wherein one identifier corresponds to a user using one of the M devices one by one;
and updating and displaying the identifier in each preset area every third preset time.
5. The information interaction device is characterized by comprising an acquisition module, a processing module and a sending module;
the acquisition module is used for acquiring a first time length under the condition that the first equipment and the second equipment are in a video call state, wherein the first time length is the time length of the sight of the first user deviating from the screen of the first equipment;
the processing module is configured to output first information and control the sending module to send second information to the second device when the first duration obtained by the obtaining module is greater than or equal to a first preset duration;
the first information is used for prompting the first user to adjust the sight line direction, the second information is used for indicating the second equipment to update a target identifier from a first display mode to a second display mode, and the target identifier is an identifier corresponding to the first user;
the sending module is further configured to send target information to the second device if it is detected that the line of sight of the first user points to the screen of the first device, where the target information is used to instruct the second device to update the target identifier from the second display mode to the first display mode;
the second device displays a plurality of identifiers including the target identifier.
6. The information interaction device according to claim 5, wherein the processing module is further configured to, after outputting the first information, stop outputting the first information if it is detected that the line of sight of the first user is directed to the screen of the first device; or if the fact that the line of sight of the first user is always deviated from the screen of the first device is detected, the first information is continuously output.
7. The information interaction device is characterized by comprising an acquisition module, a display module and a sending module;
the acquisition module is used for acquiring a second time length when the first device and the second device are in a video call state, wherein the second time length is the time length when the sight of a second user points to a target identifier in a screen of the second device, and the target identifier is an identifier corresponding to a first user using the first device;
the display module is used for updating the target identifier from a first display mode to a second display mode under the condition that the second duration acquired by the acquisition module is greater than or equal to a second preset duration;
the sending module is configured to send third information to the first device when the second duration is greater than or equal to a second preset duration;
the third information is used for prompting the first user to adjust the sight line direction; the third information is further used for outputting information for prompting the first user to adjust the sight line direction by the first equipment under the condition that the first equipment receives the third information.
8. The information interaction device according to claim 7, wherein the second device is in a video call state with M devices, the M devices including the first device; the screen of the second device comprises N preset areas; m and N are positive integers, and M is greater than N;
the display module is further configured to display at least one identifier in each of the N preset areas, where one identifier corresponds to a user using one of the M devices one to one; and updating and displaying the identifier in each preset area every third preset time.
CN202010421274.3A 2020-05-18 2020-05-18 Information interaction method and information interaction device Active CN111601064B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010421274.3A CN111601064B (en) 2020-05-18 2020-05-18 Information interaction method and information interaction device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010421274.3A CN111601064B (en) 2020-05-18 2020-05-18 Information interaction method and information interaction device

Publications (2)

Publication Number Publication Date
CN111601064A CN111601064A (en) 2020-08-28
CN111601064B true CN111601064B (en) 2022-02-22

Family

ID=72191560

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010421274.3A Active CN111601064B (en) 2020-05-18 2020-05-18 Information interaction method and information interaction device

Country Status (1)

Country Link
CN (1) CN111601064B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112261431B (en) * 2020-10-21 2022-01-14 联想(北京)有限公司 Image processing method and device and electronic equipment
CN115250379B (en) * 2021-04-25 2024-04-09 花瓣云科技有限公司 Video display method, terminal, system and computer readable storage medium
CN113778230A (en) * 2021-09-14 2021-12-10 Oppo广东移动通信有限公司 Information interaction method, device, equipment and storage medium
CN114281236B (en) * 2021-12-28 2023-08-15 建信金融科技有限责任公司 Text processing method, apparatus, device, medium, and program product
CN116193064B (en) * 2023-02-21 2023-09-01 北京洞察力科技股份有限公司 Method and system for realizing intercommunication between video conference systems

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102981620A (en) * 2012-11-27 2013-03-20 中兴通讯股份有限公司 Terminal operation method and terminal
CN104408781A (en) * 2014-12-04 2015-03-11 重庆晋才富熙科技有限公司 Concentration attendance system
CN104464406A (en) * 2013-09-12 2015-03-25 郑州学生宝电子科技有限公司 Real-time interactive online learning platform
CN104980683A (en) * 2014-04-08 2015-10-14 深圳市潮流网络技术有限公司 Implement method and device for video telephone conference
CN105005850A (en) * 2015-06-29 2015-10-28 天脉聚源(北京)科技有限公司 Method and apparatus for updating virtual audiences
CN106599881A (en) * 2016-12-30 2017-04-26 首都师范大学 Student state determination method, device and system
CN106952200A (en) * 2017-03-28 2017-07-14 安徽味唯网络科技有限公司 A kind of method that internet teaching supervises student
US9741257B1 (en) * 2016-03-30 2017-08-22 Avaya Inc. System and method for coordinated learning and teaching using a videoconference system
CN108596159A (en) * 2018-06-13 2018-09-28 钱建平 A kind of teaching method and system based on Internet education
CN109862375A (en) * 2019-01-07 2019-06-07 北京汉博信息技术有限公司 Cloud recording and broadcasting system
CN110248254A (en) * 2019-06-11 2019-09-17 Oppo广东移动通信有限公司 Display control method and Related product
CN110308790A (en) * 2019-06-04 2019-10-08 宁波视睿迪光电有限公司 The image adjusting device and system of teaching demonstration

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140255889A1 (en) * 2013-03-10 2014-09-11 Edulock, Inc. System and method for a comprehensive integrated education system
JP2016224766A (en) * 2015-06-01 2016-12-28 富士通株式会社 Remote screen display system, remote screen display method, and remote screen display program
CN109299645A (en) * 2018-07-20 2019-02-01 北京旷视科技有限公司 Method, apparatus, system and storage medium for sight protectio prompt
CN109558053A (en) * 2018-11-30 2019-04-02 努比亚技术有限公司 A kind of method, mobile terminal and computer readable storage medium for adjusting screen and showing
CN110955922B (en) * 2019-12-17 2023-03-21 联想(北京)有限公司 Display method and display device

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102981620A (en) * 2012-11-27 2013-03-20 中兴通讯股份有限公司 Terminal operation method and terminal
CN104464406A (en) * 2013-09-12 2015-03-25 郑州学生宝电子科技有限公司 Real-time interactive online learning platform
CN104980683A (en) * 2014-04-08 2015-10-14 深圳市潮流网络技术有限公司 Implement method and device for video telephone conference
CN104408781A (en) * 2014-12-04 2015-03-11 重庆晋才富熙科技有限公司 Concentration attendance system
CN105005850A (en) * 2015-06-29 2015-10-28 天脉聚源(北京)科技有限公司 Method and apparatus for updating virtual audiences
US9741257B1 (en) * 2016-03-30 2017-08-22 Avaya Inc. System and method for coordinated learning and teaching using a videoconference system
CN106599881A (en) * 2016-12-30 2017-04-26 首都师范大学 Student state determination method, device and system
CN106952200A (en) * 2017-03-28 2017-07-14 安徽味唯网络科技有限公司 A kind of method that internet teaching supervises student
CN108596159A (en) * 2018-06-13 2018-09-28 钱建平 A kind of teaching method and system based on Internet education
CN109862375A (en) * 2019-01-07 2019-06-07 北京汉博信息技术有限公司 Cloud recording and broadcasting system
CN110308790A (en) * 2019-06-04 2019-10-08 宁波视睿迪光电有限公司 The image adjusting device and system of teaching demonstration
CN110248254A (en) * 2019-06-11 2019-09-17 Oppo广东移动通信有限公司 Display control method and Related product

Also Published As

Publication number Publication date
CN111601064A (en) 2020-08-28

Similar Documents

Publication Publication Date Title
CN111601064B (en) Information interaction method and information interaction device
US11089266B2 (en) Communication processing method, terminal, and storage medium
CN110348524B (en) Human body key point detection method and device, electronic equipment and storage medium
US8508571B2 (en) Teleconference system
CN112287844B (en) Student situation analysis method and device, electronic device and storage medium
US20180324229A1 (en) Systems and methods for providing expert assistance from a remote expert to a user operating an augmented reality device
WO2015188614A1 (en) Method and device for operating computer and mobile phone in virtual world, and glasses using same
RU2665901C1 (en) Method and apparatus for testing the software of a head device of virtual reality display
WO2021227918A1 (en) Interaction method and augmented reality device
CN113965807A (en) Message pushing method, device, terminal, server and storage medium
CN114390308B (en) Interface display method, device, equipment, medium and product in live broadcast process
CN111541928A (en) Live broadcast display method, device, equipment and storage medium
WO2023016107A1 (en) Remote interaction method, apparatus and system, and electronic device and storage medium
CN111163280B (en) Asymmetric video conference system and method thereof
JPWO2018158852A1 (en) Call system
CN112702533B (en) Sight line correction method and sight line correction device
CN106201394B (en) Interact controlling terminal, interactive control method, server and mutual induction control system
US20220291752A1 (en) Distributed Application Platform Projected on a Secondary Display for Entertainment, Gaming and Learning with Intelligent Gesture Interactions and Complex Input Composition for Control
TW201608898A (en) Method and system for processing video conference
CN111835617B (en) User head portrait adjusting method and device and electronic equipment
CN114371904B (en) Data display method and device, mobile terminal and storage medium
US20230230416A1 (en) Establishing private communication channels
CN112367533B (en) Interactive service processing method, device, equipment and computer readable storage medium
JP2022142913A (en) Conference support control device and conference support method
JP6976206B2 (en) Terminal equipment, server equipment and programs

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant