CN115546877B - Visual angle adjusting method and device - Google Patents

Visual angle adjusting method and device Download PDF

Info

Publication number
CN115546877B
CN115546877B CN202211461514.8A CN202211461514A CN115546877B CN 115546877 B CN115546877 B CN 115546877B CN 202211461514 A CN202211461514 A CN 202211461514A CN 115546877 B CN115546877 B CN 115546877B
Authority
CN
China
Prior art keywords
participant
participants
display device
visual angle
docking station
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202211461514.8A
Other languages
Chinese (zh)
Other versions
CN115546877A (en
Inventor
肖艳菊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Xfanic Technology Co Ltd
Original Assignee
Shenzhen Xfanic Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Xfanic Technology Co Ltd filed Critical Shenzhen Xfanic Technology Co Ltd
Priority to CN202211461514.8A priority Critical patent/CN115546877B/en
Publication of CN115546877A publication Critical patent/CN115546877A/en
Application granted granted Critical
Publication of CN115546877B publication Critical patent/CN115546877B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Ophthalmology & Optometry (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The embodiment of the application discloses a method and a device for adjusting a visual angle, wherein the method comprises the following steps: determining communication closeness between the first participant and each participant in the plurality of participants according to the behavior information of the first participant and the behavior information of the plurality of participants; selecting a second participant from the plurality of participants based on the closeness of communication between the first participant and each of the plurality of participants; determining the eyeball sight direction of the first participant aiming at the second participant currently in the conference video; and if the eyeball sight direction of the first participant to the second participant is the non-main visual angle direction, adjusting the eyeball sight direction of the first participant to the second participant from the non-main visual angle direction to a main visual angle direction. The method and the device can adjust the difference between the visual angle of the conference room user and the visual angle of the remote user in a targeted manner.

Description

Visual angle adjusting method and device
Technical Field
The application relates to a communication technology, which is applied to the fields of internet, big data and the like, in particular to a method and a device for adjusting a visual angle.
Background
The docking station is also called a port replicator, and is an external device specially designed for notebook computers. By copying or even expanding the port of the notebook computer, the notebook computer can be conveniently connected with a plurality of accessories or external devices (such as a power adapter, a network cable, a mouse, an external keyboard, a printer and an external display) in a one-stop mode.
Due to the rapid development of communication technology, in the video conference scene, a video sharing docking station in a traditional mode and a Pro mode exists on the market, wherein the video sharing docking station in the traditional mode is low in cost, can highly integrate the problems of camera, sound box, microphone and software adaptation, and can be conveniently used by being directly connected with equipment such as a computer. In addition, the built-in operating system of the Pro mode video sharing docking station downloads conference software in the machine, and can be directly connected with large-screen equipment, but the cost of the video sharing docking station is high. However, the video sharing docking station with the two different modes cannot solve the problem that the difference exists between the viewing angle of the conference room user and the viewing angle of the remote user in the real face-to-face communication scene due to the large spatial position of the large screen position of the conference room of the remote user image and the spatial position of the video sharing docking station, so that the video conference experience of the user is poor.
Disclosure of Invention
The embodiment of the application provides a visual angle adjusting method and device, which can adjust differences between a conference room user visual angle and a remote user visual angle and visual angles in a real face-to-face communication scene in a targeted manner, and improve experience of participants.
In a first aspect, an embodiment of the present application provides a method for adjusting a viewing angle, where the method is applied to a first video sharing docking station, where the first video sharing docking station is applied to a conference scene, the conference scene further includes a second video sharing docking station, a first display device and a second display device, where the first video sharing docking station and the first display device are devices used by multiple participants to join in a conference, the second video sharing docking station and the second display device are devices used by a first participant to join in the conference, the first video sharing docking station, the second video sharing docking station, the first display device and the second display device establish a communication connection, the first participant performs conference communication with the multiple participants through the second display device, and the multiple participants perform conference communication with the first participant through the first display device, and the method includes:
acquiring behavior information of the first participant and behavior information of the plurality of participants, wherein the behavior information comprises language information, action information and face information;
determining communication closeness between the first participant and each of the plurality of participants according to the behavior information of the first participant and the behavior information of the plurality of participants;
selecting a second participant from the plurality of participants based on the closeness of communication between the first participant and each of the plurality of participants;
determining an eyeball sight direction of the first participant to the second participant currently in the conference video, wherein the eyeball sight direction comprises a main visual angle direction or a non-main visual angle direction;
if the current eyeball sight line direction of the first participant for the second participant is the non-main visual angle direction, the current eyeball sight line direction of the first participant for the second participant is adjusted from the non-main visual angle direction to the main visual angle direction.
In a video conference scene, the user experience of the video conference is poor due to the difference between the visual angle of a conference room user and the visual angle of a remote user and the visual angle of a real face-to-face communication scene. In the prior art, a human eye part of a human face part in a first image is adjusted to obtain a second image which corresponds to the first image and in which the human eye part of a user in the human face image is in a front-view state, so that in a video communication process, a video image acquired by an acquisition unit in real time is processed by adjusting parameters to obtain a corresponding image in the front-view state, and images displayed in a display unit are all adjusted and have a visual sense. According to the method, the first video sharing docking station determines the communication closeness degree between the first participant and each of the multiple participants according to the behavior information of the first participant (namely, the remote participant) and the behavior information of the multiple participants (namely, the participants in the conference room), then selects the second participant from the multiple participants according to the communication closeness degree between the first participant and each of the multiple participants (namely, the participant actually looking at by the first participant is determined to be the second participant according to the behavior information), and if the eyeball sight direction of the first participant currently aiming at the second participant is determined to be the non-main visual angle direction (namely, the sight direction of the first participant currently looking at the second participant is determined to be the non-main visual angle direction according to the angle of the first video sharing docking station in the conference room, the sight direction of the first participant currently looking at the second participant is determined to be the non-main visual angle), the first video sharing docking station adjusts the sight direction of the first participant aiming at the second participant to be the non-main visual angle direction by the non-main visual angle direction. The method and the device can adjust the difference between the visual angle of the user in the conference room and the visual angle of the remote user in the real face-to-face communication scene in a targeted manner, and the experience of the participants is improved.
In one possible embodiment, the method further comprises:
receiving first information and second information sent by a second video sharing docking station, wherein the first information is acquired eyeball motion track image information of the first participant, and the second information is an acquired area displayed on the second display device by the plurality of participants;
determining the matching degree of the visual angle of the first participant and each participant in the multiple participants according to the eyeball motion track image information of the first participant and the area displayed on the second display equipment by the multiple participants;
selecting a second participant from the plurality of participants based on the closeness of communication between the first participant and each of the plurality of participants, comprising:
selecting the second participant from the plurality of participants based on closeness of communication between the first participant and each of the plurality of participants, the perspective of the first participant and the degree of matching of each of the plurality of participants.
In the method, before the first video sharing docking station selects the second participant from the multiple participants according to the communication closeness degree between the first participant and each participant in the multiple participants, the first video sharing docking station may further receive the eye movement track image information of the first participant and the area displayed on the second display device by the multiple participants, which are sent by the second video sharing docking station, and determine the matching degree between the visual angle of the first participant and each participant in the multiple participants according to the eye movement track image information of the first participant and the area displayed on the second display device by the multiple participants. Further, a second participant is selected from the plurality of participants based on the closeness of communication between the first participant and each of the plurality of participants, the degree of matching between the perspective of the first participant and each of the plurality of participants. According to the scheme, the second participant is selected from the multiple participants according to the communication closeness degree between the first participant and each participant in the multiple participants, the visual angle of the first participant and the matching degree of each participant in the multiple participants, the second participant is selected from the multiple participants by integrating two different dimensions, and the accuracy of selecting the second participant from the multiple participants can be effectively improved.
In another possible embodiment, the determining, according to the eye movement path image information of the first participant and the area displayed on the second display device by the plurality of participants, the matching degree between the visual angle of the first participant and each of the plurality of participants includes:
determining the variable quantity of the eyeball sight line position of the first participant according to the eyeball motion track image information of the first participant;
determining a viewpoint position of the first participant falling on each of the plurality of participants according to the variation of the eyeball sight line position of the first participant;
determining a target area of the second display device watched by the first participant according to the viewpoint position of the first participant in each of the plurality of participants;
and determining the matching degree of the visual angle of the first participant and each participant in the multiple participants according to the target area of the second display equipment watched by the first participant and the area displayed on the second display equipment by the multiple participants.
In the method, the specific process of determining the matching degree between the visual angle of the first participant and each participant in the multiple participants according to the eye movement track image information of the first participant and the area displayed on the second display device by the multiple participants by the first video sharing docking station may be as follows: determining a variation of an eyeball sight line position of the first participant according to the eyeball movement track image information of the first participant (for example, an eyeball sight line direction in which the first participant gazes at the second display device is changed from 30 degrees north to 45 degrees east to west), determining a viewpoint position in which the first participant falls on each of the multiple participants according to the variation of the eyeball sight line position of the first participant and a position in which each of the multiple participants is located (for example, a position in which a participant 2 of the multiple participants is located on the second display device is a true west direction, a position in which a participant 3 is located on the second display device is a true north direction, a position in which a participant 4 is located on the second display device is a west north direction, and a position in which a participant 5 is located on the second display device is a true east direction), then, determining a target area where the first participant watches the second display device according to the viewpoint position of each participant in the plurality of participants (determining that the target area where the first participant watches the second display device is the west direction according to the information), and finally determining the matching degree of the visual angle of the first participant and each participant in the plurality of participants according to the target area where the first participant watches the second display device and the area where the plurality of participants display on the second display device (namely, because the position of the participant 2 on the second display device is the west direction, the position of the participant 3 on the second display device is the north direction, the position of the participant 4 on the second display device is the north west direction, and the position of the participant 5 on the second display device is the east direction, and the target area in which the first participant gazes at the second display device is the west-positive direction, it is determined that the degree of matching between the visual angle of the first participant and participant 2 is 95%, the degree of matching between the visual angle of the first participant and participant 3 is 10%, the degree of matching between the visual angle of the first participant and participant 4 is 54%, and the degree of matching between the visual angle of the first participant and participant 5 is 2%). The accuracy of selecting the second participant from the plurality of participants can be improved by determining the matching degree of the visual angle of the first participant and each participant in the plurality of participants.
In yet another possible implementation manner, if the target area that the first participant gazes at the second display device is an area that the second participant displays on the second display device, the matching degree between the viewing angle of the first participant and the second participant is the highest.
In the above method, the first video sharing docking station may determine the participant with the highest degree of matching with the viewing angle of the first participant among the plurality of participants by defining the area where the first participant gazes at the target area of the second display device and the area where the plurality of participants are displayed on the second display device, for example, if the plurality of participants include participant 2, participant 3, participant 4, and participant 5, the target area where participant 1 gazes at the second display device is in the west direction, and participant 2 is located on the second display device in the west direction (the area where participant 1 gazes at the target area of the second display device is 5 cm) 2 ) The position of the participant 3 on the second display device is the due north direction (the area overlapping with the target area of the participant 1 watching the second display device is 1 cm) 2 ) The position of the participant 4 on the second display device is in the northwest direction (the area coinciding with the target area of the participant 1 looking at the second display device is 2.8 cm) 2 ) The position of the participant 5 on the second display device is the east-ward direction (the area overlapping the target area where the participant 1 gazes at the second display device is 0 cm) 2 ) As can be seen from the above, the maximum overlapping area between the target area where the participant 1 watches the second display device and the area where the participant 2 is displayed on the second display device is, the highest matching degree between the viewing angle of the participant 1 and the participant 2 is obtained. According to the scheme, the accuracy rate of selecting the second participant from a plurality of participants can be improved by a specific matching degree determining mode.
In yet another possible embodiment, said determining the closeness of communication between said first participant and each of said plurality of participants based on said first participant's behavioral information and said plurality of participant's behavioral information comprises:
acquiring identity information of each participant in the multiple participants, wherein the responsibilities of the participants with different identity information are different;
generating target keywords of the first participant in a conference process according to the behavior information of the first participant and the behavior information of the multiple participants, wherein the target keywords comprise keywords of question contents provided by the first participant, keywords of answer contents of the multiple participants for the questions and keywords of communication conversation contents of the first participant and each of the multiple participants;
and matching the target keyword with the identity information of each participant in the multiple participants to obtain the communication closeness degree between the first participant and each participant in the multiple participants.
In the above method, the specific process of the first video sharing docking station determining the communication closeness between the first participant and each of the multiple participants according to the behavior information of the first participant and the behavior information of the multiple participants may be: first, identity information of each of a plurality of participants is obtained (for example, if the subject of the conference video is a project report of each project group responsible person, the first video sharing docking station may obtain the identity information of the above four participants, and the participants with different identity information have different responsibility ranges, wherein participant 2 is responsible for project a, participant 3 is responsible for project B, participant 4 is responsible for project C, and participant 5 is responsible for project D, and the four responsible projects are different), then, a target keyword of the first participant in the conference process is generated according to the behavior information of the first participant and the behavior information of the plurality of participants (wherein the target keyword may include a keyword of the content of question 1 provided by participant 1; and may include keywords of the content of participant 2, participant 3, participant 4, and participant 5 respectively corresponding to question 1; and may include keywords of the content of participant 1 matched with participant 2, participant 3, participant 4, and participant 5 respectively, and the target communication degree of the communication between participant 1 and participant 2 is 30%, and 80% (for example, the content of the participant 2, participant 4%, and participant 5% is obtained as the last communication degree of the communication between the first and second participant (for the communication), and the communication degree of the communication between participant 2, and the participant 5%, and the last communication degree of the communication between the participant (for the communication between participant 2, and the participant 5%, and the participant) (for the communication degree of the communication between the participant 5%, and the participant). According to the method and the device, the communication closeness degree between the first participant and each participant in the multiple participants is determined according to the target keyword and the identity information of each participant in the multiple participants, and the accuracy rate of selecting the second participant from the multiple participants can be effectively improved.
In yet another possible embodiment, said determining the eye gaze direction of the first participant currently directed at the second participant comprises:
calculating a deviation value between the position of each of the plurality of participants and the position of the first display device;
and if the deviation value between the position of the second participant and the position of the first display device in the multiple participants meets a preset condition, determining that the current eyeball sight line direction of the first participant for the second participant is the non-main visual angle direction.
In the above method, after the first video sharing docking station selects the second participant from the plurality of participants, it is further required to determine an eyeball sight direction of the first participant with respect to the second participant, specifically, the first video sharing docking station may first calculate an offset value between a position of each of the plurality of participants and a position of the first display device (for example, the first video sharing docking station determines that an offset value between a position of the participant 2 and a position of the display device in the conference room is 6cm, an offset value between a position of the participant 3 and a position of the display device in the conference room is 1cm, an offset value between a position of the participant 4 and a position of the display device in the conference room is 2cm, and an offset value between a position of the participant 5 and a position of the display device in the conference room is 6 cm), where the offset value between a position of the participant 2 and a position of the first display device satisfies a predetermined condition that the eyeball sight line direction of the first participant (for example, the first video sharing docking station may determine that the offset value between the position of the participant 2 and the display device in the conference room is greater than a predetermined threshold value (for the first video sharing direction) of the first participant 2). According to the scheme, after the second participant is selected from the multiple participants based on the first video sharing docking station, the eyeball sight direction of the first participant to the second participant needs to be determined, and if the eyeball sight direction of the first participant to the second participant is the non-main visual angle direction, the follow-up operation of adjusting the non-main visual angle direction to be the main visual angle direction is executed.
In yet another possible implementation, the adjusting the eye gaze direction of the first participant currently directed to the second participant from the non-primary perspective direction to the primary perspective direction includes:
and adjusting the position and size ratio of the white eye and the pupil in the eyeball of the first participant in the conference video through a correction algorithm according to preset adjustment parameters, so that the eyeball sight line direction of the first participant for the second participant is changed from the non-main visual angle direction to the main visual angle direction.
In the above method, the specific process of the first video sharing docking station adjusting the current eyeball sight line direction of the first participant to the second participant from the non-main viewing angle direction to the main viewing angle direction may be: adjusting the position and size ratio of the white eye and the pupil in the eyeball of the first participant according to preset adjustment parameters through a correction algorithm, for example, if the first video sharing docking station determines that the eyeball sight line direction of the participant 1 currently watching the participant 2 in the conference video is 45 degrees south-west, the positions of the white eye and the pupil in the eyeball of the participant 1 are rotated by 45 degrees clockwise and then translated upwards by 1cm, and the white eye and the pupil are reduced by 1 time, so that the eyeball sight line direction of the participant 1 for the participant 2 is changed from the non-main visual angle direction to the main visual angle direction. This scheme can be to being adjusted into main visual angle direction by non-main visual angle direction first participant is present to the eyeball sight direction of second participant to adjustment meeting room user visual angle and remote user visual angle and the difference between the visual angle in the real face-to-face communication scene promote participant's experience and feel.
In a second aspect, an embodiment of the present application provides a viewing angle adjusting apparatus, where the viewing angle adjusting apparatus includes an obtaining unit, a determining unit, a selecting unit, and an adjusting unit, and is configured to implement the method described in the first aspect or any one of the possible implementation manners of the first aspect.
It should be noted that the processor included in the viewing angle adjusting apparatus described in the second aspect may be a processor dedicated to execute the methods (referred to as a special-purpose processor for convenience), or may be a processor, such as a general-purpose processor, that executes the methods by calling a computer program. Optionally, at least one processor may also include both special purpose and general purpose processors.
Alternatively, the computer program may be stored in a memory. For example, the Memory may be a non-transitory (non-transitory) Memory, such as a Read Only Memory (ROM), which may be integrated with the processor on the same device or separately disposed on different devices, and the embodiment of the present application is not limited to the type of the Memory and the arrangement manner of the Memory and the processor.
In a possible embodiment, the at least one memory is located outside the viewing angle adjusting device.
In yet another possible embodiment, the at least one memory is located within the viewing angle adjustment device.
In yet another possible implementation, a part of the memory of the at least one memory is located inside the angle-of-view adjusting apparatus, and another part of the memory is located outside the angle-of-view adjusting apparatus.
In this application, it is also possible that the processor and the memory are integrated in one device, i.e. that the processor and the memory are integrated together.
In a third aspect, embodiments of the present application provide a video sharing docking station, which includes a processor and a memory; the memory has stored therein a computer program; when the processor executes the computer program, the computing device performs the method described in the first aspect or any one of the first aspects.
In a fourth aspect, embodiments of the present application provide a computer-readable storage medium having stored therein instructions that, when executed on at least one processor, implement the method described in the first aspect or any one of the first aspects.
In a fifth aspect, the present application provides a computer program product comprising computer instructions that, when run on at least one processor, implement the method described in the first aspect or any of the first aspects. The computer program product may be a software installation package, which may be downloaded and executed on a computing device in case it is desired to use the method as described above.
The advantages of the technical methods provided in the second to fifth aspects of the present application may refer to the advantages of the technical solution of the first aspect, and are not described herein again.
Drawings
The drawings that are required to be used in the description of the embodiments will now be briefly described.
Fig. 1 is an application scenario diagram of a video sharing docking station according to an embodiment of the present application;
FIG. 2 is a schematic diagram of a system architecture for adjusting a viewing angle according to an embodiment of the present application;
fig. 3 is a schematic flowchart of a method for adjusting a viewing angle according to an embodiment of the present disclosure;
FIG. 4 is a schematic diagram of a method for determining a degree of matching of a perspective of a first participant with each of a plurality of participants according to an embodiment of the present application;
fig. 5 is a schematic view illustrating an adjustment of an eye sight direction according to an embodiment of the present disclosure;
fig. 6 is a schematic structural diagram of a viewing angle adjusting apparatus 60 according to an embodiment of the present disclosure;
fig. 7 is a schematic structural diagram of a video sharing docking station 70 according to an embodiment of the present application.
Detailed Description
The embodiments of the present application will be described below with reference to the drawings.
Referring to fig. 1, fig. 1 is a view illustrating an application scenario of a video sharing docking station according to an embodiment of the present application, where a conference scenario in fig. 1 specifically includes a first participant (a remote participant, such as participant 1), a plurality of participants (participants in a conference room, such as participant 2, participant 3, participant 4, and participant 5), a first video sharing docking station, a second video sharing docking station, a first display device (i.e., a display device used by a plurality of participants in the conference room for participating in a conference), and a second display device (i.e., a display device used by the first participant for participating in the conference), where the first video sharing docking station and the first display device are devices used by the plurality of participants for participating in the conference, the second video sharing docking station and the second display device are devices used by the first participant for participating in the conference, the first video sharing docking station, the second video sharing docking station, the first video sharing docking station, the second video sharing docking station, the first display device and the second display device establish communication with the first participant, and perform communication with the second participant through the first video sharing docking station, and the second display device.
Referring to fig. 2, fig. 2 is a schematic diagram of a system architecture for adjusting a viewing angle according to an embodiment of the present application, where the system includes a first video sharing docking station 201, a second video sharing docking station 202, a first display device 203, and a second display device 204. The first video sharing docking station 201, the second video sharing docking station 202, the first display device 203, and the second display device 204 establish a Communication connection, where the connection mode may be a connection of bluetooth, wireless, near Field Communication (NFC), ultra Wide Band (UWB), or other connection modes, and taking wireless connection as an example, the first video sharing docking station 201, the second video sharing docking station 202, the first display device 203, and the second display device 204 all access the same wireless network. First video sharing docking station 201 and first display device 203 are devices used by a plurality of participants to join the conference, and second video sharing docking station 202 and second display device 204 are devices used by the first participant to join the conference.
The first video sharing docking station 201 is configured to acquire behavior information of a first participant and behavior information of a plurality of participants, determine a communication closeness degree between the first participant and each of the plurality of participants according to the behavior information of the first participant and the behavior information of the plurality of participants, select a second participant from the plurality of participants according to the communication closeness degree between the first participant and each of the plurality of participants, and finally determine an eyeball sight line direction of the first participant currently directed at the second participant in the conference video.
Second video sharing docking station 202 is configured to send the captured eye movement trajectory image information of the first participant to first video sharing docking station 201, and send the captured area displayed on second display device 204 by the plurality of participants to first video sharing docking station 201, so that subsequent first video sharing docking station 201 determines the matching degree between the viewing angle of the first participant and each of the plurality of participants according to the eye movement trajectory image information of the first participant and the area displayed on second display device 204 by the plurality of participants, and finally combines the closeness between the first participant and each of the plurality of participants, the matching degree between the viewing angle of the first participant and each of the plurality of participants, and the second participant selected from the plurality of participants.
The first display device 203 is a device having processing capability and data transmission and reception capability, and can generate a transmission instruction and a reception instruction. For example, the first display device 203 may be a Computer, a notebook, a tablet, a palmtop, a desktop, a diagnostic, a mobile phone, an Ultra-mobile Personal Computer (UMPC), a netbook, a Personal Digital Assistant (PDA), or the like.
The second display device 204 is also a device having processing capability and data transceiving capability, and can generate a transmission instruction and a reception instruction. In this embodiment, the second display device 204 may be a large screen television for video conference in a conference room.
The method of the embodiments of the present application is described in detail below.
Referring to fig. 3, fig. 3 is a schematic flow chart of a method for adjusting a viewing angle according to an embodiment of the present disclosure. Alternatively, the method may employ the system described in fig. 2.
The method for adjusting the viewing angle as shown in fig. 3 at least includes steps S301 to S305.
Step S301: the first video sharing docking station obtains behavioral information of the first participant and behavioral information of the plurality of participants.
It should be noted that the first video sharing docking station is applied to a conference scene, the conference scene further includes a second video sharing docking station, a first display device (i.e., a display device in a conference room, where multiple participants are used for participating in the conference), and a second display device (i.e., a display device in a conference, where the first video sharing docking station and the first display device are devices used by the multiple participants to join in the conference, the second video sharing docking station and the second display device are devices used by the first participant to join in the conference, the first video sharing docking station, the second video sharing docking station, the first display device and the second display device establish communication connections, and taking a wireless connection as an example, the first video sharing docking station, the second video sharing docking station, the first display device and the second display device all access to the same wireless network. The first participant carries out conference communication with the plurality of participants through the second display device, and the plurality of participants carry out conference communication with the first participant through the first display device.
Specifically, the first video sharing docking station may acquire behavior information of the first participant and behavior information of the multiple participants, where the behavior information includes language information, motion information, and facial information, for example, during the video conference, the first video sharing docking station may acquire language information output by the first participant (e.g., participant 1) and the multiple participants (e.g., participant 2, participant 3, participant 4, and participant 5), for example, the content of question 1 presented to the multiple participants by participant 1 during the video conference, the content of verbal responses of participant 2 and participant 3 among the multiple participants to question 1, and the content of communication conversations conducted by participant 1 with participant 2, participant 3, participant 4, and participant 5, respectively.
As another example, during the video conference, the first video sharing docking station may also obtain the action information output by the first participant (e.g., participant 1) and the multiple participants (e.g., including participant 2, participant 3, participant 4, and participant 5), for example, when participant 1 proposes question 1, the first video sharing docking station may obtain whether any of the multiple participants answers question 1 through a positive feedback action, e.g., participant 2 and participant 3 interactively discuss and actively answer question 1 proposed by participant 1, and participant 4 and participant 5 keep question 1 proposed by participant 1 silent.
For example, when participant 1 asks for question 1, first video sharing docking station may capture facial expressions of participant 1 and participants to obtain which participants need to answer question 1 asked by participant 1 or which participants are interested in question 1 asked by participant 1, for example, when participant 2 and participant 3 ask for question 1, facial expressions are warm, and the question 1 is actively thought of, and when participant 1 asks for question 1, participant 4 and participant 5 have no obvious emotional feedback.
Step S302: the first video sharing docking station determines the closeness of communication between the first participant and each of the plurality of participants based on the behavioral information of the first participant and the behavioral information of the plurality of participants.
Specifically, the specific process of the first video sharing docking station determining the communication closeness degree between the first participant and each of the multiple participants according to the behavior information of the first participant and the behavior information of the multiple participants may be as follows: first, identity information of each participant in the multiple participants is obtained (for example, if the topic of the video conference is that each project group responsible person reports to the first participant, the first video sharing docking station can obtain identity information of the above four participants, and the participants with different identity information have different responsibility ranges, wherein participant 2 is responsible for project a, participant 3 is responsible for project B, participant 4 is responsible for project C, and participant 5 is responsible for project D, and the projects for which the above four participants are responsible are different), then a target keyword of the first participant in the conference process is generated according to the behavior information of the first participant and the behavior information of the multiple participants (wherein the target keyword may include a keyword of the content of question 1 provided by participant 1, for example, question 1 provided by participants is "how item a current project a is, whether the participant can be handed over in a week, the term of project B is tense, and the participant is in time to follow up", and the extracted keyword may be "keyword of the first video sharing docking station project a" a, question B "and keyword B" if the participant is a keyword of question 1, question of question 5, and the participant can answer the target keyword 2, and the keyword of the participant 2, if the participant can answer the target keyword of the participant 2, and the keyword of the participant to follow up the topic of the participant respectively, and finally, matching the extracted target keyword with the identity information of each participant in the multiple participants to obtain the communication intimacy degree between the first participant and each participant in the multiple participants (for example, the first video sharing docking station matches the extracted target keyword with the identity information of participant 2, participant 3, participant 4, and participant 5 to obtain the communication intimacy degrees between participant 1 and participant 2, participant 3, participant 4, and participant 5, which are respectively 80%, 30%, 56%, and 78%). According to the scheme, the communication closeness degree between the first participant and each participant in the multiple participants is determined according to the target keyword and the identity information of each participant in the multiple participants, and the accuracy of selecting the second participant from the multiple participants can be effectively improved.
Optionally, before the first video sharing docking station selects a second participant from the multiple participants according to the communication closeness between the first participant and each participant in the multiple participants, the first video sharing docking station may further receive first information and second information sent by the second video sharing docking station, where the first information is eyeball motion trajectory image information of the first participant collected by the second video sharing docking station, and the second information is an area displayed on the second display device by the multiple participants obtained by the second video sharing docking station; and then determining the matching degree of the visual angle of the first participant and each participant in the multiple participants according to the eyeball motion track image information of the first participant and the area displayed on the second display equipment by the multiple participants.
Specifically, the specific process of determining the matching degree between the visual angle of the first participant and each participant in the multiple participants by the first video sharing docking station according to the eyeball motion trajectory image information of the first participant and the area displayed on the second display device by the multiple participants may be: determining a variation of an eyeball sight line position of the first participant according to the eyeball movement track image information of the first participant (for example, the eyeball sight line direction in which the first participant gazes at the second display device is changed from 30 degrees to 45 degrees from north to south and then to exactly the west direction), determining a viewpoint position of the first participant falling on each of the plurality of participants according to the variation of the eyeball sight line position of the first participant and the position of each of the plurality of participants (for example, the position of participant 2 in the plurality of participants on the second display device is exactly the west direction, the position of participant 3 on the second display device is exactly the north direction, the position of participant 4 on the second display device is west direction, the position of participant 5 on the second display device is exactly east direction), determining a target display area in which the first participant is watched at the second display device according to the viewpoint position of each of the first participant in the plurality of participants, and determining the target display area in which the second participant is displayed on the second display device according to the first and the second display target display area of the plurality of participants.
The first video sharing docking station determines that the visual angle of the first participant is the highest matching degree of the second participant displayed on the second display device if the visual angle of the first participant is determined to be the area displayed on the second display device by the second participant according to the target area of the second display device watched by the first participant and the area displayed on the second display device by the plurality of participants.
In particular, the first video sharing docking station may determine the participant of the plurality of participants who has the highest degree of match with the perspective of the first participant by defining an area of coincidence between a target area at which the first participant gazes at the second display device and an area at which the plurality of participants are displayed on the second display device. For example, fig. 4 is a schematic diagram for determining a matching degree between a viewing angle of a first participant and each of multiple participants according to an embodiment of the present disclosure, and as shown in fig. 4, if the multiple participants include a participant 2, a participant 3, a participant 4, and a participant 5, a target area where the participant 1 gazes at the second display device is a western direction, and a position where the participant 2 is located on the second display device is a western direction (an area overlapping with the target area where the participant 1 gazes at the second display device is 5 cm) 2 ) The position of the participant 3 on the second display device is the due north direction (the area coinciding with the target area of the participant 1 looking at the second display device is 1 cm) 2 ) The position of the participant 4 on the second display device is in the northwest direction (the area coinciding with the target area of the participant 1 looking at the second display device is 2.8 cm) 2 ) The position of the participant 5 on the second display device is the east-ward direction (the area overlapping the target area where the participant 1 gazes at the second display device is 0 cm) 2 ) As can be seen from the above, the overlapping area between the target area where the participant 1 watches the second display device and the area where the participant 2 displays on the second display device is the largest, and then the matching degree between the viewing angle of the participant 1 and the participant 2 is the highest, and further the target area where the participant 1 watches the second display device and each participant displayThe area of overlap between the regions shown on the second display device determines that the degree of matching between the first participant's perspective and participant 2 is 95%, the degree of matching between the first participant's perspective and participant 3 is 10%, the degree of matching between the first participant's perspective and participant 4 is 54%, and the degree of matching between the first participant's perspective and participant 5 is 2%. According to the method and the device, the matching degree of the visual angle of the first participant and each participant in the multiple participants can be determined in a specific matching degree determining mode, and the accuracy of selecting the second participant from the multiple participants is improved.
Step S303: the first video sharing docking station selects the second participant from the plurality of participants based on the closeness of communication between the first participant and each of the plurality of participants.
Specifically, according to the scheme, the second participant is selected from the multiple participants according to the communication closeness degree between the first participant and each participant in the multiple participants, the visual angle of the first participant and the matching degree of each participant in the multiple participants, the second participant is selected from the multiple participants according to the two different dimensions of the communication closeness degree and the matching degree, and the accuracy of selecting the second participant from the multiple participants can be effectively improved.
Step S304: the first video sharing docking station determines an eye gaze direction of a first participant in the conference video that is currently directed at a second participant.
It should be noted that, after the first video sharing docking station selects the second participant from the multiple participants, it is further required to determine an eyeball sight direction of the first participant currently directed at the second participant to determine whether the eyeball sight direction of the first participant currently directed at the second participant is a main visual angle direction, and if the eyeball sight direction of the first participant currently directed at the second participant is the main visual angle direction, the first video sharing docking station does not need to adjust the eyeball sight direction; if the current eyeball sight direction of the first participant to the second participant is not the primary viewing direction, the first video sharing docking station needs to adjust the eyeball sight direction from the non-primary viewing direction to the primary viewing direction.
Specifically, the first video sharing docking station may calculate an offset value between the position of each of the participants and the position of the first display device (for example, if the first video sharing docking station first determines that the offset value between the position of the participant 2 and the position of the display device in the conference room is 6cm, the offset value between the position of the participant 3 and the position of the display device in the conference room is 1cm, the offset value between the position of the participant 4 and the position of the display device in the conference room is 2cm, and the offset value between the position of the participant 5 and the position of the display device in the conference room is 6 cm), where the offset value between the position of the participant 2 and the position of the first display device satisfies a preset condition (the preset condition may be that, for example, the offset value between the position of the participant 2 and the position of the first display device is greater than a preset 2 cm), and the first video sharing docking station determines that the current eye gaze direction of the participant (1) is the second video sharing direction (the main conference participant). According to the scheme, after the second participant is selected from the multiple participants based on the first video sharing docking station, the eyeball sight direction of the first participant to the second participant needs to be determined, and if the eyeball sight direction of the first participant to the second participant is a non-main visual angle direction, the follow-up operation of adjusting the non-main visual angle direction to be a main visual angle direction is executed.
Step S305: if the current eyeball sight direction of the first participant to the second participant is the non-main visual angle direction, the first video sharing docking station adjusts the current eyeball sight direction of the first participant to the second participant from the non-main visual angle direction to the main visual angle direction.
Specifically, the specific process of the first video sharing docking station adjusting the current eyeball sight direction of the first participant to the second participant from the non-main viewing angle direction to the main viewing angle direction may be: adjusting the position and size ratio of the white eye and the pupil in the eyeball of the first participant in the conference video through a correction algorithm according to preset adjustment parameters, for example, fig. 5 is a schematic diagram for adjusting the eye sight direction provided by the embodiment of the application, as shown in fig. 5, if the first video sharing docking station determines that the eye sight direction of the participant 1 currently watching the participant 2 in the conference video is 45 ° south away, the positions of the white eye and the pupil in the eyeball of the participant 1 are rotated by 45 ° clockwise and then translated by 1cm upwards, and the white eye and the pupil are reduced by 1 time, so that the eye sight direction of the participant 1 to the participant 2 is changed from the non-main visual angle direction to the main visual angle direction. This scheme can be to being adjusted into main visual angle direction by non-main visual angle direction first participant's eyeball sight direction to the second participant at present to adjustment meeting room user visual angle and remote user visual angle are with the difference between the visual angle in the real face-to-face communication scene, promote participant's experience and feel.
In a video conference scene, the difference exists between the visual angle of a conference room user and the visual angle of a remote user and the visual angle in a real face-to-face communication scene, so that the video conference experience of the user is poor. In the prior art, a human eye part of a human face part in a first image is adjusted to obtain a second image which corresponds to the first image and in which the human eye part of a user in the human face image is in a front-view state, so that in a video communication process, a video image acquired by an acquisition unit in real time is processed by adjusting parameters to obtain a corresponding image in the front-view state, and images displayed in a display unit are all adjusted and have a visual sense. According to the method, firstly, the first video sharing extension dock determines the communication closeness degree between a first participant and each of a plurality of participants according to the behavior information of the first participant (namely, a remote participant) and the behavior information of the plurality of participants (namely, participants in a conference room), then, according to the communication closeness degree between the first participant and each of the plurality of participants, a second participant is selected from the plurality of participants (namely, a person which the first participant actually wants to watch is determined to be a second participant according to the behavior information), and if the eyeball sight direction of the first participant relative to the second participant is determined to be a non-main visual angle direction (namely, the direction in which the first participant watches the second participant currently is determined to be a non-orthographic visual angle direction according to the angle of the first video sharing extension dock in the conference room), the first video sharing extension dock adjusts the eyeball sight direction of the first participant relative to the second participant to be the main visual angle direction. According to the method and the device, the difference between the visual angle of the meeting room user and the visual angle of the remote user in the real face-to-face communication scene can be adjusted in a targeted manner, and the experience of the meeting participants is improved.
The method of the embodiments of the present application is explained in detail above, and the apparatus of the embodiments of the present application is provided below.
It is to be understood that a plurality of apparatuses, such as a viewing angle adjusting apparatus, provided in the embodiments of the present application include a hardware structure, a software module, or a combination of a hardware structure and a software structure, which performs respective functions, in order to implement the functions in the above method embodiments.
Those of skill in the art will readily appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as hardware or combinations of hardware and computer software. Whether a function is performed as hardware or computer software drives hardware depends upon the particular application and design constraints imposed on the solution. A person skilled in the art may implement the foregoing method embodiments in different usage scenarios by using different device implementations, and the different implementation manners of the device should not be considered as exceeding the scope of the embodiments of the present application.
The embodiment of the application can divide the functional modules of the device. For example, each functional module may be divided for each function, or two or more functions may be integrated into one functional module. The integrated module can be realized in a hardware mode, and can also be realized in a software functional module mode. It should be noted that, in the embodiment of the present application, the division of the module is schematic, and is only one logic function division, and there may be another division manner in actual implementation.
For example, in the case where the respective functional modules of the apparatus are divided in an integrated manner, the present application exemplifies several possible processing apparatuses.
Referring to fig. 6, fig. 6 is a schematic structural diagram of a viewing angle adjusting apparatus 60 according to an embodiment of the present disclosure, where the viewing angle adjusting apparatus 60 may be the video sharing docking station shown in fig. 2 or a device in the video sharing docking station, such as a chip, a software module, an integrated circuit, and the like. The viewing angle adjusting device 60 is used to implement the aforementioned viewing angle adjusting method, such as the viewing angle adjusting method described in fig. 3.
In one possible embodiment, the viewing angle adjusting apparatus 60 may include an acquiring unit 601, a determining unit 602, a selecting unit 603, and an adjusting unit 604.
The acquiring unit 601 is configured to acquire behavior information of the first participant and behavior information of the plurality of participants, where the behavior information includes language information, motion information, and facial information;
the determining unit 602 is configured to determine, according to the behavior information of the first participant and the behavior information of the multiple participants, a communication closeness degree between the first participant and each of the multiple participants;
the selecting unit 603 is configured to select a second participant from the multiple participants according to the communication closeness between the first participant and each of the multiple participants;
the determining unit 602 is further configured to determine an eyeball sight direction of the first participant with respect to the second participant in the conference video, where the eyeball sight direction includes a main viewing direction or a non-main viewing direction;
if the current eyeball sight line direction of the first participant to the second participant is the non-main viewing angle direction, the adjusting unit 604 is configured to adjust the current eyeball sight line direction of the first participant to the second participant from the non-main viewing angle direction to the main viewing angle direction.
In a video conference scene, the user experience of the video conference is poor due to the difference between the visual angle of a conference room user and the visual angle of a remote user and the visual angle of a real face-to-face communication scene. In the prior art, a human eye part of a human face part in a first image is adjusted to obtain a second image which corresponds to the first image and in which the human eye part of a user in the human face image is in a front-view state, so that in a video communication process, a video image acquired by an acquisition unit in real time is processed by adjusting parameters to obtain a corresponding image in the front-view state, and images displayed in a display unit are all adjusted and have a visual sense. According to the method, firstly, the first video sharing extension dock determines the communication closeness degree between a first participant and each of a plurality of participants according to the behavior information of the first participant (namely, a remote participant) and the behavior information of the plurality of participants (namely, participants in a conference room), then, according to the communication closeness degree between the first participant and each of the plurality of participants, a second participant is selected from the plurality of participants (namely, a person which the first participant actually wants to watch is determined to be a second participant according to the behavior information), and if the eyeball sight direction of the first participant relative to the second participant is determined to be a non-main visual angle direction (namely, the direction in which the first participant watches the second participant currently is determined to be a non-orthographic visual angle direction according to the angle of the first video sharing extension dock in the conference room), the first video sharing extension dock adjusts the eyeball sight direction of the first participant relative to the second participant to be the main visual angle direction. The method and the device can adjust the difference between the visual angle of the user in the conference room and the visual angle of the remote user in the real face-to-face communication scene in a targeted manner, and the experience of the participants is improved.
In another possible embodiment, the system further comprises a receiving unit;
the receiving unit is configured to receive first information and second information sent by a second video sharing docking station, where the first information is acquired eyeball motion trail image information of the first participant, and the second information is an area displayed on the second display device by the acquired multiple participants;
the determining unit 602 is further configured to determine, according to the eye movement trajectory image information of the first participant and the area displayed on the second display device by the multiple participants, a matching degree between the viewing angle of the first participant and each of the multiple participants;
in said selecting a second participant from said plurality of participants based on closeness of communication between said first participant and each of said plurality of participants, said selecting unit 603 is specifically configured to:
and selecting the second participant from the plurality of participants according to the communication closeness degree between the first participant and each of the plurality of participants and the matching degree of the visual angle of the first participant and each of the plurality of participants.
In this embodiment, before the first video sharing docking station selects the second participant from the plurality of participants according to the closeness of communication between the first participant and each of the plurality of participants, the first video sharing docking station may further receive the eye movement track image information of the first participant and the area displayed on the second display device by the plurality of participants, which are sent by the second video sharing docking station, and determine the matching degree between the viewing angle of the first participant and each of the plurality of participants according to the eye movement track image information of the first participant and the area displayed on the second display device by the plurality of participants. Further, a second participant is selected from the plurality of participants based on the closeness of communication between the first participant and each of the plurality of participants, the perspective of the first participant, and the degree of matching of each of the plurality of participants. According to the scheme, the second participant is selected from the multiple participants according to the communication closeness degree between the first participant and each participant in the multiple participants, the visual angle of the first participant and the matching degree of each participant in the multiple participants, the second participant is selected from the multiple participants by integrating two different dimensions, and the accuracy of selecting the second participant from the multiple participants can be effectively improved.
In another possible implementation manner, in the aspect that the degree of matching between the visual angle of the first participant and each of the multiple participants is determined according to the information of the eye movement trajectory image of the first participant and the area displayed on the second display device by the multiple participants, the determining unit 602 is specifically configured to:
determining the variation of the eyeball sight line position of the first participant according to the eyeball motion track image information of the first participant;
determining a viewpoint position of the first participant falling on each of the plurality of participants according to the variation of the eyeball sight line position of the first participant;
determining a target area of the second display equipment watched by the first participant according to the viewpoint position of the first participant in each of the plurality of participants;
and determining the matching degree of the visual angle of the first participant and each participant in the multiple participants according to the target area of the second display equipment watched by the first participant and the area displayed on the second display equipment by the multiple participants.
In this embodiment of the application, the specific process of determining the matching degree between the viewing angle of the first participant and each participant in the multiple participants according to the eye movement track image information of the first participant and the area displayed on the second display device by the multiple participants may be as follows: determining a variation of an eyeball sight line position of the first participant according to the eyeball movement track image information of the first participant (for example, an eyeball sight line direction in which the first participant gazes at the second display device is changed from 30 degrees north to 45 degrees east to west), determining a viewpoint position in which the first participant falls on each of the multiple participants according to the variation of the eyeball sight line position of the first participant and a position in which each of the multiple participants is located (for example, a position in which a participant 2 of the multiple participants is located on the second display device is a true west direction, a position in which a participant 3 is located on the second display device is a true north direction, a position in which a participant 4 is located on the second display device is a west north direction, and a position in which a participant 5 is located on the second display device is a true east direction), then, determining a target area where the first participant watches the second display device according to the viewpoint position of each participant in the plurality of participants (determining that the target area where the first participant watches the second display device is the west direction according to the information), and finally determining the matching degree of the visual angle of the first participant and each participant in the plurality of participants according to the target area where the first participant watches the second display device and the area where the plurality of participants display on the second display device (namely, because the position of the participant 2 on the second display device is the west direction, the position of the participant 3 on the second display device is the north direction, the position of the participant 4 on the second display device is the north west direction, and the position of the participant 5 on the second display device is the east direction, and the target area in which the first participant gazes at the second display device is the west-positive direction, it is determined that the degree of matching between the visual angle of the first participant and participant 2 is 95%, the degree of matching between the visual angle of the first participant and participant 3 is 10%, the degree of matching between the visual angle of the first participant and participant 4 is 54%, and the degree of matching between the visual angle of the first participant and participant 5 is 2%). The accuracy of selecting the second participant from the multiple participants can be improved by determining the matching degree of the visual angle of the first participant and each participant in the multiple participants.
In yet another possible implementation, if the target area where the first participant gazes at the second display device is an area where the second participant is displayed on the second display device, the matching degree between the viewing angle of the first participant and the second participant is the highest.
In this embodiment, the first video sharing docking station may determine the participant with the highest degree of matching with the viewing angle of the first participant among the multiple participants by defining the area where the first participant gazes at the target area of the second display device and the area where the multiple participants are displayed on the second display device, for example, if the multiple participants include participant 2, participant 3, participant 4, and participant 5, the target area where participant 1 gazes at the second display device is in the western direction, and the position where participant 2 is located on the second display device is in the western direction (the area where participant 1 gazes at the target area of the second display device is 5 cm) 2 ) The position of the participant 3 on the second display device is the due north direction (the area overlapping with the target area of the participant 1 watching the second display device is 1 cm) 2 ) The position of the participant 4 on the second display device is in the northwest direction (the area coinciding with the target area of the participant 1 looking at the second display device is 2.8 cm) 2 ) The position of the participant 5 on the second display device is the east-ward direction (the area overlapping the target area where the participant 1 gazes at the second display device is 0 cm) 2 ) As can be seen from the above, the maximum overlapping area between the target area where the participant 1 watches the second display device and the area where the participant 2 is displayed on the second display device is, the matching degree between the viewing angle of the participant 1 and the participant 2 is the highest. According to the scheme, the accuracy rate of selecting the second participant from a plurality of participants can be improved by a specific matching degree determining mode.
In yet another possible implementation manner, in the determining the communication closeness between the first participant and each of the multiple participants according to the behavior information of the first participant and the behavior information of the multiple participants, the determining unit 602 is specifically configured to:
acquiring identity information of each participant in the multiple participants, wherein the responsibilities of the participants with different identity information are different;
generating target keywords of the first participant in a conference process according to the behavior information of the first participant and the behavior information of the multiple participants, wherein the target keywords comprise keywords of question contents provided by the first participant, keywords of answer contents of the questions by the multiple participants, and keywords of communication conversation contents of the first participant and each of the multiple participants;
and matching the target keyword with the identity information of each participant in the multiple participants to obtain the communication closeness degree between the first participant and each participant in the multiple participants.
In this embodiment, the specific process of the first video sharing docking station determining the communication closeness between the first participant and each of the multiple participants according to the behavior information of the first participant and the behavior information of the multiple participants may be: first, identity information of each of a plurality of participants is obtained (for example, if the subject of the conference video is a project report of each project group responsible person, the first video sharing docking station may obtain the identity information of the above four participants, and the participants with different identity information have different responsibility ranges, wherein participant 2 is responsible for project a, participant 3 is responsible for project B, participant 4 is responsible for project C, and participant 5 is responsible for project D, and the four responsible projects are different), then, a target keyword of the first participant in the conference process is generated according to the behavior information of the first participant and the behavior information of the plurality of participants (wherein the target keyword may include a keyword of the content of question 1 provided by participant 1; and may include keywords of the content of participant 2, participant 3, participant 4, and participant 5 respectively corresponding to question 1; and may include keywords of the content of participant 1 matched with participant 2, participant 3, participant 4, and participant 5 respectively, and the target communication degree of the communication between participant 1 and participant 2 is 30%, and 80% (for example, the content of the participant 2, participant 4%, and participant 5% is obtained as the last communication degree of the communication between the first and second participant (for the communication), and the communication degree of the communication between participant 2, and the participant 5%, and the last communication degree of the communication between the participant (for the communication between participant 2, and the participant 5%, and the participant) (for the communication degree of the communication between the participant 5%, and the participant). According to the method and the device, the communication closeness degree between the first participant and each participant in the multiple participants is determined according to the target keyword and the identity information of each participant in the multiple participants, and the accuracy rate of selecting the second participant from the multiple participants can be effectively improved.
In another possible implementation manner, in the aspect of determining the current eyeball sight direction of the first participant with respect to the second participant, the determining unit 602 is specifically configured to:
calculating a deviation value between the position of each participant in the plurality of participants and the position of the first display device;
and if the deviation value between the position of the second participant and the position of the first display device in the multiple participants meets a preset condition, determining that the current eyeball sight line direction of the first participant for the second participant is the non-main visual angle direction.
In the embodiment of the present application, after the first video sharing docking station selects the second participant from the multiple participants, it is further necessary to determine the eyeball sight direction of the first participant with respect to the second participant, specifically, the first video sharing docking station may first calculate an offset value between the position of each of the multiple participants and the position of the first display device (for example, the first video sharing docking station determines that the offset value between the position of participant 2 and the position of the display device in the conference room is 6cm, the offset value between the position of participant 3 and the position of the display device in the conference room is 1cm, the offset value between the position of participant 4 and the position of the display device in the conference room is 2cm, the offset value between the position of participant 5 and the position of the display device in the conference room is 6cm, and the offset value between the position of participant 2 and the display device in the conference room is 6 cm), where the offset value between the position of participant 2 and the position of the first display device satisfies a preset condition (that the eyeball sight-line sight direction of the first participant is larger than the offset value of the position of the second participant 2), and then determine that the offset value between the position of the first video sharing docking station (for the first video sharing device) and the second participant is larger than the position of the first display device (for the first video sharing docking station). According to the scheme, after the second participant is selected from the multiple participants based on the first video sharing docking station, the eyeball sight direction of the first participant to the second participant needs to be determined, and if the eyeball sight direction of the first participant to the second participant is the non-main visual angle direction, the follow-up operation of adjusting the non-main visual angle direction to be the main visual angle direction is executed.
In another possible implementation manner, in the aspect of adjusting the current eyeball sight line direction of the first participant with respect to the second participant from the non-main viewing angle direction to the main viewing angle direction, the adjusting unit 604 is specifically configured to:
and adjusting the position and size ratio of the white eye and the pupil in the eyeball of the first participant in the conference video through a correction algorithm according to preset adjustment parameters, so that the eyeball sight line direction of the first participant for the second participant is changed from the non-main visual angle direction to the main visual angle direction.
In this embodiment, the specific process of the first video sharing docking station adjusting the current eyeball sight direction of the first participant relative to the second participant from the non-primary viewing angle direction to the primary viewing angle direction may be: adjusting the position and size ratio of the white eye and the pupil in the eyeball of the first participant according to preset adjustment parameters through a correction algorithm, for example, if the first video sharing docking station determines that the eyeball sight line direction of the participant 1 currently watching the participant 2 in the conference video is 45 degrees south-west, the positions of the white eye and the pupil in the eyeball of the participant 1 are rotated by 45 degrees clockwise and then translated upwards by 1cm, and the white eye and the pupil are reduced by 1 time, so that the eyeball sight line direction of the participant 1 for the participant 2 is changed from the non-main visual angle direction to the main visual angle direction. This scheme can be to being adjusted into main visual angle direction by non-main visual angle direction first participant is present to the eyeball sight direction of second participant to adjustment meeting room user visual angle and remote user visual angle and the difference between the visual angle in the real face-to-face communication scene promote participant's experience and feel.
Referring to fig. 7, fig. 7 is a schematic structural diagram of a video sharing docking station 70 according to an embodiment of the present disclosure, such as a chip, a software module, an integrated circuit, and the like. The video sharing docking station 70 may include at least one processor 701. Optionally, at least one memory 703 may also be included. Further optionally, the video sharing docking station 70 may also include a communication interface 702. Still further optionally, a bus 704 may be included, wherein the processor 701, the communication interface 702, and the memory 703 are connected via the bus 704.
The processor 701 is a module for performing arithmetic operation and/or logical operation, and may specifically be one or a combination of multiple Processing modules, such as a Central Processing Unit (CPU), a picture Processing Unit (GPU), a Microprocessor (MPU), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA), a Complex Programmable Logic Device (CPLD), a coprocessor (assisting the CPU to complete corresponding Processing and Application), and a Micro Control Unit (MCU).
Communication interface 702 may be used to provide information input or output to the at least one processor. And/or, the communication interface 702 may be used to receive externally transmitted data and/or transmit data to the outside, and may be a wired link interface such as an ethernet cable, and may also be a wireless link (Wi-Fi, bluetooth, general wireless transmission, vehicle-mounted short-range communication technology, other short-range wireless communication technologies, and the like) interface. Optionally, the communication interface 702 may also include a transmitter (e.g., a radio frequency transmitter, an antenna, etc.) or a receiver, etc. coupled to the interface.
The memory 703 is used to provide a storage space in which data such as an operating system and computer programs can be stored. The Memory 703 may be one or a combination of Random Access Memory (RAM), read-only Memory (ROM), erasable Programmable Read-only Memory (EPROM), or portable Read-only Memory (CD-ROM), among others.
The at least one processor 701 in the video sharing docking station 70 is configured to perform the aforementioned method, such as the method described in the embodiment illustrated in fig. 3.
Alternatively, the processor 701 may be a processor dedicated to performing the methods (referred to as a special-purpose processor for convenience), or may be a processor for executing the methods by calling a computer program, such as a general-purpose processor. Optionally, at least one processor may also include both special purpose and general purpose processors. Optionally, in case the computing device comprises at least one processor 701, the computer program described above may be stored in the memory 703.
Optionally, the at least one processor 701 in the video sharing docking station 70 is configured to execute calling computer instructions to:
acquiring behavior information of the first participant and behavior information of the plurality of participants, wherein the behavior information comprises language information, action information and face information;
determining communication closeness between the first participant and each of the plurality of participants according to the behavior information of the first participant and the behavior information of the plurality of participants;
selecting a second participant from the plurality of participants based on the closeness of communication between the first participant and each of the plurality of participants;
determining an eyeball sight direction of the first participant to the second participant currently in the conference video, wherein the eyeball sight direction comprises a main visual angle direction or a non-main visual angle direction;
if the current eyeball sight line direction of the first participant for the second participant is the non-main visual angle direction, the current eyeball sight line direction of the first participant for the second participant is adjusted from the non-main visual angle direction to the main visual angle direction.
In a video conference scene, the user experience of the video conference is poor due to the difference between the visual angle of a conference room user and the visual angle of a remote user and the visual angle of a real face-to-face communication scene. In the prior art, the human eye part of the human face part in the first image is adjusted to obtain a second image which corresponds to the first image and in which the human eye part of the user in the human face image is in the front-view state, so that the video image acquired by the acquisition unit in real time is processed by adjusting parameters in the video communication process to obtain a corresponding image in the front-view state, and the images displayed in the display unit are all adjusted images with visual sense. According to the method, firstly, the first video sharing extension dock determines the communication closeness degree between a first participant and each of a plurality of participants according to the behavior information of the first participant (namely, a remote participant) and the behavior information of the plurality of participants (namely, participants in a conference room), then, according to the communication closeness degree between the first participant and each of the plurality of participants, a second participant is selected from the plurality of participants (namely, a person which the first participant actually wants to watch is determined to be a second participant according to the behavior information), and if the eyeball sight direction of the first participant relative to the second participant is determined to be a non-main visual angle direction (namely, the direction in which the first participant watches the second participant currently is determined to be a non-orthographic visual angle direction according to the angle of the first video sharing extension dock in the conference room), the first video sharing extension dock adjusts the eyeball sight direction of the first participant relative to the second participant to be the main visual angle direction. The method and the device can adjust the difference between the visual angle of the user in the conference room and the visual angle of the remote user in the real face-to-face communication scene in a targeted manner, and the experience of the participants is improved.
Optionally, the processor 701 is further configured to:
receiving first information and second information sent by a second video sharing docking station, wherein the first information is acquired eyeball motion track image information of the first participant, and the second information is an acquired area displayed on the second display device by the plurality of participants;
determining the matching degree of the visual angle of the first participant and each participant in the multiple participants according to the eyeball motion track image information of the first participant and the area displayed on the second display equipment by the multiple participants;
selecting a second participant from the plurality of participants based on the closeness of communication between the first participant and each of the plurality of participants, comprising:
and selecting the second participant from the plurality of participants according to the communication closeness degree between the first participant and each of the plurality of participants and the matching degree of the visual angle of the first participant and each of the plurality of participants.
In this embodiment, before the first video sharing docking station selects the second participant from the plurality of participants according to the closeness of communication between the first participant and each of the plurality of participants, the first video sharing docking station may further receive the eye movement track image information of the first participant and the area displayed on the second display device by the plurality of participants, which are sent by the second video sharing docking station, and determine the matching degree between the viewing angle of the first participant and each of the plurality of participants according to the eye movement track image information of the first participant and the area displayed on the second display device by the plurality of participants. Further, a second participant is selected from the plurality of participants based on the closeness of communication between the first participant and each of the plurality of participants, the degree of matching between the perspective of the first participant and each of the plurality of participants. According to the scheme, the second participant is selected from the multiple participants according to the communication closeness degree between the first participant and each participant in the multiple participants, the visual angle of the first participant and the matching degree of each participant in the multiple participants, the second participant is selected from the multiple participants by integrating two different dimensions, and the accuracy of selecting the second participant from the multiple participants can be effectively improved.
Optionally, the processor 701 is further configured to:
determining the variation of the eyeball sight line position of the first participant according to the eyeball motion track image information of the first participant;
determining a viewpoint position of the first participant falling on each of the plurality of participants according to the variation of the eyeball sight line position of the first participant;
determining a target area of the second display device watched by the first participant according to the viewpoint position of the first participant in each of the plurality of participants;
and determining the matching degree of the visual angle of the first participant and each participant in the multiple participants according to the target area of the second display equipment watched by the first participant and the area displayed on the second display equipment by the multiple participants.
In this embodiment, the specific process of determining, by the first video sharing docking station, the matching degree between the visual angle of the first participant and each participant in the multiple participants according to the eye movement track image information of the first participant and the area displayed on the second display device by the multiple participants may be: determining a variation of an eyeball sight line position of the first participant according to the eyeball movement track image information of the first participant (for example, an eyeball sight line direction in which the first participant gazes at the second display device is changed from 30 degrees north to 45 degrees east to west), determining a viewpoint position in which the first participant falls on each of the multiple participants according to the variation of the eyeball sight line position of the first participant and a position in which each of the multiple participants is located (for example, a position in which a participant 2 of the multiple participants is located on the second display device is a true west direction, a position in which a participant 3 is located on the second display device is a true north direction, a position in which a participant 4 is located on the second display device is a west north direction, and a position in which a participant 5 is located on the second display device is a true east direction), then, determining a target area where the first participant watches the second display device according to the viewpoint position of each participant in the plurality of participants (determining that the target area where the first participant watches the second display device is the west direction according to the information), and finally determining the matching degree of the visual angle of the first participant and each participant in the plurality of participants according to the target area where the first participant watches the second display device and the area where the plurality of participants display on the second display device (namely, because the position of the participant 2 on the second display device is the west direction, the position of the participant 3 on the second display device is the north direction, the position of the participant 4 on the second display device is the north west direction, and the position of the participant 5 on the second display device is the east direction, and the target area in which the first participant gazes at the second display device is the west-positive direction, it is determined that the degree of matching between the visual angle of the first participant and participant 2 is 95%, the degree of matching between the visual angle of the first participant and participant 3 is 10%, the degree of matching between the visual angle of the first participant and participant 4 is 54%, and the degree of matching between the visual angle of the first participant and participant 5 is 2%). The accuracy of selecting the second participant from the multiple participants can be improved by determining the matching degree of the visual angle of the first participant and each participant in the multiple participants.
Optionally, if the target area that the first participant gazes at the second display device is an area that the second participant displays on the second display device, the matching degree between the visual angle of the first participant and the second participant is the highest.
In this embodiment, the first video sharing docking station may determine the participant with the highest degree of matching with the viewing angle of the first participant among the multiple participants by defining the area where the first participant gazes at the target area of the second display device and the area where the multiple participants are displayed on the second display device, for example, if the multiple participants include participant 2, participant 3, participant 4, and participant 5, the target area where participant 1 gazes at the second display device is in the western direction, and the position where participant 2 is located on the second display device is in the western direction (the area where participant 1 gazes at the target area of the second display device is 5 cm) 2 ) The position of the participant 3 on the second display device is the due north direction (the area overlapping with the target area of the participant 1 watching the second display device is 1 cm) 2 ) The position of the participant 4 on the second display device is in the northwest direction (the area coinciding with the target area of the participant 1 looking at the second display device is 2.8 cm) 2 ) The position of the participant 5 on the second display device is the east-ward direction (the area overlapping the target area where the participant 1 gazes at the second display device is 0 cm) 2 ) As can be seen from the above, the maximum overlapping area between the target area where the participant 1 watches the second display device and the area where the participant 2 is displayed on the second display device is, the highest matching degree between the viewing angle of the participant 1 and the participant 2 is obtained. Book (I)The scheme can improve the accuracy of selecting the second participant from a plurality of participants through a specific matching degree determining mode.
Optionally, the processor 701 is further configured to:
acquiring identity information of each participant in the multiple participants, wherein the responsibilities of the participants with different identity information are different;
generating target keywords of the first participant in a conference process according to the behavior information of the first participant and the behavior information of the multiple participants, wherein the target keywords comprise keywords of question contents provided by the first participant, keywords of answer contents of the questions by the multiple participants, and keywords of communication conversation contents of the first participant and each of the multiple participants;
and matching the target keyword with the identity information of each participant in the multiple participants to obtain the communication closeness degree between the first participant and each participant in the multiple participants.
In this embodiment, the specific process of the first video sharing docking station determining the communication closeness between the first participant and each of the multiple participants according to the behavior information of the first participant and the behavior information of the multiple participants may be: first, identity information of each of a plurality of participants is obtained (for example, if the subject of the conference video is a project report of each project group responsible person, the first video sharing docking station may obtain the identity information of the above four participants, and the participants with different identity information have different responsibility ranges, wherein participant 2 is responsible for project a, participant 3 is responsible for project B, participant 4 is responsible for project C, and participant 5 is responsible for project D, and the four responsible projects are different), then, a target keyword of the first participant in the conference process is generated according to the behavior information of the first participant and the behavior information of the plurality of participants (wherein the target keyword may include a keyword of the content of question 1 provided by participant 1; and may include keywords of the content of participant 2, participant 3, participant 4, and participant 5 respectively corresponding to question 1; and may include keywords of the content of participant 1 matched with participant 2, participant 3, participant 4, and participant 5 respectively, and the target communication degree of the communication between participant 1 and participant 2 is 30%, and 80% (for example, the content of the participant 2, participant 4%, and participant 5% is obtained as the last communication degree of the communication between the first and second participant (for the communication), and the communication degree of the communication between participant 2, and the participant 5%, and the last communication degree of the communication between the participant (for the communication between participant 2, and the participant 5%, and the participant) (for the communication degree of the communication between the participant 5%, and the participant). According to the method and the device, the communication closeness degree between the first participant and each participant in the multiple participants is determined according to the target keyword and the identity information of each participant in the multiple participants, and the accuracy rate of selecting the second participant from the multiple participants can be effectively improved.
Optionally, the processor 701 is further configured to:
calculating a deviation value between the position of each participant in the plurality of participants and the position of the first display device;
and if the deviation value between the position of the second participant and the position of the first display device in the multiple participants meets a preset condition, determining that the current eyeball sight line direction of the first participant for the second participant is the non-main visual angle direction.
In this embodiment, after the first video sharing docking station selects the second participant from the plurality of participants, it is further required to determine an eyeball sight direction of the first participant with respect to the second participant, specifically, the first video sharing docking station may first calculate an offset value between the position of each participant in the plurality of participants and the position of the first display device (for example, the first video sharing docking station determines that the offset value between the position of the participant 2 in the plurality of participants (including the participant 2, the participant 3, the participant 4, the participant 5) and the position of the display device in the conference room is 6cm, the offset value between the position of the participant 3 and the position of the display device in the conference room is 1cm, the offset value between the position of the participant 4 and the position of the display device in the conference room is 2cm, the offset value between the position of the participant 5 and the position of the display device in the conference room is 6 cm), where the offset value between the position of the participant 2 and the position of the display device in the conference room satisfies a predetermined condition that the eyeball sight direction of the first participant (for example, the first video sharing docking station satisfies a predetermined condition that the first video sharing direction of the first participant 2 is greater than the first video sharing direction (for the first participant 2). According to the scheme, after the second participant is selected from the multiple participants based on the first video sharing docking station, the eyeball sight direction of the first participant to the second participant needs to be determined, and if the eyeball sight direction of the first participant to the second participant is the non-main visual angle direction, the follow-up operation of adjusting the non-main visual angle direction to be the main visual angle direction is executed.
Optionally, the processor 701 is further configured to:
and adjusting the position and size ratio of the white eye and the pupil in the eyeball of the first participant in the conference video through a correction algorithm according to preset adjustment parameters, so that the eyeball sight line direction of the first participant for the second participant is changed from the non-main visual angle direction to the main visual angle direction.
In this embodiment of the application, a specific process of the first video sharing docking station adjusting the current eyeball sight direction of the first participant to the second participant from the non-primary viewing angle direction to the primary viewing angle direction may be: adjusting the position and size ratio of the white eye and the pupil in the eyeball of the first participant according to preset adjustment parameters through a correction algorithm, for example, if the first video sharing docking station determines that the eyeball sight line direction of the participant 1 currently watching the participant 2 in the conference video is 45 degrees south-west, the positions of the white eye and the pupil in the eyeball of the participant 1 are rotated by 45 degrees clockwise and then translated upwards by 1cm, and the white eye and the pupil are reduced by 1 time, so that the eyeball sight line direction of the participant 1 for the participant 2 is changed from the non-main visual angle direction to the main visual angle direction. This scheme can be to being adjusted into main visual angle direction by non-main visual angle direction first participant is present to the eyeball sight direction of second participant to adjustment meeting room user visual angle and remote user visual angle and the difference between the visual angle in the real face-to-face communication scene promote participant's experience and feel.
The present application further provides a computer-readable storage medium having stored therein instructions that, when executed on at least one processor, implement the aforementioned perspective adjustment method, such as the method described in fig. 3.
The present application also provides a computer program product comprising computer instructions, which when executed by a computing device, implement the aforementioned method of adjusting a viewing angle, such as the method described in fig. 3.
In the embodiments of the present application, words such as "for example" or "like" are used to mean serving as an example, instance, or illustration. Any embodiment or design described herein as "for example" or "such as" is not necessarily to be construed as preferred or advantageous over other embodiments or designs. Rather, use of the words "for example" or "such as" are intended to present relevant concepts in a concrete fashion.
In the present application, the embodiments refer to "at least one" and "a plurality" and two or more. "at least one of the following" or similar expressions refer to any combination of these items, including any combination of the singular or plural items. For example, at least one (one) of a, b, or c, may represent: a. b, c, (a and b), (a and c), (b and c), or (a and b and c), wherein a, b and c can be single or multiple. "and/or" describes the association relationship of the associated object, indicating that there may be three relationships, for example, a and/or B, which may indicate: a alone, A and B together, and B alone, wherein A, B may be singular or plural. The character "/" generally indicates that the former and latter associated objects are in an "or" relationship.
And unless stated to the contrary, the ordinal numbers such as "first", "second", etc. are used in the embodiments of the present application to distinguish a plurality of objects and are not used to limit the sequence, timing, priority, or importance of the plurality of objects. For example, a first device and a second device are for convenience of description only and do not represent differences in structure, importance, etc. of the first device and the second device, and in some embodiments, the first device and the second device may be the same device.
As used in the above embodiments, the term "when … …" may be interpreted to mean "if … …" or "after … …" or "in response to determination … …" or "in response to detection of … …", depending on the context. The above description is only exemplary of the present application and is not intended to limit the present application, and any modifications, equivalents, improvements, etc. made within the spirit and principles of the present application are intended to be included within the scope of the present application.
It will be understood by those skilled in the art that all or part of the steps for implementing the above embodiments may be implemented by hardware, or may be implemented by a program instructing relevant hardware, where the program may be stored in a computer-readable storage medium, and the storage medium may be a read-only memory, a magnetic disk or an optical disk.
While the invention has been described with reference to specific embodiments, the scope of the invention is not limited thereto, and those skilled in the art can easily conceive various equivalent modifications or substitutions within the technical scope of the invention. Therefore, the protection scope of the present application shall be subject to the protection scope of the claims.

Claims (10)

1. A method for adjusting a viewing angle, the method being applied to a first video sharing docking station, the first video sharing docking station being applied to a conference scene, the conference scene further including a second video sharing docking station, a first display device and a second display device, the first video sharing docking station and the first display device being devices used by a plurality of participants to join a conference, the second video sharing docking station and the second display device being devices used by a first participant to join a conference, the first video sharing docking station, the second video sharing docking station, the first display device and the second display device establishing a communication connection, the first participant having conference communication with the plurality of participants through the second display device, the plurality of participants having conference communication with the first participant through the first display device, the method comprising:
acquiring behavior information of the first participant and behavior information of the plurality of participants, wherein the behavior information comprises language information, action information and face information;
determining communication closeness between the first participant and each of the plurality of participants according to the behavior information of the first participant and the behavior information of the plurality of participants;
selecting a second participant from said plurality of participants based on closeness of communication between said first participant and each of said plurality of participants;
determining an eyeball sight direction of the first participant to the second participant currently in the conference video, wherein the eyeball sight direction comprises a main visual angle direction or a non-main visual angle direction;
if the current eyeball sight line direction of the first participant for the second participant is the non-main visual angle direction, the current eyeball sight line direction of the first participant for the second participant is adjusted from the non-main visual angle direction to the main visual angle direction.
2. The method of claim 1, further comprising:
receiving first information and second information sent by a second video sharing docking station, wherein the first information is acquired eyeball motion track image information of the first participant, and the second information is an acquired area displayed on the second display device by the plurality of participants;
determining the matching degree of the visual angle of the first participant and each participant in the multiple participants according to the eyeball motion track image information of the first participant and the area displayed on the second display equipment by the multiple participants;
selecting a second participant from the plurality of participants based on the closeness of communication between the first participant and each of the plurality of participants, comprising:
and selecting the second participant from the plurality of participants according to the communication closeness degree between the first participant and each of the plurality of participants and the matching degree of the visual angle of the first participant and each of the plurality of participants.
3. The method of claim 2, wherein said determining the degree of match of the perspective of the first participant with each of the plurality of participants based on the eye movement path image information of the first participant and the area of the plurality of participants displayed on the second display device comprises:
determining the variation of the eyeball sight line position of the first participant according to the eyeball motion track image information of the first participant;
determining the viewpoint position of the first participant falling on each participant in the plurality of participants according to the variation of the eyeball sight-line position of the first participant;
determining a target area of the second display device watched by the first participant according to the viewpoint position of the first participant in each of the plurality of participants;
and determining the matching degree of the visual angle of the first participant and each participant in the multiple participants according to the target area of the second display equipment watched by the first participant and the area displayed on the second display equipment by the multiple participants.
4. The method of claim 3, wherein:
and if the target area of the second display equipment watched by the first participant is the area displayed on the second display equipment by the second participant, the matching degree of the visual angle of the first participant and the second participant is the highest.
5. The method of claim 1 or 2, wherein determining the closeness of communication between the first participant and each of the plurality of participants based on the behavioral information of the first participant and the behavioral information of the plurality of participants comprises:
acquiring identity information of each participant in the multiple participants, wherein the responsibilities of the participants with different identity information are different;
generating target keywords of the first participant in a conference process according to the behavior information of the first participant and the behavior information of the multiple participants, wherein the target keywords comprise keywords of question contents provided by the first participant, keywords of answer contents of the questions by the multiple participants, and keywords of communication conversation contents of the first participant and each of the multiple participants;
and matching the target keyword with the identity information of each participant in the multiple participants to obtain the communication closeness degree between the first participant and each participant in the multiple participants.
6. The method of claim 1, wherein the determining the current eye gaze direction of the first participant with respect to the second participant comprises:
calculating a deviation value between the position of each participant in the plurality of participants and the position of the first display device;
and if the deviation value between the position of the second participant and the position of the first display device in the multiple participants meets a preset condition, determining that the current eyeball sight line direction of the first participant for the second participant is the non-main visual angle direction.
7. The method of claim 1, wherein the adjusting the eye gaze direction of the first participant currently directed to the second participant from the non-primary viewing direction to the primary viewing direction comprises:
and adjusting the position and size ratio of the white eye and the pupil in the eyeball of the first participant in the conference video through a correction algorithm according to preset adjustment parameters, so that the eyeball sight line direction of the first participant for the second participant is changed from the non-main visual angle direction to the main visual angle direction.
8. A visual angle adjusting apparatus applied to a first video sharing docking station, where the first video sharing docking station is applied to a conference scene, where the conference scene further includes a second video sharing docking station, a first display device and a second display device, where the first video sharing docking station and the first display device are devices used by a plurality of participants to join a conference, the second video sharing docking station and the second display device are devices used by a first participant to join a conference, the first video sharing docking station, the second video sharing docking station, the first display device and the second display device establish a communication connection, the first participant performs conference communication with the plurality of participants through the second display device, and the plurality of participants performs conference communication with the first participant through the first display device, the apparatus comprising an acquiring unit, a determining unit, a selecting unit and an adjusting unit, where:
the acquisition unit is used for acquiring the behavior information of the first participant and the behavior information of the plurality of participants, wherein the behavior information comprises language information, action information and face information;
the determining unit is used for determining the communication closeness degree between the first participant and each participant in the participants according to the behavior information of the first participant and the behavior information of the participants;
the selection unit is used for selecting a second participant from the multiple participants according to the communication closeness degree between the first participant and each participant in the multiple participants;
the determining unit is further configured to determine an eyeball sight direction of the first participant with respect to the second participant currently in the conference video, where the eyeball sight direction includes a dominant perspective direction or a non-dominant perspective direction;
if the current eyeball sight line direction of the first participant for the second participant is a non-main visual angle direction, the adjusting unit is used for adjusting the current eyeball sight line direction of the first participant for the second participant from the non-main visual angle direction to the main visual angle direction.
9. A video sharing docking station characterized in that it is a first video sharing docking station comprising a processor and a memory, the memory for storing computer instructions, the processor for invoking the computer instructions to implement the method of any of claims 1-7.
10. A computer-readable storage medium having stored therein instructions which, when executed on at least one processor, implement the method of any one of claims 1-7.
CN202211461514.8A 2022-11-16 2022-11-16 Visual angle adjusting method and device Active CN115546877B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211461514.8A CN115546877B (en) 2022-11-16 2022-11-16 Visual angle adjusting method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211461514.8A CN115546877B (en) 2022-11-16 2022-11-16 Visual angle adjusting method and device

Publications (2)

Publication Number Publication Date
CN115546877A CN115546877A (en) 2022-12-30
CN115546877B true CN115546877B (en) 2023-03-31

Family

ID=84719923

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211461514.8A Active CN115546877B (en) 2022-11-16 2022-11-16 Visual angle adjusting method and device

Country Status (1)

Country Link
CN (1) CN115546877B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116489502B (en) * 2023-05-12 2023-10-31 深圳星河创意科技开发有限公司 Remote conference method based on AI camera docking station and AI camera docking station

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0730877A (en) * 1993-07-12 1995-01-31 Oki Electric Ind Co Ltd Inter-multi location multimedia communications conference system
WO2021227988A1 (en) * 2020-05-11 2021-11-18 华为技术有限公司 Face image processing method, apparatus and device, and computer readable storage medium
CN114554135A (en) * 2022-02-28 2022-05-27 联想(北京)有限公司 Online conference method and electronic equipment

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6753900B2 (en) * 2002-05-07 2004-06-22 Avaya Techology Corp. Method and apparatus for overcoming the limitations of camera angle in video conferencing applications

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0730877A (en) * 1993-07-12 1995-01-31 Oki Electric Ind Co Ltd Inter-multi location multimedia communications conference system
WO2021227988A1 (en) * 2020-05-11 2021-11-18 华为技术有限公司 Face image processing method, apparatus and device, and computer readable storage medium
CN114554135A (en) * 2022-02-28 2022-05-27 联想(北京)有限公司 Online conference method and electronic equipment

Also Published As

Publication number Publication date
CN115546877A (en) 2022-12-30

Similar Documents

Publication Publication Date Title
US20210343041A1 (en) Method and apparatus for obtaining position of target, computer device, and storage medium
US20170195650A1 (en) Method and system for multi point same screen broadcast of video
CN111510645B (en) Video processing method and device, computer readable medium and electronic equipment
US20200094138A1 (en) Game Picture Display Method and Apparatus, Storage Medium and Electronic Device
CN110022363B (en) Method, device and equipment for correcting motion state of virtual object and storage medium
CN115546877B (en) Visual angle adjusting method and device
US10304352B2 (en) Electronic device and method for sharing image
CN109660738B (en) Exposure control method and system based on double cameras
EP4000700A1 (en) Camera shot movement control method, device, apparatus, and storage medium
CN111836114A (en) Video interaction method and device, electronic equipment and storage medium
CN109639987B (en) Bracelet shooting method and device and computer readable storage medium
CN110969061A (en) Neural network training method, neural network training device, visual line detection method, visual line detection device and electronic equipment
US10114221B1 (en) Toggling a 360-degree camera to selfie view
CN112702533B (en) Sight line correction method and sight line correction device
WO2017138143A1 (en) Terminal device, method for adjusting screen resolution, display control method, screen resolution adjustment program, and display control program
CN112533070B (en) Video sound and picture adjusting method, terminal and computer readable storage medium
US11042215B2 (en) Image processing method and apparatus, storage medium, and electronic device
CN112307941A (en) Image conversion method and terminal equipment
US10922825B2 (en) Image data processing method and electronic device
US11876632B2 (en) Audio transcription for electronic conferencing
KR101964446B1 (en) Art Activity System Using Virtual Reality
KR20170127354A (en) Apparatus and method for providing video conversation using face conversion based on facial motion capture
CN113362243A (en) Model training method, image processing method and apparatus, medium, and electronic device
KR20200129885A (en) Online remote group exercise system
CN111582822A (en) AR-based conference method and device and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant