CN114500911A - Method, device, equipment, system and medium for realizing video call - Google Patents

Method, device, equipment, system and medium for realizing video call Download PDF

Info

Publication number
CN114500911A
CN114500911A CN202210088623.3A CN202210088623A CN114500911A CN 114500911 A CN114500911 A CN 114500911A CN 202210088623 A CN202210088623 A CN 202210088623A CN 114500911 A CN114500911 A CN 114500911A
Authority
CN
China
Prior art keywords
target
subsystems
candidate
image
relative position
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210088623.3A
Other languages
Chinese (zh)
Inventor
闫文龙
张建伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Deep Vision Technology Nanjing Co ltd
Original Assignee
Deep Vision Technology Nanjing Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Deep Vision Technology Nanjing Co ltd filed Critical Deep Vision Technology Nanjing Co ltd
Priority to CN202210088623.3A priority Critical patent/CN114500911A/en
Publication of CN114500911A publication Critical patent/CN114500911A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/141Systems for two-way working between two video terminals, e.g. videophone
    • H04N7/142Constructional details of the terminal equipment, e.g. arrangements of the camera and the display
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/141Systems for two-way working between two video terminals, e.g. videophone
    • H04N7/147Communication arrangements, e.g. identifying the communication as a video-communication, intermediate storage of the signals

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Abstract

The embodiment of the invention discloses a method, a device, equipment, a system and a medium for realizing video call. Wherein the method is performed by a current subsystem, comprising: receiving the relative position of a first target sent by other subsystems, and determining a first target binocular video stream according to the relative position of the first target; sending the first target binocular video stream to other subsystems so that the other subsystems display the first target binocular video stream through first 3D display equipment; sending the relative position of the second target to other subsystems so that the other subsystems determine a second target binocular video stream according to the relative position of the second target; and receiving second target binocular video streams sent by other subsystems, and displaying the second target binocular video streams through second 3D display equipment. By executing the scheme, the 3D video call can be realized, and the stereoscopic watching experience of a video call user can be guaranteed while the bandwidth resource is saved.

Description

Method, device, equipment, system and medium for realizing video call
Technical Field
The embodiment of the invention relates to the technical field of 3D video call, in particular to a method, a device, equipment, a system and a medium for realizing video call.
Background
The 3D display device utilizes the principle that the angles of the human eyes for observing an object are slightly different, so that the distance of the object can be distinguished, the stereoscopic vision is generated, and the images seen by the left eye and the right eye are separated, so that a user can experience stereoscopic sensation by means of stereoscopic glasses or without the aid of stereoscopic glasses (namely naked eyes). 3D live broadcast display service can be realized through 3D display equipment.
The realization of 3D video call in the related art is limited to a scene that one end shoots and the other end watches. The method is not applicable to a scene that two ends simultaneously carry out video calls, and is not friendly to the user experience.
Disclosure of Invention
Embodiments of the present invention provide a method, an apparatus, a device, a system, and a medium for implementing a video call, which can implement a 3D video call, save bandwidth resources, and ensure stereoscopic viewing experience of a video call user.
In a first aspect, an embodiment of the present invention provides a method for implementing a video call, where the method is executed by a current subsystem, and the method includes:
receiving a first target relative position sent by other subsystems, and determining a first target binocular video stream according to the first target relative position; wherein the first target relative position is determined by a first user position determining device in the other subsystem;
sending the first target binocular video stream to the other subsystems so that the other subsystems display the first target binocular video stream through first 3D display equipment;
sending the relative position of the second target to the other subsystems so that the other subsystems determine a second target binocular video stream according to the relative position of the second target; wherein the second target relative position is determined by a second user position determination device in the current subsystem;
and receiving second target binocular video streams sent by the other subsystems, and displaying the second target binocular video streams through second 3D display equipment.
In a second aspect, an embodiment of the present invention further provides an apparatus for implementing a video call, where the apparatus is configured in a current subsystem, and the apparatus includes:
the first target binocular video stream determining module is used for receiving the relative positions of the first targets sent by other subsystems and determining a first target binocular video stream according to the relative positions of the first targets; wherein the first target relative position is determined by a first user position determining device in the other subsystem;
the first target binocular video stream sending module is used for sending the first target binocular video stream to the other subsystems so that the other subsystems can display the first target binocular video stream through first 3D display equipment;
the second target relative position determining module is used for sending the second target relative position to the other subsystems so that the other subsystems can determine a second target binocular video stream according to the second target relative position; wherein the second target relative position is determined by a second user position determination device in the current subsystem;
and the display module is used for receiving the second target binocular video stream sent by the other subsystems and displaying the second target binocular video stream through second 3D display equipment.
In a third aspect, an embodiment of the present invention further provides an electronic device, where the electronic device includes:
one or more processors;
a storage device for storing one or more programs,
when executed by the one or more processors, cause the one or more processors to implement a method for implementing a video call as described in any of the embodiments of the present invention as performed by the current subsystem.
In a fourth aspect, an embodiment of the present invention further provides a system for implementing a video call, where the system includes at least two subsystems, and includes: the system comprises a current subsystem and other subsystems, wherein the other subsystems comprise first user position determining equipment, first 3D display equipment, a first server and a first image collector;
the current subsystem comprises a second user position determining device, a second 3D display device, a second server and a second image collector;
the first user position determining device determines a first target relative position of a user relative to a screen and sends the first target relative position to the second server;
the second server determines a first target binocular video stream acquired by a second image acquisition device according to the relative position of the first target, and sends the first target binocular video stream to the first server;
the first 3D display device is used for carrying out 3D display on the first target binocular video stream in the first server;
the second server further sends a second target relative position determined by the second user position determining device to the first server;
the first server determines a second target binocular video stream acquired by the first image acquirer according to the relative position of the second target, and sends the second target binocular video stream to the second server;
and the second 3D display equipment is used for carrying out 3D display on the second target binocular video stream in the second server.
In a fifth aspect, an embodiment of the present invention further provides a computer-readable storage medium, on which a computer program is stored, where the computer program is used to implement, when executed by a processor, a method for implementing a video call performed by a current subsystem according to any one of the embodiments of the present invention.
According to the technical scheme provided by the embodiment of the invention, when the current subsystem is executed, the first target relative position sent by other subsystems is received, and the first target binocular video stream is determined according to the first target relative position; sending the first target binocular video stream to other subsystems so that the other subsystems display the first target binocular video stream through first 3D display equipment; sending the relative position of the second target to other subsystems so that the other subsystems determine a second target binocular video stream according to the relative position of the second target; and receiving second target binocular video streams sent by other subsystems, and displaying the second target binocular video streams through second 3D display equipment. By executing the scheme, the 3D video call can be realized, and the stereoscopic watching experience of a video call user can be guaranteed while the bandwidth resource is saved.
Drawings
Fig. 1 is a flowchart of a method for implementing a video call executed by a current subsystem according to an embodiment of the present invention;
fig. 2 is a schematic view of a 3D video call scene according to an embodiment of the present invention;
fig. 3 is a flowchart of another method for implementing a video call performed by a current subsystem according to an embodiment of the present invention;
fig. 4 is a schematic structural diagram of an apparatus for implementing a video call configured in a current subsystem according to an embodiment of the present invention;
fig. 5 is a schematic structural diagram of an electronic device according to an embodiment of the present invention;
fig. 6 is a schematic structural diagram of a system for implementing a video call according to an embodiment of the present invention.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings and examples. It is to be understood that the specific embodiments described herein are merely illustrative of the invention and are not limiting of the invention. It should be further noted that, for the convenience of description, only some structures related to the present invention are shown in the drawings, not all of them.
Fig. 1 is a flowchart of a method for implementing a video call, which is performed by a current subsystem according to an embodiment of the present invention, and the method may be performed by an apparatus for implementing a video call, which may be implemented by software and/or hardware, and may be used in an electronic device for implementing a video call. The method is applied to the scene of the 3D video call. As shown in fig. 1, the technical solution provided by the embodiment of the present invention specifically includes:
and S110, receiving the relative positions of the first targets sent by other subsystems, and determining a first target binocular video stream according to the relative positions of the first targets.
Wherein the first target relative position is determined by a first user position determining device in the other subsystem.
Illustratively, the current subsystem and the other subsystems are respectively related equipment at the end of the video call user. The current subsystem and other subsystems comprise a server, a 3D display device, a user position determining device and an image collector array for collecting image data of a shooting area where a video user is located. Wherein, the image collector array can be arranged in an arc shape. The server in the other subsystem is the first server. The 3D display device in the other subsystem is the first 3D display device. The user position determining device in the other subsystem is the first user position determining device. And the image collector in other subsystems is a first image collector. The server in the current subsystem is the second server. The 3D display device in the current subsystem is the second 3D display device. The user position determining device in the current subsystem is the second user position determining device. The image collector in the current subsystem is the second image collector. The relative position of the first target can be set according to actual needs, and can be coordinate information of the head, two eyes or one eye of other video users in a screen reference frame. The screen reference system is a coordinate system taking the screen center of other subsystems as a coordinate origin. The first target relative position may also be the angle of the head, eyes or monocular of other video users with respect to the screen center normal of other subsystems. The first target relative position may also be an offset of the head, eyes or eyes of other video users relative to the screen of the other subsystem.
Wherein, the current subsystem and other subsystems can have two data output links including a control signal link and a video stream link. Wherein the control signal link may be used to communicate a target relative position of the video call user with respect to the screen between the other subsystem and the current subsystem. The video stream link is used for transmitting the video streams shot at two ends between the other subsystem and the current subsystem. The scheme can receive the relative position of the first target sent by other subsystems, and determine the binocular video stream of the first target according to the relative position of the first target. Specifically, according to the scheme, at least two image collector serial numbers corresponding to the relative position of the first target can be determined according to the corresponding position relationship between the relative position of the first target and the placement of the second image collector, then the image data collected by the at least two image collectors is obtained from the second server, the image data is processed to obtain a first target binocular video stream, and the first target binocular video stream is respectively transmitted back or combined and then transmitted back to other subsystems for display.
As shown in fig. 2, for example, when the video call user a is at position 1, the opposite-end subsystem (the subsystem on the side of the video call user B) determines that the two corresponding video streams are the video streams captured by the image collector No. 9 and the image collector No. 8, and sends the video streams captured by the image collector No. 9 and the image collector No. 8 to the 3D display subsystem at one end of the video call user a for display.
When the video call user A moves from the position 1 to the position 2, the video call user A indicates that the watching visual angle of the video call user A is expected to rotate to the left, at the moment, the opposite terminal subsystem (the subsystem on the side of the video call user B) determines that the two corresponding video streams are the video streams shot by the No. 8 image collector and the No. 7 image collector respectively, and sends the video streams shot by the No. 8 image collector and the No. 7 image collector to the 3D display subsystem of the video call user A to be displayed respectively or in a combined mode.
Similarly, when the video call user B moves from the position 3 to the position 4, the two-way video stream sent by the video call user a terminal system to the video call user B terminal system is also switched to the video stream shot by the image collector No. 4 and the image collector No. 3, and then is switched to the video stream shot by the image collector No. 3 and the image collector No. 2.
The first user position determining device may be, for example, a user position tracking sensor, and the user position tracking sensor may be, for example, a tracking sensor based on a single camera or at least two cameras, or a tracking sensor based on a combination of an image sensor and a depth information sensor, or an eye tracking sensor based on a gaze point tracking. A first target relative position of the other video users with respect to the screen may be determined by the first user position determining device.
And S120, sending the first target binocular video stream to the other subsystems so that the other subsystems display the first target binocular video stream through first 3D display equipment.
Wherein, the first 3D display device can realize the 3D effect display of the first target binocular video stream. The first 3D display device may be a naked eye 3D display or 3D glasses. Wherein, the naked eye 3D display may be a lenticular type naked eye 3D display. The naked eye 3D display may also be a parallax barrier naked eye 3D display. The naked eye 3D display may also be a backlit naked eye 3D display. The 3D glasses may be polarized glasses or shutter 3D glasses. According to the scheme, the first target binocular video stream can be sent to other subsystems, and the other subsystems can carry out 3D display on the first target binocular video stream through the first 3D display equipment.
For example, when the naked eye 3D display is used for 3D display, the first 3D display device may further multiplex the first target relative position of the other video user with respect to the screen, which is obtained by the first user position determining device, and adjust the layout parameter of the 3D display or adjust the position of the relevant device in the naked eye 3D display, so that the other video user may obtain an ideal 3D viewing effect at the current position. When 3D glasses are used for 3D display, other video users can obtain an ideal 3D watching effect at the current position in a mode of 3D matching of the glasses.
And S130, sending the relative position of the second target to the other subsystems so that the other subsystems can determine a second target binocular video stream according to the relative position of the second target.
Wherein the second target relative position is determined by a second user position determination device in the current subsystem.
The relative position of the second target may be set according to actual needs, for example, the relative position of the second target may be coordinate information of the head, eyes or a single eye of the current video user in a screen reference frame. The screen reference system is a coordinate system taking the screen center of the current subsystem as a coordinate origin. The second target relative position may also be an angle of the head, both eyes or a single eye of the current video user with respect to the screen center normal in the current subsystem. The second target relative position may also be an offset of the head, eyes or monocular of the current video user relative to the screen of the current subsystem. The relevant introduction of the second user position determining device may refer to the detailed introduction of the first user position determining device. The scheme can send the relative position of the second target to other subsystems, so that the other subsystems determine the binocular video stream of the second target according to the relative position of the second target. Specifically, as shown in fig. 2, the other subsystems may determine at least two image collector serial numbers corresponding to the relative position of the second target according to the corresponding position relationship of the relative position of the second target and the placement of the first image collector in the other subsystems, then obtain image data collected by the at least two image collectors from the first server, further process the image data to obtain a second target binocular video stream, and respectively return or merge the second target binocular video stream and then return to the current subsystem for display.
And S140, receiving the second target binocular video stream sent by the other subsystems, and displaying the second target binocular video stream through second 3D display equipment.
Wherein the second 3D display device may implement 3D effect display of the second target binocular video stream. The second 3D display device may be a naked eye 3D display or 3D glasses. According to the scheme, the second target binocular video stream sent by other subsystems can be received, and the second target binocular video stream is displayed through the second 3D display equipment.
For example, when the second 3D display device is a naked eye 3D display, the second 3D display device may further multiplex the relative position of the second target, and adjust the layout parameter of the 3D display or adjust the position of a relevant device in the naked eye 3D display, so that the current position of the current video user may obtain an ideal 3D viewing effect. When 3D glasses are used for 3D display, other video users can obtain an ideal 3D watching effect at the current position in a mode of 3D matching of the glasses.
It should be noted that the sequence of the current subsystem and other subsystems is not limited, and the current subsystem and other subsystems can be operated simultaneously or sequentially.
According to the technical scheme provided by the embodiment of the invention, when the current subsystem is executed, the first target relative position sent by other subsystems is received, and the first target binocular video stream is determined according to the first target relative position; sending the first target binocular video stream to other subsystems so that the other subsystems display the first target binocular video stream through first 3D display equipment; sending the relative position of the second target to other subsystems so that the other subsystems determine a second target binocular video stream according to the relative position of the second target; and receiving second target binocular video streams sent by other subsystems, and displaying the second target binocular video streams through second 3D display equipment. By executing the scheme, the 3D video call can be realized, and the stereoscopic watching experience of a video call user can be guaranteed while the bandwidth resource is saved.
Fig. 3 is a flowchart of a method for implementing a video call executed by a current subsystem according to an embodiment of the present invention, which is optimized based on the foregoing embodiments. As shown in fig. 3, the method for implementing a video call in the embodiment of the present invention may include:
and S210, receiving the relative positions of the first targets sent by other subsystems, and determining at least two target image collectors from the candidate image collectors of the current subsystem according to the relative positions of the first targets and the incidence relation between the candidate relative positions in the other subsystems and the candidate image collector in the current subsystem.
Wherein the first target relative position comprises at least an offset direction of the other user relative to the screen of the other subsystem.
The number of candidate image collectors in the current subsystem is at least three, and the candidate image collectors may be arranged in an arc shape, as shown in fig. 2. The scheme can pre-establish the incidence relation between the candidate relative positions in other subsystems and the candidate image collector in the current subsystem. Since the relative positions of other users relative to the screens of other subsystems are continuously distributed, and the set of shooting angles formed by all image collectors in the current subsystem is usually discretely distributed, the association relationship can be expressed in that a plurality of candidate relative positions of other subsystems correspond to one image collector of the current subsystem. The scheme can receive multiple paths of videos shot by the image collector array, determine at least two suitable target image collectors from the candidate image collectors of the current subsystem according to the incidence relation between the first target relative position and the candidate relative position in the other subsystems and the candidate image collectors in the current subsystem, and process image data collected by the target image collectors into two paths of videos to be output or combined and then output to the other subsystems.
In this embodiment, optionally, before determining at least two target image collectors from the candidate image collectors of the current subsystem according to the first target relative position and the association relationship between the candidate relative positions in the other subsystems and the candidate image collector in the current subsystem, the method further includes: and determining the incidence relation between the candidate relative positions in the other subsystems and the first candidate image collector in the current subsystem according to the viewing angles of the other users on the candidate relative positions, the offset directions of the other users relative to the screens of the other subsystems and the image collecting angle of the first candidate image collector of the current subsystem.
And the viewing angles of the other users in the candidate relative position can be included angles of the other users in the candidate relative position relative to the normal of the screen center of the other subsystem. The offset direction of the other user relative to the screen of the other subsystem may be that the other user is offset to the left relative to the screen of the other subsystem, or that the other user is offset to the right relative to the screen of the other subsystem. And the image acquisition angle of the candidate image acquirer of the current subsystem can be determined according to the arc-shaped position information of the candidate image acquirer arrangement of the current subsystem. According to the scheme, if the viewing angles of other users at the candidate relative positions are determined to be consistent with the image acquisition angle of the first candidate image acquirer of the current subsystem, and the offset directions of the other users at the candidate relative positions relative to the screen of the other subsystem are determined to be consistent with the offset direction of the first candidate image acquirer of the current subsystem relative to the current user, the incidence relation between the candidate relative positions in the other subsystems and the first candidate image acquirer of the current subsystem is determined.
Or, if it is determined that the viewing angles of the other users at the candidate relative positions are within the image acquisition angle interval of the current subsystem determined by the two first candidate image collectors, and the offset directions of the other users at the candidate relative positions relative to the screens of the other subsystems are consistent with the offset directions of the two candidate image collectors of the current subsystem relative to the current user, determining the association relationship between the candidate relative positions in the other subsystems and the first candidate image collector of the current subsystem. The included angle is used as the relative position of the user relative to the screen, and is closer to the arc-shaped characteristics of the image collector array arrangement, so that the accuracy of the server for determining the target image collector can be further improved, and the complexity of the matching algorithm can be reduced.
Therefore, the incidence relation between the candidate relative positions in the other subsystems and the first candidate image collector in the current subsystem is determined according to the viewing angles of the other users on the candidate relative positions, the offset directions of the other users relative to the screens of the other subsystems and the image collecting angle of the first candidate image collector of the current subsystem. The target binocular video stream can be accurately determined, and reliable data support is provided for normal display of a 3D video call.
In a possible embodiment, optionally, determining the association relationship between the candidate relative positions in the other subsystems and the candidate image collector in the current subsystem according to the viewing angles of the other users at the candidate relative positions, the offset directions of the other users relative to the screens of the other subsystems, and the image collection angle of the candidate image collector in the current subsystem includes: and if the viewing angles of other users at the candidate relative positions are consistent with the image acquisition angle of the first candidate image acquirer in the current subsystem, and the offset directions of the other users relative to the screen are consistent with the offset direction of the first candidate image acquirer in the current subsystem relative to the current user, establishing the association relationship between the candidate relative positions in the other subsystems and the first candidate image acquirer in the current subsystem.
Illustratively, as shown in fig. 2, the viewing angles of the other users at the candidate relative positions are known, for example, the viewing angles of the other users at the candidate relative positions may be 20 degrees. And the offset direction of other users with respect to the screen center of other subsystems is known, for example, to the left of the screen, or to the right of the screen. Similarly, as shown in fig. 2, the current subsystem arranges 5 first candidate image collectors uniformly in an arc array, the image capturing angle of the first candidate image collectors with respect to the current user is known, and the offset direction of the first candidate image collectors with respect to the current user may be determined, for example, to the left of the viewing area or to the right of the viewing area. Therefore, if the viewing angle of the other user at the candidate relative position is consistent with the image capturing angle of the first candidate image collector of the current subsystem, and the offset direction of the other user relative to the screen of the other subsystem is also consistent with the offset direction of the first image collector of the current subsystem relative to the screen of the current subsystem, the association relationship between the candidate relative position in the other subsystem and the first candidate image collector in the current subsystem can be established. Therefore, if the viewing angle of the other user at the candidate relative position is consistent with the image acquisition angle of the first candidate image acquirer in the current subsystem, and the offset direction of the other user relative to the screen is consistent with the offset direction of the first candidate image acquirer in the current subsystem relative to the current user, the association relationship between the candidate relative position in the other subsystem and the first candidate image acquirer in the current subsystem is established. The method and the device can accurately determine the target binocular video stream matched with the video call user, and further provide reliable data support for normal display of the 3D video call.
In this embodiment, optionally, before determining at least two target image collectors from the candidate image collectors of the current subsystem according to the first target relative position and the incidence relation between the candidate relative positions in the other subsystems and the candidate image collector in the current subsystem, the method further includes: and determining the incidence relation between the candidate relative positions in the other subsystems and the second candidate image collector in the current subsystem according to the viewing angles of the other users on the candidate relative positions, the offset directions of the other users relative to the screens of the other subsystems and the distance between the current user and the second candidate image collector of the current subsystem.
In two images obtained by shooting an object in front of a screen through two image collectors, the parallax size of the object on the two images is related to the spacing distance between the two image collectors. For example, if the separation distance between two image collectors is larger, the parallax of the object on the two images is larger. When the human eyes watch the stereo image formed by two images, the requirement of the parallax range of the object on the image is provided by the normal stereo perception. Therefore, when a farther object is shot, the distance between the two image collectors is larger, and the parallax formed by the two shot images can be comfortably distinguished by the front and rear distance information of a viewer.
According to the scheme, the second candidate image collector in the current subsystem matched with the candidate relative position can be determined according to the viewing angles of other users at the candidate relative position and the offset directions of the other users relative to the screens of the other subsystems, the image parallax determined by any two second candidate image collectors in each second candidate image collector is determined according to the distance between the current user and the second candidate image collector of the current subsystem, and if the image parallax is consistent with the preset image parallax, the incidence relation between the candidate relative position in the other subsystems and the second candidate image collector in the current subsystem is determined.
Therefore, the incidence relation between the candidate relative position in the other subsystem and the second candidate image collector in the current subsystem is determined according to the viewing angle of the other user on the candidate relative position, the offset direction of the other user relative to the screen of the other subsystem and the distance between the current user and the second candidate image collector in the current subsystem. The target binocular video stream can be accurately determined, and reliable data support is provided for normal display of a 3D video call.
In another possible embodiment, optionally, determining the association relationship between the candidate relative positions in the other subsystems and the candidate image collector in the current subsystem according to the viewing angles of the other users at the candidate relative positions, the offset directions of the other users relative to the screens of the other subsystems, and the distance between the current user and the candidate image collector in the current subsystem includes: determining a second candidate image collector in the target angle of the current subsystem according to the viewing angle of other users at the candidate relative position and the offset direction of other users relative to the screen; determining the distance between a second candidate image collector in the target angle and the current user, and determining the image parallax determined by any two second candidate image collectors in the target angle according to the distance; and if the image parallax is determined to be consistent with the preset image parallax, establishing the incidence relation between the candidate relative positions in other subsystems and the two second candidate image collectors in the current subsystem within the target angle.
In the present application, when determining the association relationship between the candidate relative positions of the other subsystems and the candidate image collector of the current subsystem, the distance between the position of the current user and the image collector in the current subsystem may be further considered in addition to the viewing angle of the other users, the offset direction of the other users relative to the screen of the other subsystems, the image collection angle of the candidate image collector of the current subsystem, and the offset direction of the candidate image collector in the current subsystem relative to the current user in the above-mentioned candidate relative positions. The farther the position of the current user is from the image collector in the current subsystem, the images shot by the two image collectors with larger distance can be used to synthesize the first target binocular video stream.
The target angle may be 30 degrees to the left relative to the screen in the current subsystem, or 40 degrees to the right relative to the screen in the current subsystem, and the target angle may be determined by the viewing angle of the other user at the candidate relative position and the offset direction of the other user relative to the screen. The preset image parallax may be a parallax range of an object on an image with normal stereoscopic sensation when human eyes watch a stereoscopic image formed by two images. According to the scheme, the candidate image collector in the target angle in the current subsystem, namely the second candidate image collector, can be determined according to the viewing angle of other users at the candidate relative position and the offset direction of other users relative to the screen. And further determining the distance between the second candidate image collector in the target angle and the current user and the image data collected by the second candidate image collector in the target angle. And then combining any two image data in the target angle to obtain an image parallax, comparing the image parallax with a preset image parallax, and if the image parallax is consistent with the preset image parallax, establishing an association relation between a second candidate image collector in the current subsystem and candidate relative positions in other subsystems for obtaining the two image data.
For example, in the scenario shown in fig. 2, assuming that the video call user a still moves to the left to position 2, and the video call user B moves away from the screen in the current subsystem in this example, the first target binocular video stream is spliced by using the pictures collected by the number 6 image collector and the number 9 image collector, which have similar overall shooting angles but are separated by a larger distance, instead of being spliced by using the pictures collected by the number 7 image collector and the number 8 image collector. This helps to ensure that the captured parallax distance information of the video call user B is comfortably perceived by the video call user a.
Thus, a second candidate image collector in the target angle of the current subsystem is determined according to the viewing angle of the other user at the candidate relative position and the offset direction of the other user relative to the screen; determining the distance between a second candidate image collector in the target angle and the current user, and determining the image parallax determined by any two second candidate image collectors in the target angle according to the distance; and if the image parallax is determined to be consistent with the preset image parallax, establishing the incidence relation between the candidate relative positions in other subsystems and two second candidate image collectors in the current subsystem in the target angle. The method and the device can accurately determine the target binocular video stream matched with the video call user, and further provide reliable data support for normal display of the 3D video call.
And S220, determining a first target binocular video stream according to the image data collected by the at least two target image collectors.
In this embodiment, optionally, the disparity between the video contents determined by the image data collected by the at least two target image collectors is smaller than or equal to the binocular disparity of the user.
If the parallax between the video contents determined by the image data collected by the at least two target image collectors is smaller than the binocular parallax of the user, the current subsystem can perform operations such as merging processing on the image data as required to obtain the video contents with parallax equivalent to the binocular parallax of the user, and then respectively output or merge the video streams and output the video contents to other subsystems. If the parallax between the video contents determined by the image data collected by the at least two target image collectors has parallax equivalent to the parallax of the eyes of the person, the server can output the video streams to the second subsystem respectively or after merging.
Thus, the parallax between the video contents determined by the image data collected by the at least two target image collectors is set to be less than or equal to the user binocular parallax. 3D display can be realized according to the human eye imaging principle, and reliable data support is provided for normal display of 3D video call.
And S230, sending the first target binocular video stream to the other subsystems so that the other subsystems display the first target binocular video stream through first 3D display equipment.
And S240, sending the relative position of the second target to the other subsystems so that the other subsystems can determine a second target binocular video stream according to the relative position of the second target.
And S250, receiving the second target binocular video stream sent by the other subsystems, and displaying the second target binocular video stream through second 3D display equipment.
According to the technical scheme provided by the embodiment of the invention, when the current subsystem is executed, the first target relative position sent by other subsystems is received, and at least two target image collectors are determined from the candidate image collectors of the current subsystem according to the first target relative position and the incidence relation between the candidate relative positions in the other subsystems and the candidate image collectors in the current subsystem; wherein the first target relative position at least comprises the offset direction of other users relative to the screen of other subsystems; determining the first target binocular video stream according to image data acquired by at least two target image collectors; sending the first target binocular video stream to other subsystems so that the other subsystems display the first target binocular video stream through first 3D display equipment; sending the relative position of the second target to other subsystems so that the other subsystems determine a second target binocular video stream according to the relative position of the second target; and receiving second target binocular video streams sent by other subsystems, and displaying the second target binocular video streams through second 3D display equipment. By executing the scheme, the 3D video call can be realized, and the stereoscopic watching experience of a video call user can be guaranteed while the bandwidth resource is saved.
Fig. 4 is a schematic structural diagram of an apparatus for implementing a video call configured in a current subsystem according to an embodiment of the present invention, where the apparatus may be configured in an electronic device for implementing a video call. As shown in fig. 4, the apparatus includes:
the first target binocular video stream determining module 410 is configured to receive the first target relative position sent by the other subsystem, and determine a first target binocular video stream according to the first target relative position; wherein the first target relative position is determined by a first user position determining device in the other subsystem;
a first target binocular video stream sending module 420, configured to send the first target binocular video stream to the other subsystems, so that the other subsystems display the first target binocular video stream through a first 3D display device;
the second target relative position determining module 430 is configured to send the second target relative position to the other subsystems, so that the other subsystems determine a second target binocular video stream according to the second target relative position; wherein the second target relative position is determined by a second user position determination device in the current subsystem;
and the display module 440 is configured to receive the second target binocular video stream sent by the other subsystem, and display the second target binocular video stream through a second 3D display device.
Optionally, the first target binocular video stream determining module 410 includes a target image collector determining unit, configured to determine at least two target image collectors from the candidate image collectors of the current subsystem according to the relative position of the first target and the association relationship between the candidate relative positions in the other subsystems and the candidate image collectors of the current subsystem; wherein the first target relative position comprises at least an offset direction of the other user relative to a screen of the other subsystem; and the target binocular video stream determining unit is used for determining the first target binocular video stream according to the image data acquired by the at least two target image acquirers.
Optionally, the apparatus further includes a first association determining module, configured to determine, before determining at least two target image collectors from the candidate image collectors of the current subsystem according to the first target relative position and the association between the candidate relative positions in the other subsystems and the candidate image collector in the current subsystem, the association between the candidate relative positions in the other subsystems and the first candidate image collector in the current subsystem according to the viewing angles of the other users at the candidate relative positions, the offset directions of the other users relative to the screens of the other subsystems, and the image collecting angle of the first candidate image collector of the current subsystem.
Optionally, the first association relation determining module is specifically configured to, if the viewing angle of the other user at the candidate relative position is consistent with the image acquisition angle of the first candidate image acquirer in the current subsystem, and the offset direction of the other user with respect to the screen is consistent with the offset direction of the first candidate image acquirer in the current subsystem with respect to the current user, establish an association relation between the candidate relative position in the other subsystem and the first candidate image acquirer in the current subsystem.
Optionally, the apparatus further includes a second association relation determining module, configured to determine, before determining at least two target image collectors from the candidate image collectors of the current subsystem according to the first target relative position and the association relation between the candidate relative positions in the other subsystems and the candidate image collector in the current subsystem, the association relation between the candidate relative positions in the other subsystems and the second candidate image collector in the current subsystem according to the viewing angle of the other users at the candidate relative positions, the offset direction of the other users relative to the screens of the other subsystems, and the distance between the current user and the second candidate image collector of the current subsystem.
Optionally, the second association relation determining module includes a second candidate image collector determining unit, configured to determine a second candidate image collector within a target angle in the current subsystem according to the viewing angle of the other user at the candidate relative position and the offset direction of the other user relative to the screen; the image parallax determining unit is used for determining the distance between the second candidate image collectors in the target angle and the current user and determining the image parallax determined by any two second candidate image collectors in the target angle according to the distance; and the second incidence relation determining unit is used for establishing incidence relations between the candidate relative positions in other subsystems and the two second candidate image collectors in the current subsystem within the target angle if the image parallax is determined to be consistent with the preset image parallax.
Optionally, the first target relative position includes: the angle of the other users relative to the normal of the center of the screen of the other subsystem or the offset of the other users relative to the screen of the other subsystem.
Optionally, the disparity between the video contents determined by the image data collected by the at least two target image collectors is smaller than or equal to the binocular disparity of the user.
The device provided by the above embodiment can execute the method for implementing the video call executed by the current subsystem provided by any embodiment of the present invention, and has the corresponding functional modules and beneficial effects of the execution method.
Fig. 5 is a schematic structural diagram of an electronic device according to an embodiment of the present invention, and as shown in fig. 5, the electronic device includes:
one or more processors 510, one processor 510 being illustrated in FIG. 5;
a memory 520;
the apparatus may further include: an input device 530 and an output device 540.
The processor 510, the memory 520, the input device 530 and the output device 540 of the apparatus may be connected by a bus or other means, and fig. 5 illustrates the connection by a bus as an example.
The memory 520 is a non-transitory computer-readable storage medium, and can be used to store software programs, computer-executable programs, and modules, such as program instructions/modules corresponding to a method for implementing a video call in an embodiment of the present invention. The processor 510 executes various functional applications and data processing of the computer device by executing the software programs, instructions and modules stored in the memory 520, namely, implementing a method for implementing a video call executed by a current subsystem, of the above method embodiments, that is:
receiving a first target relative position sent by other subsystems, and determining a first target binocular video stream according to the first target relative position; wherein the first target relative position is determined by a first user position determining device in the other subsystem;
sending the first target binocular video stream to the other subsystems so that the other subsystems display the first target binocular video stream through first 3D display equipment;
sending the relative position of the second target to the other subsystems so that the other subsystems determine a second target binocular video stream according to the relative position of the second target; wherein the second target relative position is determined by a second user position determination device in the current subsystem;
and receiving second target binocular video streams sent by the other subsystems, and displaying the second target binocular video streams through second 3D display equipment.
The memory 520 may include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application program required for at least one function; the storage data area may store data created according to use of the computer device, and the like. Further, the memory 520 may include high speed random access memory, and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid state storage device. In some embodiments, memory 520 may optionally include memory located remotely from processor 510, which may be connected to a terminal device via a network. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof.
The input device 530 may be used to receive input numeric or character information and generate key signal inputs related to user settings and function control of the computer apparatus. The output device 540 may include a display device such as a display screen.
Fig. 6 is a schematic structural diagram of a system for implementing a video call according to an embodiment of the present invention, and as shown in fig. 6, the system includes at least two subsystems, including: a current subsystem 61 and other subsystems 62, the other subsystems 62 including a first user position determining device 621, a first 3D display device 622, a first server 623 and a first image collector 624;
the current subsystem 61 comprises a second user position determining device 611, a second 3D display device 612, a second server 613 and a second image collector 614;
wherein the first user position determining device 621 determines a first target relative position of the other user with respect to the screen of the other subsystem, and sends the first target relative position to the second server 613;
the second server 613 determines a first target binocular video stream acquired by the second image acquirer 614 according to the first target relative position, and sends the first target binocular video stream to the first server 623;
the first 3D display device 622, which performs 3D display on the first target binocular video stream in the first server 623;
the second server 613, further sending the second target relative position determined by the second user position determining device 611 to the first server 623;
the first server 623 determines a second target binocular video stream acquired by the first image acquirer 624 according to the second target relative position, and sends the second target binocular video stream to the second server 613;
the second 3D display device 612 performs 3D display on the second target binocular video stream in the second server 613.
The system comprises at least two subsystems, wherein any one of the at least two subsystems can be used as a current subsystem in the embodiment of the application, and at least one other subsystem is used as the other subsystem.
The video call implementation system and the video call implementation method provided by the embodiment belong to the same public concept, and details which are not described in detail in the system embodiment may refer to the description in the embodiment.
Optionally, the second user position determination device 611 comprises at least one of a tracking sensor consisting of at least one camera, a tracking sensor consisting of an image and depth information sensor and an eye tracking sensor based on gaze point tracking.
The eyeball tracking technology belongs to machine vision, an image of an eyeball is captured through an image sensor, characteristics in the pupil of the eyeball of each person are identified according to the processing of the image, and a fixation point of a watching screen is calculated in real time through the characteristics. The method has the advantages of high precision, low time delay and large visual field range.
Therefore, by using a tracking sensor composed of at least one camera or a tracking sensor composed of an image and depth information sensor or an eyeball tracking sensor based on gaze point tracking as the user position determination equipment, the target relative position of the user relative to the screen can be accurately obtained, reliable data support is further provided for determining binocular video streams according to the target relative position, and the quality of 3D video calls can be guaranteed.
The technical scheme provided by the embodiment of the invention comprises at least two subsystems, including: the system comprises a current subsystem and other subsystems, wherein the other subsystems comprise first user position determining equipment, first 3D display equipment, a first server and a first image collector; the current subsystem comprises a second user position determining device, a second 3D display device, a second server and a second image collector; the first user position determining device determines a first target relative position of a user relative to a screen and sends the first target relative position to the second server; the second server determines a first target binocular video stream acquired by the second image acquisition device according to the relative position of the first target and sends the first target binocular video stream to the first server; the first 3D display equipment is used for carrying out 3D display on the first target binocular video stream in the first server; the second server is used for sending the second target relative position determined by the second user position determining equipment to the first server; the first server determines a second target binocular video stream acquired by the first image acquisition device according to the relative position of the second target and sends the second target binocular video stream to the second server; and the second 3D display equipment is used for carrying out 3D display on the second target binocular video stream in the second server. By executing the scheme, the 3D video call can be realized, and the stereoscopic watching experience of a video call user can be guaranteed while the bandwidth resource is saved.
The embodiment of the invention provides a computer readable storage medium, on which a computer program is stored, and when the program is executed by a processor, the method for implementing video call executed by a current subsystem, as provided by the embodiment of the invention, is implemented:
receiving a first target relative position sent by other subsystems, and determining a first target binocular video stream according to the first target relative position; wherein the first target relative position is determined by a first user position determining device in the other subsystem;
sending the first target binocular video stream to the other subsystems so that the other subsystems display the first target binocular video stream through first 3D display equipment;
sending the relative position of the second target to the other subsystems so that the other subsystems determine a second target binocular video stream according to the relative position of the second target; wherein the second target relative position is determined by a second user position determination device in the current subsystem;
and receiving second target binocular video streams sent by the other subsystems, and displaying the second target binocular video streams through second 3D display equipment.
Any combination of one or more computer-readable media may be employed. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C + + or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any type of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet service provider).
It is to be noted that the foregoing is only illustrative of the preferred embodiments of the present invention and the technical principles employed. It will be understood by those skilled in the art that the present invention is not limited to the particular embodiments described herein, but is capable of various obvious changes, rearrangements and substitutions as will now become apparent to those skilled in the art without departing from the scope of the invention. Therefore, although the present invention has been described in greater detail by the above embodiments, the present invention is not limited to the above embodiments, and may include other equivalent embodiments without departing from the spirit of the present invention, and the scope of the present invention is determined by the scope of the appended claims.

Claims (13)

1. A method for implementing a video call, performed by a current subsystem, includes:
receiving a first target relative position sent by other subsystems, and determining a first target binocular video stream according to the first target relative position; wherein the first target relative position is determined by a first user position determining device in the other subsystem;
sending the first target binocular video stream to the other subsystems so that the other subsystems display the first target binocular video stream through first 3D display equipment;
sending the relative position of the second target to the other subsystems so that the other subsystems determine a second target binocular video stream according to the relative position of the second target; wherein the second target relative position is determined by a second user position determination device in the current subsystem;
and receiving second target binocular video streams sent by the other subsystems, and displaying the second target binocular video streams through second 3D display equipment.
2. The method of claim 1, wherein determining a first target binocular video stream based on the first target relative position comprises:
determining at least two target image collectors from the candidate image collectors of the current subsystem according to the relative position of the first target and the incidence relation between the candidate relative positions in the other subsystems and the candidate image collectors in the current subsystem; wherein the first target relative position comprises at least an offset direction of the other user relative to a screen of the other subsystem;
and determining the first target binocular video stream according to the image data acquired by at least two target image acquirers.
3. The method of claim 2, wherein before determining at least two target image collectors from the candidate image collectors of the current subsystem according to the first target relative position and the association relationship between the candidate relative positions in the other subsystems and the candidate image collectors of the current subsystem, the method further comprises:
and determining the incidence relation between the candidate relative positions in the other subsystems and the first candidate image collector in the current subsystem according to the viewing angles of the other users on the candidate relative positions, the offset directions of the other users relative to the screens of the other subsystems and the image collecting angle of the first candidate image collector of the current subsystem.
4. The method of claim 2, wherein before determining at least two target image collectors from the candidate image collectors of the current subsystem according to the first target relative position and the association relationship between the candidate relative positions in the other subsystems and the candidate image collectors of the current subsystem, the method further comprises:
and determining the incidence relation between the candidate relative positions in the other subsystems and the second candidate image collector in the current subsystem according to the viewing angles of the other users on the candidate relative positions, the offset directions of the other users relative to the screens of the other subsystems and the distance between the current user and the second candidate image collector of the current subsystem.
5. The method of claim 3, wherein determining the association relationship between the candidate relative positions in the other subsystems and the first candidate image collector in the current subsystem according to the viewing angles of the other users in the candidate relative positions, the offset directions of the other users relative to the screens of the other subsystems, and the image collecting angle of the first candidate image collector in the current subsystem comprises:
and if the viewing angles of other users at the candidate relative positions are consistent with the image acquisition angle of the first candidate image acquirer in the current subsystem, and the offset directions of the other users relative to the screen are consistent with the offset direction of the first candidate image acquirer in the current subsystem relative to the current user, establishing the association relationship between the candidate relative positions in the other subsystems and the first candidate image acquirer in the current subsystem.
6. The method of claim 4, wherein determining the association relationship between the candidate relative positions in the other subsystems and the second candidate image collector in the current subsystem according to the viewing angles of the other users in the candidate relative positions, the offset directions of the other users relative to the screens of the other subsystems, and the distance between the current user and the second candidate image collector in the current subsystem comprises:
determining a second candidate image collector in the target angle of the current subsystem according to the viewing angle of other users at the candidate relative position and the offset direction of other users relative to the screen;
determining the distance between a second candidate image collector in the target angle and the current user, and determining the image parallax determined by any two second candidate image collectors in the target angle according to the distance;
and if the image parallax is determined to be consistent with the preset image parallax, establishing the incidence relation between the candidate relative positions in other subsystems and the two second candidate image collectors in the current subsystem within the target angle.
7. The method of claim 1, wherein the first target relative position comprises:
the angle of the other users relative to the normal of the center of the screen of the other subsystem or the offset of the other users relative to the screen of the other subsystem.
8. The method of claim 2, wherein the disparity between the video content determined from the image data captured by the at least two target image collectors is less than or equal to a user binocular disparity.
9. An apparatus for implementing a video call, configured to a current subsystem, comprising:
the first target binocular video stream determining module is used for receiving the relative positions of the first targets sent by other subsystems and determining a first target binocular video stream according to the relative positions of the first targets; wherein the first target relative position is determined by a first user position determining device in the other subsystem;
the first target binocular video stream sending module is used for sending the first target binocular video stream to the other subsystems so that the other subsystems can display the first target binocular video stream through first 3D display equipment;
the second target relative position determining module is used for sending the second target relative position to the other subsystems so that the other subsystems can determine a second target binocular video stream according to the second target relative position; wherein the second target relative position is determined by a second user position determination device in the current subsystem;
and the display module is used for receiving the second target binocular video stream sent by the other subsystems and displaying the second target binocular video stream through second 3D display equipment.
10. An electronic device, comprising:
one or more processors;
a storage device for storing one or more programs,
when executed by the one or more processors, cause the one or more processors to implement a method for video call implementation as claimed in any of claims 1-8 performed by the current subsystem.
11. A system for implementing a video call, the system comprising at least two subsystems, comprising: the system comprises a current subsystem and other subsystems, wherein the other subsystems comprise first user position determining equipment, first 3D display equipment, a first server and a first image collector;
the current subsystem comprises a second user position determining device, a second 3D display device, a second server and a second image collector;
the first user position determining device determines a first target relative position of other users relative to screens of other subsystems and sends the first target relative position to the second server;
the second server determines a first target binocular video stream acquired by a second image acquisition device according to the relative position of the first target, and sends the first target binocular video stream to the first server;
the first 3D display device is used for carrying out 3D display on the first target binocular video stream in the first server;
the second server further sends a second target relative position determined by the second user position determining device to the first server;
the first server determines a second target binocular video stream acquired by the first image acquirer according to the relative position of the second target, and sends the second target binocular video stream to the second server;
and the second 3D display equipment is used for carrying out 3D display on the second target binocular video stream in the second server.
12. The system of claim 11, wherein the second user position determining device comprises at least one of a tracking sensor comprising at least one camera, a tracking sensor comprising an image and depth information sensor, and an eye tracking sensor based on gaze point tracking.
13. A computer-readable storage medium, on which a computer program is stored, which, when being executed by a processor, carries out a method for carrying out a video call, as claimed in any one of claims 1 to 8, carried out by a current subsystem.
CN202210088623.3A 2022-01-25 2022-01-25 Method, device, equipment, system and medium for realizing video call Pending CN114500911A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210088623.3A CN114500911A (en) 2022-01-25 2022-01-25 Method, device, equipment, system and medium for realizing video call

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210088623.3A CN114500911A (en) 2022-01-25 2022-01-25 Method, device, equipment, system and medium for realizing video call

Publications (1)

Publication Number Publication Date
CN114500911A true CN114500911A (en) 2022-05-13

Family

ID=81474720

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210088623.3A Pending CN114500911A (en) 2022-01-25 2022-01-25 Method, device, equipment, system and medium for realizing video call

Country Status (1)

Country Link
CN (1) CN114500911A (en)

Similar Documents

Publication Publication Date Title
KR102239686B1 (en) Single depth tracking acclimatization-convergence solution
US10523919B2 (en) Apparatus and method for adjusting stereoscopic image parallax and stereo camera
US20150358539A1 (en) Mobile Virtual Reality Camera, Method, And System
US10951804B2 (en) Photographing synchronization method and apparatus
US20130141550A1 (en) Method, apparatus and computer program for selecting a stereoscopic imaging viewpoint pair
CN111970524B (en) Control method, device, system, equipment and medium for interactive live broadcast and microphone connection
US20150244756A1 (en) Method, Apparatus and System for Determining Terminal That is to Share Real-Time Video
WO2022262839A1 (en) Stereoscopic display method and apparatus for live performance, medium, and system
CN105516639A (en) Headset device, three-dimensional video call system and three-dimensional video call implementing method
US20160373725A1 (en) Mobile device with 4 cameras to take 360°x360° stereoscopic images and videos
KR20120054746A (en) Method and apparatus for generating three dimensional image in portable communication system
WO2017130472A1 (en) Three-dimensionalizing video display device for two-dimensional video
CN114500911A (en) Method, device, equipment, system and medium for realizing video call
EP2852149A1 (en) Method and apparatus for generation, processing and delivery of 3D video
CN114040184A (en) Image display method, system, storage medium and computer program product
CN113641247A (en) Sight angle adjusting method and device, electronic equipment and storage medium
WO2017092369A1 (en) Head-mounted device, three-dimensional video call system and three-dimensional video call implementation method
CN110784728B (en) Image data processing method and device and computer readable storage medium
CN114422819A (en) Video display method, device, equipment, system and medium
KR100703713B1 (en) 3D mobile devices capable offer 3D image acquisition and display
KR101373611B1 (en) System and method for sharing 3d video with glasses type camera/display apparatuses
CN114339188B (en) Multi-view virtual reality immersive display method, device, medium and electronic equipment
CN114697602B (en) Conference device and conference system
CN218998126U (en) High-definition low-delay 3D wireless remote image transmission system
CN114173109A (en) Watching user tracking method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination