KR102044003B1 - Electronic apparatus for a video conference and operation method therefor - Google Patents
Electronic apparatus for a video conference and operation method therefor Download PDFInfo
- Publication number
- KR102044003B1 KR102044003B1 KR1020170155550A KR20170155550A KR102044003B1 KR 102044003 B1 KR102044003 B1 KR 102044003B1 KR 1020170155550 A KR1020170155550 A KR 1020170155550A KR 20170155550 A KR20170155550 A KR 20170155550A KR 102044003 B1 KR102044003 B1 KR 102044003B1
- Authority
- KR
- South Korea
- Prior art keywords
- image
- user
- information
- face
- camera
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/15—Conference systems
- H04N7/157—Conference systems defining a virtual conference space and using avatars or agents
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/332—Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/141—Systems for two-way working between two video terminals, e.g. videophone
- H04N7/142—Constructional details of the terminal equipment, e.g. arrangements of the camera and the display
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/141—Systems for two-way working between two video terminals, e.g. videophone
- H04N7/147—Communication arrangements, e.g. identifying the communication as a video-communication, intermediate storage of the signals
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Controls And Circuits For Display Device (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
A first electronic device and a method of operating the same are disclosed. The first electronic device may include a first camera that generates a first image of the first user; A receiver configured to receive, from the second camera, a second image photographing a specific area of the first user not photographed by the first camera through a second camera worn by the first user; Detecting a face image of the first user from the first image, generating a third image which is a 3D stereoscopic image corresponding to the face of the first user based on the first image, and generating the third image and the A processor configured to generate a synthesized image obtained by synthesizing the second image; And a transmitter for transmitting the synthesized image to another electronic device. Therefore, the first electronic device may provide a vivid expression of the first user to the second electronic device.
Description
The present invention relates to an image processing technology, and more particularly, to an electronic device for a video conference and a method of operating the same.
A video conference system can acquire video and audio signals for each user through each camera and microphone located remotely. The video conferencing system may transmit the acquired video and audio signals to each other user via a network. In addition, the video conferencing system may transmit each received video and audio signal to each user through each display and speaker.
The user may wear conference equipment, such as a headphone mic set or a head mounted display (HMD). The video conferencing system may provide an image of a user wearing the conference equipment to another user. At this time, the other user may have a problem that the immersion feeling for the video conference is reduced due to the conference equipment worn by the user in the video.
An object of the present invention for solving the above problems is to provide an electronic device and method for video conferencing for transmitting and displaying by replacing the image of the user wearing the conference equipment with the image of the user not wearing the conference equipment have.
According to an aspect of the present invention, there is provided a first electronic device including: a first camera generating a first image of a first user; A receiver configured to receive, from the second camera, a second image photographing a specific area of the first user not photographed by the first camera through a second camera worn by the first user; Detecting a face image of the first user from the first image, generating a third image which is a 3D stereoscopic image corresponding to the face of the first user based on the first image, and generating the third image and the A processor configured to generate a synthesized image obtained by synthesizing the second image; And a transmitter for transmitting the synthesized image to another electronic device.
The processor generates position information and size information of the face part image of the first user from the first image, and generates a second camera image which is an image of a portion where the second camera is worn on the face part image of the first user. And detect location information and size information of the second camera image.
The processor generates a face image model of the first user based on position information and size information of the face image of the first user, and generates a frame of the first image based on the face image image of the first user. The apparatus may determine whether a face image of the first user exists and generate face information of the first user based on the face image model of the first user.
The processor generates the second camera image model based on position information and size information of the second camera image, and the second camera image in a frame of the first image based on the second camera image model. It can be determined whether this exists.
The processor generates 3D image information on the face part of the first user based on the first image, and the first user based on direction information and the 3D image information of the face part image of the first user. The third image corresponding to the direction of the face portion may be generated.
The processor is configured to generate first image illuminance information on an average illuminance of the first image, generate second image illuminance information on an average illuminance of the second image, and generate a first image related to the average illuminance of the third image. 3 Image illuminance information can be generated.
The processor may change the second image illuminance information and the third image illuminance information to correspond to the first image illuminance information based on the first image illuminance information.
The processor may correct the distortion of the second image having the illumination information changed, and change the direction of the second image to correspond to the direction of the face part of the first user based on the direction information of the face part image of the first user. have.
The processor generates a modified image in which the face image of the first user is changed to the third image, and includes position information, size information and direction information of the face image of the first user, and position information of the second camera image. And generate the synthesized image obtained by synthesizing the changed image and the second image based on the direction information.
The processor may monitor a face image of the first user in the first image and update the composite image based on the monitoring result.
According to an embodiment of the present disclosure, a method of operating a first electronic device may include generating a first image of a first user by using a first camera; Receiving, from the second camera, a second image photographing a specific area of the first user not photographed by the first camera through a second camera worn by the first user; Detecting a face image of the first user from the first image; Generating a third image, which is a 3D stereoscopic image, corresponding to the face part of the first user based on the first image; Generating a synthesized image obtained by synthesizing the third image and the second image; And transmitting the composite image to another electronic device.
The detecting of the face unit image of the first user may include generating location information and size information of the face unit image of the first user from the first image; Detecting a second camera image, which is an image of a part where the second camera is worn, from a face image of the first user; And generating location information and size information of the second camera image.
The detecting of the face image of the first user may include generating a face image model of the first user based on location information and size information of the face image of the first user; Determining whether a face image of the first user exists in a frame of the first image based on the face image model of the first user; And generating facial part direction information of the first user based on the facial part image model of the first user.
The detecting of the face image of the first user may include generating the second camera image model based on location information and size information of the second camera image; And determining whether the second camera image exists in a frame of the first image based on the second camera image model.
The generating of the third image may include generating three-dimensional image information of the face part of the first user based on the first image; And generating the third image corresponding to the direction of the face part of the first user based on the direction information of the face part image of the first user and the 3D image information.
Generating a third image may include generating first image illumination information regarding an average illumination of the first image; Generating second image illuminance information regarding an average illuminance of the second image; And generating third image illumination information regarding an average illumination of the third image.
The generating of the third image may further include changing the second image illuminance information and the third image illuminance information to correspond to the first image illuminance information based on the first image illuminance information. can do.
The generating of the third image may include correcting a distortion of the second image having the illumination information changed; And changing the direction of the second image to correspond to the direction of the face part of the first user based on the direction information of the face part image of the first user.
The generating of the synthesized image may include: generating a changed image of changing a face image of the first user into the third image; And generating the composite image by combining the changed image and the second image based on the position information, the size information and the orientation information of the face image of the first user, and the position information and the orientation information of the second camera image. It can include;
The generating of the synthesized image may include: monitoring a face image of the first user in the first image; And updating the composite image based on the monitoring result.
According to the present invention, by providing a video conferencing electronic device and method for transmitting and outputting the image of the user wearing the conference equipment to the image of the user not wearing the conference equipment, participating in the video conference Can be provided to other users.
1A is a conceptual diagram of a video conferencing system providing a user image, according to an exemplary embodiment.
1B is a conceptual diagram of a video conference system providing a synthesized user image, according to an exemplary embodiment.
2 is a block diagram illustrating a configuration of a video conference system according to an exemplary embodiment.
3 is a block diagram illustrating a configuration of an electronic device according to an embodiment.
4 is a block diagram illustrating a configuration of a camera in an embodiment.
5 is a block diagram illustrating a configuration of a head mounted display according to an embodiment.
6 is a block diagram illustrating a configuration of an image processing apparatus according to an exemplary embodiment.
7 is a flowchart illustrating an operation sequence of an electronic device, according to an exemplary embodiment.
8A is a flowchart illustrating an operation sequence of an electronic device for detecting an object and estimating a direction of a face of a user, according to an exemplary embodiment.
8B is a flowchart illustrating an operation sequence of an electronic device for generating a composite image, according to an exemplary embodiment.
As the present invention allows for various changes and numerous embodiments, particular embodiments will be illustrated in the drawings and described in detail in the written description. However, this is not intended to limit the present invention to specific embodiments, it should be understood to include all modifications, equivalents, and substitutes included in the spirit and scope of the present invention.
Terms such as first and second may be used to describe various components, but the components should not be limited by the terms. The terms are used only for the purpose of distinguishing one component from another. For example, without departing from the scope of the present invention, the first component may be referred to as the second component, and similarly, the second component may also be referred to as the first component. The term and / or includes a combination of a plurality of related items or any item of a plurality of related items.
When a component is referred to as being "connected" or "connected" to another component, it may be directly connected to or connected to that other component, but it may be understood that other components may be present in between. Should be. On the other hand, when a component is said to be "directly connected" or "directly connected" to another component, it should be understood that there is no other component in between.
The terminology used herein is for the purpose of describing particular example embodiments only and is not intended to be limiting of the present invention. Singular expressions include plural expressions unless the context clearly indicates otherwise. In this application, the terms "comprise" or "have" are intended to indicate that there is a feature, number, step, operation, component, part, or combination thereof described in the specification, and one or more other features. It is to be understood that the present invention does not exclude the possibility of the presence or the addition of numbers, steps, operations, components, components, or a combination thereof.
Unless defined otherwise, all terms used herein, including technical or scientific terms, have the same meaning as commonly understood by one of ordinary skill in the art. Terms such as those defined in the commonly used dictionaries should be construed as having meanings consistent with the meanings in the context of the related art and shall not be construed in ideal or excessively formal meanings unless expressly defined in this application. Do not.
Hereinafter, with reference to the accompanying drawings, it will be described in detail a preferred embodiment of the present invention. In the following description of the present invention, the same reference numerals are used for the same elements in the drawings and redundant descriptions of the same elements will be omitted.
1A is a conceptual diagram of a video conferencing system providing a user image, according to an exemplary embodiment.
Referring to FIG. 1A, a video conferencing system may include a first user system 110 and a second user system 120. The first user system 110 and the second user system 120 may transmit and receive real-time video and audio data of each user through a network or point-to-point communication through a server (not shown).
The first user system 110 may photograph the
For example, the first user system 110 may acquire a real time image and audio of the
The first user system 110 may transmit a real time image and audio of the
In this case, each of the first user system 110 and the second user system 120 may be disposed in a separate video conference room (not shown) for improving immersion in video conferences and providing vivid video and audio. In addition, the first user system 110 and the second user system 120 may utilize augmented reality (AR) and virtual reality (VR) technologies to reduce equipment costs for the video conference room. Can be. For example, the first user system 110 and the second user system 120 replace each user image with a virtual avatar in a video conference room represented by a two-dimensional or three-dimensional virtual space, respectively. The user's motions and emotions can be expressed through the avatar.
The
The first head mounted
In addition, the first head mounted
The first head mounted
The second user system 120 based on the image signal received from the first user system 110, the
1B is a conceptual diagram of a video conference system providing a synthesized user image, according to an exemplary embodiment.
Referring to FIG. 1B, the first user system 110 displays a first head mounted display of an image of the
In a general video conferencing system or a virtual video conferencing system, the first user system 110 may include a separate user before joining the video conference when the
That is, the first user system 110 may synthesize a part of the
2 is a block diagram illustrating a configuration of a video conference system according to an exemplary embodiment.
The video conferencing system according to an embodiment may be a virtual video conferencing system that provides a video conference to a user using a head mounted display in a virtual video conference space using augmented reality and virtual reality.
Referring to FIG. 2, the video conferencing system may include a
The
Each of the first and second head mounted
The first user face photographing camera included in the first head mounted
3 is a block diagram illustrating a configuration of an electronic device according to an embodiment.
Referring to FIG. 3, the
The
The
The
The
The
The
The
The
The
In addition, although not shown, when the
4 is a block diagram illustrating a configuration of a camera in an embodiment.
Referring to FIG. 4,
The
The
The
The
The
5 is a block diagram illustrating a configuration of a head mounted display according to an embodiment.
Referring to FIG. 5, the head mounted
The
The
The interior of the head mounted
The
In addition, there may be at least one or more optical sensors fixed to the head mounted
The
The
The
The
The
Any operation or method for the head mounted
The
The head mounted
6 is a block diagram illustrating a configuration of an image processing apparatus according to an exemplary embodiment.
Referring to FIG. 6, the
The
The
The
The object
The
The
The
The 3D scanning
Also, the 3D scanning
The
7 is a flowchart illustrating an operation sequence of an electronic device, according to an exemplary embodiment.
Referring to FIG. 7, the electronic device may generate a captured image (S701).
The electronic device may operate in the same or similar manner as the first and second
In this case, the user may wear an object such as conference equipment or glasses or a hat. In addition, the user can wear a head mounted display. The electronic device can take a picture of the user wearing the head mounted display.
The electronic device may generate a second image of the face of the user. The user may wear a head mounted display. The head mounted display may comprise a camera. The head mounted display may photograph the face of the user through a camera. The head mounted display may transmit an image of the face of the user to the electronic device. The electronic device may generate a second image of the face part of the user captured by the camera of the head mounted display. In addition, the electronic device may generate a second image of the face of the user through a separate external camera.
The electronic device may photograph a user who has not worn an object. The electronic device may generate a third image by photographing a face of a user who has not worn an object such as glasses or a hat through a camera. The electronic device may three-dimensionally process the third image by the image processing device. The electronic device may generate a 3D image through the image processing device. The electronic device may generate an image or image information having a predetermined angular interval according to the x-axis, y-axis, and z-axis directions of the 3D image through the image processing apparatus. The electronic device may display the first to third images on the display of the electronic device. The electronic device may transmit the first to third images to another electronic device.
The electronic device may detect an object worn by the user and determine a direction of the face part of the user (S702).
The electronic device may detect an object worn by the user in the first image. For example, the electronic device may detect an object such as glasses or a hat or a head mounted display worn by the user of the first image. The electronic device may determine the position and size of the object. The electronic device may determine whether the object is the same as a predefined object. In addition, the electronic device may determine the direction of the face part of the user based on the third image.
The electronic device may generate a video call or a video for a meeting in operation S703.
The electronic device may adjust distortion and change illuminance according to characteristics of the lens with respect to the second image. The electronic device may determine an image or image that matches the direction of the face part of the current user based on the image or image data of a predetermined angular interval included in the third image. In addition, the electronic device may process the determined image or image.
The electronic device may change the object image worn by the user into another image. The electronic device may change the image of the portion where the user and the object overlap, into another image. For example, the electronic device may change the image of the portion where the user and the object worn by the user overlap with the image before the user wears the object, based on the second image and the third image. The electronic device may arrange the changed user image in the virtual video conference space based on the predefined layout information.
8A is a flowchart illustrating an operation sequence of an electronic device for detecting an object and estimating a direction of a face of a user, according to an exemplary embodiment.
The electronic device may operate in the same or similar manner as the first
Referring to FIG. 8A, the electronic device may receive a first image and a second image (S801).
The electronic device may receive the first image from the first camera. The first camera may be located in front of the user to photograph the user. The first camera may generate a first image of the user and the background of the user. The first camera may transmit the first image to the electronic device. The electronic device may include a first camera. Alternatively, the first camera may be a separate external device connectable with the electronic device.
The electronic device may receive the second image from the head mounted display. The user may wear a head mounted display on a portion of the user's head or face. The head mounted display may include a second camera. The second camera may photograph the face of the user. The head mounted display may generate a second image of the face of the user. The head mounted display may transmit the second image to the electronic device.
The electronic device may determine whether to detect the object (S802).
For example, the electronic device may determine whether to detect the head mounted display. If the head mounted display is not detected, the electronic device may proceed to detect the head mounted display in the first image. Alternatively, when the head mounted display is detected, the electronic device may skip the object detection step and proceed to track the object in the first image.
The electronic device may detect an object in a frame of the first image (S803).
For example, the electronic device may detect a head mounted display object worn by a user in a specific frame of the first image.
The electronic device may generate position and size information of the head mounted display. For example, the electronic device may generate location information including x-axis and y-axis coordinate information of the head mounted display. In addition, the electronic device may generate size information including size information on the x-axis and size information on the y-axis of the head mounted display.
The electronic device may perform a complicated calculation process when performing a full search for the user's face image in the first image or the second image. The electronic device may detect a user face area based on a user face detection algorithm to reduce a complicated calculation process. In addition, the electronic device may search for the coordinates of the head mounted display in the detected user face area. In this case, the electronic device may generate region-based convolutional neural network (CNN) training data based on the plurality of user facial images or image data included in the second image. The convolutional neural network may be a model of one of deep neural networks widely used in various applications such as object classification and object detection in an image. For example, the convolutional neural network may have a structure suitable for learning two-dimensional data. The convolutional neural network may generate trained learning data through a backpropagation algorithm.
The electronic device may track the object in the first image (S804).
When an object is detected in a specific frame of the first image, the electronic device may not detect the object in a frame after the specific frame. The electronic device may track the object based on the location information and the size information of the object detected in the specific frame. The electronic device may generate location information and size information about the tracked object.
The electronic device may generate an object model based on the position information and the size information of the first detected object, and sequentially track the object with respect to frames continuously input based on the object model. The electronic device may generate an error of tracking an object that matches the object model even if the object to be tracked does not exist in the input frame. The electronic device may generate location information and size information of the tracked object. The electronic device may perform an operation of checking whether the object exists in the input frame based on the tracked position information and the size information of the object.
The electronic device may track the direction of the object in the first image (S805).
The electronic device may track the direction of the object to determine whether the object exists in the input frame based on the tracked position information and the size information of the object. For example, if it is determined that the tracked object is the head mounted display, the electronic device may estimate the direction of the head mounted display, that is, the direction of the user's face. The electronic device may generate user face portion direction information based on a result of estimating the direction of the user face portion.
The direction of the user of the first image captured by the first camera may be changed according to a posture and an operation of the user. On the other hand, the direction of the face of the user captured by the second camera included in the head mounted display may be constant regardless of the posture and motion of the user because the head mounted display is fixed to the head or part of the face of the user. have.
8B is a flowchart illustrating an operation sequence of an electronic device for generating a composite image, according to an exemplary embodiment.
Referring to FIG. 8B, the electronic device may determine a third image corresponding to the direction of the face of the user in the user image database (S806).
The electronic device may determine, in the user image database, a third image having the same or similar direction as the face direction of the current user according to the x-axis, y-axis, and z-axis directions based on the user's face portion direction information. The electronic device may convert the direction of the second image to the direction of the first image based on the user facial part direction information.
The electronic device may generate illuminance information of the first to third images in operation S807.
The electronic device may determine the average illuminance of the first image. The electronic device may generate first image illuminance information based on the average illuminance of the first image. The electronic device may determine the average illuminance of the second image. The electronic device may generate second image illuminance information based on the average illuminance of the second image. In addition, the electronic device may determine an average illuminance of the third image. The electronic device may generate third image illuminance information based on the average illuminance of the third image.
The electronic device may change the illuminance of the second and third images to be the same as the illuminance of the first image (S808).
The electronic device may change the average illuminance of the second image to be the same as the average illuminance of the user face region of the first image. In addition, the electronic device may change the average illuminance of the third image to be the same as the average illuminance of the user face region of the first image.
The electronic device may correct the distortion of the second image (S809).
The second camera included in the head mounted display may use a wide angle lens for capturing a wide range with a relatively short focal length as compared to the first camera. The electronic device may perform image processing for removing distortion by the wide-angle lens on the second image having the changed illuminance. In addition, the electronic device may perform image processing to match the direction of the second image with the direction of the user's face part of the first image.
The electronic device may synthesize the first to third images (S810).
The electronic device may copy an image of the user face region from the third image. The electronic device may generate the first composite image by synthesizing the copied image of the user face region to the user face region of the first image.
The electronic device may copy a portion of the second image and synthesize the portion of the first composite image. For example, the electronic device may copy an image of the eye area of the user from the second image. The electronic device may synthesize the copied image of the eye region of the user from the first synthesized image to the eye region of the user.
The electronic device may include: a first camera generating a first image photographing a first user; A receiver configured to receive, from the second camera, a second image photographing a specific area of the first user not photographed by the first camera through a second camera worn by the first user; 3D processing the first image to generate a 3D image, detecting a region of the second camera from the first image, and synthesizing a region of the second camera from the first image with a portion of the 3D image. A processor configured to generate the synthesized image; And a transmitter for transmitting the synthesized image to another electronic device.
The electronic device may include a first camera configured to generate a first image signal photographing a first user; A second camera disposed on a part of the first user and generating a second image signal photographing another part of the first user; A third camera configured to generate a third image signal of the first user wearing the second camera; Based on the first image signal, a plurality of first user face image images corresponding to a plurality of predetermined angles and sizes of the face portion of the first user are generated, and the second camera region is detected from the third image signal. And detecting the second camera area in a specific frame included in the third image signal, estimating a front direction of the face part of the first user based on the first image signal and the second image signal, Determine a first user face image corresponding to a front direction of the face of the first user from among first user face images, and determine a copy image corresponding to the second camera region among the first user face images A processor configured to synthesize the radiated image with a second camera area of the third image signal to generate a synthesized image; And a transceiver configured to transmit the synthesized image to another electronic device.
According to an embodiment of the present disclosure, a first electronic device may include a first camera configured to generate a first image of a first user; A receiver configured to receive, from the second camera, a second image photographing a specific area of the first user not photographed by the first camera through a second camera worn by the first user; Detecting a face image of the first user from the first image, generating a third image which is a 3D stereoscopic image corresponding to the face of the first user based on the first image, and generating the third image and the A processor configured to generate a synthesized image obtained by synthesizing the second image; And a transmitter for transmitting the synthesized image to another electronic device.
In the first electronic device, the processor may generate location information and size information of the face part image of the first user from the first image, and the image of the part where the second camera is worn on the face part image of the first user. The second camera image may be detected, and location information and size information of the second camera image may be generated. The second camera image may be a virtual reality image.
The processor generates a face image model of the first user based on position information and size information of the face image of the first user, and generates a frame of the first image based on the face image image of the first user. The apparatus may determine whether a face image of the first user exists and generate face information of the first user based on the face image model of the first user.
The processor generates the second camera image model based on position information and size information of the second camera image, and the second camera image in a frame of the first image based on the second camera image model. It can be determined whether this exists.
The processor generates 3D image information on the face part of the first user based on the first image, and the first user based on direction information and the 3D image information of the face part image of the first user. The third image corresponding to the direction of the face portion may be generated.
The processor is configured to generate first image illuminance information on an average illuminance of the first image, generate second image illuminance information on an average illuminance of the second image, and generate a first image related to the average illuminance of the third image. 3 Image illuminance information can be generated.
The processor may change the second image illuminance information and the third image illuminance information to correspond to the first image illuminance information based on the first image illuminance information.
The processor may correct the distortion of the second image having the illumination information changed, and change the direction of the second image to correspond to the direction of the face part of the first user based on the direction information of the face part image of the first user. have.
The processor generates a modified image in which the face image of the first user is changed to the third image, and includes position information, size information and direction information of the face image of the first user, and position information of the second camera image. And generate the synthesized image obtained by synthesizing the changed image and the second image based on the direction information.
The processor may monitor a face image of the first user in the first image and update the composite image based on the monitoring result.
According to an embodiment of the present disclosure, a method of operating a first electronic device may include generating a first image of a first user by using a first camera; Receiving, from the second camera, a second image photographing a specific area of the first user not photographed by the first camera through a second camera worn by the first user; Detecting a face image of the first user from the first image; Generating a third image, which is a 3D stereoscopic image, corresponding to the face part of the first user based on the first image; Generating a synthesized image obtained by synthesizing the third image and the second image; And transmitting the composite image to another electronic device.
The detecting of the face unit image of the first user may include generating location information and size information of the face unit image of the first user from the first image; Detecting a second camera image, which is an image of a part where the second camera is worn, from a face image of the first user; And generating location information and size information of the second camera image. The second camera image may be a virtual reality image.
The detecting of the face image of the first user may include generating a face image model of the first user based on location information and size information of the face image of the first user; Determining whether a face image of the first user exists in a frame of the first image based on the face image model of the first user; And generating facial part direction information of the first user based on the facial part image model of the first user.
The detecting of the face image of the first user may include generating the second camera image model based on location information and size information of the second camera image; And determining whether the second camera image exists in a frame of the first image based on the second camera image model.
The generating of the third image may include generating three-dimensional image information of the face part of the first user based on the first image; And generating the third image corresponding to the direction of the face part of the first user based on the direction information of the face part image of the first user and the 3D image information.
Generating a third image may include generating first image illumination information regarding an average illumination of the first image; Generating second image illuminance information regarding an average illuminance of the second image; And generating third image illumination information regarding an average illumination of the third image.
The generating of the third image may further include changing the second image illuminance information and the third image illuminance information to correspond to the first image illuminance information based on the first image illuminance information. can do.
The generating of the third image may include correcting a distortion of the second image having the illumination information changed; And changing the direction of the second image to correspond to the direction of the face part of the first user based on the direction information of the face part image of the first user.
The generating of the synthesized image may include: generating a changed image of changing a face image of the first user into the third image; And generating the composite image by combining the changed image and the second image based on the position information, the size information and the orientation information of the face image of the first user, and the position information and the orientation information of the second camera image. It can include;
The generating of the synthesized image may include: monitoring a face image of the first user in the first image; And updating the composite image based on the monitoring result.
The methods according to the invention can be implemented in the form of program instructions that can be executed by various computer means and recorded on a computer readable medium. Computer-readable media may include, alone or in combination with the program instructions, data files, data structures, and the like. The program instructions recorded on the computer readable medium may be those specially designed and constructed for the present invention, or may be known and available to those skilled in computer software.
Examples of computer readable media include hardware devices that are specifically configured to store and execute program instructions, such as ROM, RAM, flash memory, and the like. Examples of program instructions include machine language code, such as produced by a compiler, as well as high-level language code that can be executed by a computer using an interpreter or the like. The hardware device described above may be configured to operate with at least one software module to perform the operations of the present invention, and vice versa.
Although described with reference to the embodiments above, those skilled in the art will understand that the present invention can be variously modified and changed without departing from the spirit and scope of the invention as set forth in the claims below. Could be.
Claims (20)
A first camera generating a first image of the first user;
A receiver configured to receive, from the second camera, a second image photographing a specific area of the first user not photographed by the first camera through a second camera worn by the first user;
Detecting a face image of the first user from the first image, generating a third image which is a 3D stereoscopic image corresponding to the face of the first user based on the first image, and generating the third image and the A processor configured to generate a synthesized image obtained by synthesizing the second image; And
And a transmitter configured to transmit the composite image to another electronic device.
The processor generates position information and size information of the face part image of the first user from the first image, and generates a second camera image which is an image of a portion where the second camera is worn on the face part image of the first user. And detecting location information and generating location information and size information of the second camera image.
The processor generates a face image model of the first user based on position information and size information of the face image of the first user, and generates a frame of the first image based on the face image image of the first user. And determining whether a face image of the first user exists and generating face direction information of the first user based on the face image model of the first user.
The processor generates the second camera image model based on location information and size information of the second camera image.
And determining whether the second camera image exists in a frame of the first image based on the second camera image model.
The processor generates 3D image information on the face part of the first user based on the first image, and the first user based on direction information and the 3D image information of the face part image of the first user. The first electronic device generates the third image corresponding to the direction of the face part of the face.
The processor is configured to generate first image illuminance information on an average illuminance of the first image, generate second image illuminance information on an average illuminance of the second image, and generate a second image illuminant on the average illuminance of the third image. The first electronic device to generate 3 image illuminance information.
The processor is further configured to change the second image illuminance information and the third image illuminance information to correspond to the first image illuminance information based on the first image illuminance information.
The processor may be configured to correct distortion of the second image in which illuminance information is changed, and to change the direction of the second image to correspond to the direction of the face part of the first user based on the direction information of the face part image of the first user. , First electronic device.
The processor generates a modified image in which the face image of the first user is changed to the third image, and includes position information, size information and direction information of the face image of the first user, and position information of the second camera image. And generating the composite image by combining the changed image and the second image based on direction information.
The processor is configured to monitor a face image of the first user in the first image and to update the composite image based on the monitoring result.
Generating a first image of the first user by using the first camera;
Receiving, from the second camera, a second image photographing a specific area of the first user not photographed by the first camera through a second camera worn by the first user;
Detecting a face image of the first user from the first image;
Generating a third image, which is a 3D stereoscopic image, corresponding to the face part of the first user based on the first image;
Generating a synthesized image obtained by synthesizing the third image and the second image; And
Transmitting the composite image to another electronic device;
The detecting of the face part image of the first user may include:
Generating location information and size information of a face image of the first user from the first image;
Detecting a second camera image, which is an image of a part where the second camera is worn, from a face image of the first user; And
Generating location information and size information of the second camera image.
The detecting of the face part image of the first user may include:
Generating a face image model of the first user based on location information and size information of the face image of the first user;
Determining whether a face image of the first user exists in a frame of the first image based on the face image model of the first user; And
And generating facial part direction information of the first user based on the facial part image model of the first user.
The detecting of the face part image of the first user may include:
Generating the second camera image model based on location information and size information of the second camera image; And
And determining whether the second camera image exists in a frame of the first image based on the second camera image model.
Generating the third image,
Generating 3D image information on the face part of the first user based on the first image; And
And generating the third image corresponding to the direction of the face part of the first user based on the direction information of the face part image of the first user and the three-dimensional image information.
Generating the third image,
Generating first image illumination information regarding an average illumination of the first image;
Generating second image illuminance information regarding an average illuminance of the second image; And
Generating third image illuminance information regarding the average illuminance of the third image.
Generating the third image,
And changing the second image illuminance information and the third image illuminance information to correspond to the first image illuminance information based on the first image illuminance information. .
Generating the third image,
Correcting distortion of the second image in which illuminance information is changed; And
And changing the direction of the second image to correspond to the direction of the face part of the first user based on the direction information of the face part image of the first user.
Generating the composite image,
Generating a changed image in which the face image of the first user is changed to the third image;
Generating the synthesized image by combining the changed image and the second image based on the position information, the size information and the orientation information of the face image of the first user, and the position information and the orientation information of the second camera image; And operating the first electronic device.
Generating the composite image,
Monitoring a face image of the first user in the first image; And
And updating the synthesized image based on the monitoring result.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020160156459 | 2016-11-23 | ||
KR20160156459 | 2016-11-23 |
Publications (2)
Publication Number | Publication Date |
---|---|
KR20180058199A KR20180058199A (en) | 2018-05-31 |
KR102044003B1 true KR102044003B1 (en) | 2019-11-12 |
Family
ID=62454562
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020170155550A KR102044003B1 (en) | 2016-11-23 | 2017-11-21 | Electronic apparatus for a video conference and operation method therefor |
Country Status (1)
Country | Link |
---|---|
KR (1) | KR102044003B1 (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110084775B (en) * | 2019-05-09 | 2021-11-26 | 深圳市商汤科技有限公司 | Image processing method and device, electronic equipment and storage medium |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2820842A1 (en) * | 2012-02-27 | 2015-01-07 | ETH Zürich | Method and system for image processing in video conferencing for gaze correction |
US9524588B2 (en) * | 2014-01-24 | 2016-12-20 | Avaya Inc. | Enhanced communication between remote participants using augmented and virtual reality |
-
2017
- 2017-11-21 KR KR1020170155550A patent/KR102044003B1/en active IP Right Grant
Also Published As
Publication number | Publication date |
---|---|
KR20180058199A (en) | 2018-05-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10445917B2 (en) | Method for communication via virtual space, non-transitory computer readable medium for storing instructions for executing the method on a computer, and information processing system for executing the method | |
CN110506249B (en) | Information processing apparatus, information processing method, and recording medium | |
US20200410713A1 (en) | Generating pose information for a person in a physical environment | |
US20180315364A1 (en) | Information Processing Apparatus and Image Generation Method | |
US10546407B2 (en) | Information processing method and system for executing the information processing method | |
EP3422149B1 (en) | Methods, apparatus, systems, computer programs for enabling consumption of virtual content for mediated reality | |
US10410395B2 (en) | Method for communicating via virtual space and system for executing the method | |
CN102959616A (en) | Interactive reality augmentation for natural interaction | |
JPWO2017122299A1 (en) | Facial expression recognition system, facial expression recognition method and facial expression recognition program | |
JP2018532173A (en) | Shared reality content sharing | |
JP7081052B2 (en) | Displaying device sharing and interactivity in simulated reality (SR) | |
US20180299948A1 (en) | Method for communicating via virtual space and system for executing the method | |
US20150138301A1 (en) | Apparatus and method for generating telepresence | |
US10564801B2 (en) | Method for communicating via virtual space and information processing apparatus for executing the method | |
KR20200038111A (en) | electronic device and method for recognizing gestures | |
WO2017061890A1 (en) | Wireless full body motion control sensor | |
US20190227695A1 (en) | Immersive displays | |
KR102044003B1 (en) | Electronic apparatus for a video conference and operation method therefor | |
JP6518645B2 (en) | INFORMATION PROCESSING APPARATUS AND IMAGE GENERATION METHOD | |
CN110968248B (en) | Generating a 3D model of a fingertip for visual touch detection | |
JP2021180425A (en) | Remote control system, remote work device thereof, video processing device and program | |
US20200342833A1 (en) | Head mounted display system and scene scanning method thereof | |
JP2021009647A (en) | Virtual reality control apparatus, virtual reality head set, virtual reality control method, and program | |
US11882172B2 (en) | Non-transitory computer-readable medium, information processing method and information processing apparatus | |
KR102657318B1 (en) | Personalized apparatus for virtual reality based on remote experience and method for providing virtual reality experience |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
A201 | Request for examination | ||
E902 | Notification of reason for refusal | ||
E701 | Decision to grant or registration of patent right | ||
GRNT | Written decision to grant |