WO2017122299A1 - 表情認識システム、表情認識方法及び表情認識プログラム - Google Patents
表情認識システム、表情認識方法及び表情認識プログラム Download PDFInfo
- Publication number
- WO2017122299A1 WO2017122299A1 PCT/JP2016/050869 JP2016050869W WO2017122299A1 WO 2017122299 A1 WO2017122299 A1 WO 2017122299A1 JP 2016050869 W JP2016050869 W JP 2016050869W WO 2017122299 A1 WO2017122299 A1 WO 2017122299A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- user
- camera
- expression recognition
- unit
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/70—Multimodal biometrics, e.g. combining information from different biometric modalities
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/012—Head tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/10—Image acquisition
- G06V10/12—Details of acquisition arrangements; Constructional details thereof
- G06V10/14—Optical characteristics of the device performing the acquisition or on the illumination arrangements
- G06V10/143—Sensing or illuminating at different wavelengths
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
- G06V40/175—Static expression
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/19—Sensors therefor
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/25—Image signal generators using stereoscopic image cameras using two or more image sensors with different characteristics other than in their location or field of view, e.g. having different resolutions or colour pickup characteristics; using image signals from one sensor to control the characteristics of another sensor
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/254—Image signal generators using stereoscopic image cameras in combination with electromagnetic radiation sources for illuminating objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/366—Image reproducers using viewer tracking
- H04N13/383—Image reproducers using viewer tracking for tracking with gaze detection, i.e. detecting the lines of sight of the viewer's eyes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/10—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths
- H04N23/11—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths for generating image signals from visible and infrared light wavelengths
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
- G06T2207/10012—Stereo images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10048—Infrared image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10141—Special mode during image acquisition
- G06T2207/10152—Varying illumination
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
- G06T2207/30201—Face
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/171—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
Definitions
- the present invention relates to a head mounted display.
- Some head mounted displays have a function of presenting a three-dimensional image to a user who has worn it.
- the head mounted display is generally mounted and used so as to cover the field of view of the user. Then, as described above, in the content that uses the user's gaze direction as a pointing device, it is desirable to provide content that is more interesting to the user.
- the present invention has been made in view of the above-described needs, and an object thereof is to provide a head mounted display that can output information for providing content that the user may be more interested in.
- one mode of the present invention is the 1st camera which picturizes a user's eye, the 2nd camera which picturizes a user's mouth, and the 1st picture and 2nd which were picturized by the 1st camera
- a head mounted display having an output unit for outputting a second image captured by a camera, a receiving unit for receiving the first image and the second image output by the output unit, and based on the first image and the second image
- a facial expression recognition device having a facial expression recognition unit that recognizes a user's facial expression.
- the head mounted display further includes a light source for irradiating the invisible light to the user's eye, and a third camera for capturing the invisible light reflected by the user's eye, and the output unit is the third camera.
- the facial expression recognition apparatus may further include a gaze detection unit that detects the gaze direction of the user based on the third image received by the reception unit.
- the facial expression recognition apparatus further includes a combining unit that combines the first image and the second image received by the receiving unit to create a composite image, and the facial expression recognition unit recognizes the user's facial expression based on the composite image. You may do it.
- the second camera may be detachably attached to the head mounted display.
- the second camera may be attached to the head mounted display so as to have an angle of view capable of imaging the range from the user's nose to the shoulder when the user mounts the head mounted display.
- the facial expression recognition apparatus may further include a posture estimation unit that estimates the posture of the user based on the second image received by the reception unit.
- the head mounted display may be configured to cover the periphery of the user's eye and not to cover the user's mouth.
- the first camera and the second camera are cameras that acquire depth information indicating the distance to the imaging target, and the facial expression recognition system further includes the eye of the user captured by the first camera and the user captured by the second camera
- An avatar image that identifies the three-dimensional shape of the user's eyes and mouth based on the image with the mouth and generates an avatar image reflected on the user's avatar's eye and mouth shape based on the identified three-dimensional shape
- a generation unit may be provided.
- the user's face and the mouth are separately captured and synthesized to synthesize the face image of the user.
- An expected composite image can be acquired to perform facial expression recognition processing. Therefore, content in which the user's expression is reflected can be provided.
- FIG. 1 is a perspective view schematically showing an overview of an image display system of a head mounted display according to an embodiment. It is a figure which shows typically the optical structure of the image display system of the head mounted display which concerns on embodiment. It is a block diagram showing composition of a head mount display system concerning an embodiment. It is a schematic diagram explaining the calibration for the detection of the gaze direction which concerns on embodiment. It is a schematic diagram explaining the position coordinate of a user's cornea. (A) is an example of a picture around a user's eyes which a head mount display system concerning an embodiment picturizes.
- (B) is an example of a picture of a user's mouth circumference which a head mount display system concerning an embodiment picturizes. It is an example of the synthetic
- (A), (b) is an external view which shows the structure at the time of making a camera detachable to a head mounted display. It is an external view which shows the example which changes the camera angle of the camera 160 provided in the head mounted display.
- (A) is an image obtained by imaging a user, and (b) is an image obtained by reflecting the captured user's expression on an avatar image.
- head mounted displays are usually shaped so as to cover the periphery of the eyes of the user's head.
- the reason why the head mount display has such a shape is that the full helmet type not only gives the user a sense of oppression, but also increases the weight as the head mount display and gives a load to the user. It is because there is.
- a camera can be provided in the head mounted display to capture an image around the user's eye, it is impossible to acquire a face image of the entire user.
- Non-Patent Document 1 There is a technique described in Non-Patent Document 1 as a method for realizing facial expression recognition in a head mounted display of such a shape.
- a curved arm is attached to the outside of the head mount display, and the camera is placed on the opposite side to the attached side, thereby capturing an image of the user's mouth and realizing facial expression recognition.
- the center of gravity of the head mounted display as a whole is biased to the front of the user due to the attached curved arm, which makes handling difficult.
- the inventors have found that there is a problem of increased weight.
- Non-Patent Document 1 the inventors detect the movement of expression muscles around the eyes of the user with a myoelectric sensor (Strain Sensor), and realize expression recognition around the eyes.
- a myoelectric sensor Stress Sensor
- the method using the myoelectric sensor is not suitable for detecting the user's gaze.
- the inventors of the present invention have invented a configuration capable of performing gaze detection while performing facial expression recognition in a head mounted display of a type that covers the current user's field of view.
- the head mounted display according to the present invention will be described in detail.
- a facial expression recognition system 1 includes a first camera (181) for imaging a user's eye, a second camera (180) for imaging a user's mouth, and a first camera imaged by the first camera.
- a head mounted display (100) having an output unit (118) for outputting an image and a second image captured by a second camera, and a receiving unit (220 for receiving the first image and the second image output by the output unit Recognition of the user's expression based on the combined image created by the combining unit (222) that creates the combined image by combining the first image and the second image received by the receiving unit, and It is an expression recognition system provided with the expression recognition device (200) which has an expression recognition part (223).
- the head mounted display further includes a light source (103) for irradiating the user's eyes with the invisible light, and a third camera (161) for capturing the invisible light reflected by the user's eyes, and an output unit
- the third aspect of the present invention outputs the third image captured by the third camera
- the facial expression recognition apparatus further includes a gaze detection unit (221) that detects the gaze direction of the user based on the third image received by the reception unit. .
- FIG. 1 is a view schematically showing an overview of the facial expression recognition system 1 according to the embodiment.
- the facial expression recognition system 1 according to the embodiment includes a head mounted display 100 and a gaze detection apparatus 200. As shown in FIG. 1, the head mounted display 100 is used by being attached to the head of the user 300.
- the gaze detection apparatus 200 detects the gaze direction of at least one of the right eye and the left eye of the user wearing the head mounted display 100, and gazes at the user's focus, ie, the three-dimensional image displayed on the head mounted display by the user. Identify where it is
- the visual axis detection device 200 also functions as an image generation device that generates an image displayed by the head mounted display 100.
- the sight line detection device 200 is a device capable of reproducing video such as a stationary game machine, a portable game machine, a PC, a tablet, a smartphone, a Fablet, a video player, and a television.
- the gaze detection apparatus 200 is connected to the head mounted display 100 wirelessly or by wire. In the example shown in FIG.
- the visual axis detection device 200 is wirelessly connected to the head mounted display 100.
- the wireless connection between the sight line detection device 200 and the head mounted display 100 can be realized, for example, using a known wireless communication technology such as Wi-Fi (registered trademark) or Bluetooth (registered trademark).
- Wi-Fi registered trademark
- Bluetooth registered trademark
- transmission of an image between the head mounted display 100 and the gaze detection apparatus 200 is performed according to standards such as Miracast (trademark), WiGig (trademark), WHDI (trademark) and the like.
- FIG. 1 shows an example in which the head mounted display 100 and the sight line detection device 200 are different devices.
- the visual axis detection device 200 may be incorporated in the head mounted display 100.
- the head mounted display 100 includes a housing 150, a mounting tool 160, headphones 170, and a camera 180.
- the housing 150 accommodates an image display system for presenting an image to the user 300 such as an image display element, and a wireless transmission module such as a Wi-Fi module or a Bluetooth (registered trademark) module not shown.
- the mounting tool 160 mounts the head mounted display 100 on the head of the user 300.
- the mounting tool 160 can be realized by, for example, a belt or a stretchable band.
- the headphones 170 output the audio of the video reproduced by the visual axis detection device 200.
- the headphones 170 may not be fixed to the head mounted display 100.
- the user 300 can freely attach and detach the headphones 170 even when the head mounted display 100 is attached using the attachment 160.
- the camera 180 when the user 300 wears the head mounted display 100, the camera 180 is arranged to be able to capture an image including a half face of the user. That is, the imaging angle of view is arranged so as to be capable of imaging the lower half of the face of the user 300 (from the lower part of the nose of the user to the shoulder of the user). That is, the first image 801 as shown in FIG. 7B is captured. In the present specification, the image (an image including the lower half of the user's face) is referred to as a first image.
- the camera 180 is connected to a first communication unit 118 described later. The first image captured by the camera 180 is output by the first communication unit 118 to the gaze detection apparatus 200.
- a visible light camera or a depth camera is used as the camera 180.
- a depth camera is a camera that can obtain depth information from the camera to an object or a camera that can acquire a three-dimensional shape of an object.
- Specific examples of the depth camera include a stereo camera, a light field camera, and Structure Light. The camera used, the camera using the photometric stereo method, etc. are mentioned.
- FIG. 2 is a perspective view schematically showing an overview of the image display system 130 of the head mounted display 100 according to the embodiment. More specifically, FIG. 2 is a view showing a region of the housing 150 according to the embodiment, which faces the cornea 302 of the user 300 when the head mounted display 100 is mounted.
- the left-eye convex lens 114 a is disposed so as to face the cornea 302 a of the left eye of the user 300 when the user 300 wears the head mounted display 100.
- the right-eye convex lens 114 b is disposed so as to face the cornea 302 b of the right eye of the user 300 when the user 300 wears the head mounted display 100.
- the left-eye convex lens 114a and the right-eye convex lens 114b are held by the left-eye lens holding portion 152a and the right-eye lens holding portion 152b, respectively.
- the convex lens 114 a for the left eye and the convex lens 114 b for the right eye are simply referred to as “convex lens 114” unless otherwise specified.
- the cornea 302 a of the user 300 and the cornea 302 b of the right eye of the user 300 are simply referred to as “cornea 302” except in the case where they are particularly distinguished.
- the left-eye lens holding portion 152a and the right-eye lens holding portion 152b are also referred to as "lens holding portions 152" unless otherwise specified.
- the lens holding unit 152 is provided with a plurality of infrared light sources 103.
- an infrared light source for emitting infrared light to the cornea 302a of the left eye of the user 300 is collectively shown by the infrared light source 103a, and the cornea 302b of the right eye of the user 300 is shown.
- An infrared light source for emitting infrared light is collectively shown by an infrared light source 103b.
- the infrared light source 103a and the infrared light source 103b are described as "infrared light source 103" except in the case where they are particularly distinguished. In the example shown in FIG.
- the left eye lens holding unit 152a is provided with six infrared light sources 103a.
- six infrared light sources 103b are also provided in the right-eye lens holding unit 152b.
- the infrared light source 103 is arranged in the lens holding portion 152 that holds the convex lens 114 instead of directly arranging the infrared light source 103, the attachment of the infrared light source 103 is facilitated.
- the lens holding portion 152 is made of resin or the like, processing for attaching the infrared light source 103 is easier than the convex lens 114 made of glass or the like.
- the lens holding portion 152 is a member that holds the convex lens 114. Therefore, the infrared light source 103 provided in the lens holding unit 152 is disposed around the convex lens 114.
- the number is not limited to this, and at least one corresponding to each eye. It is preferable that there be two or more.
- FIG. 3 is a view schematically showing an optical configuration of the image display system 130 accommodated in the case 150 according to the embodiment, and a view of the case 150 shown in FIG. 2 when viewed from the side surface on the left eye side. is there.
- the image display system 130 includes an infrared light source 103, an image display element 108, a hot mirror 112, a convex lens 114, a camera 116, a first communication unit 118, and a camera 181.
- the infrared light source 103 is a light source capable of emitting light in a wavelength band of near infrared (about 700 nm to 2500 nm). Near infrared light is generally light in a wavelength band of invisible light that can not be observed by the naked eye of the user 300.
- the image display element 108 displays an image to be presented to the user 300.
- the image output unit 224 in the visual axis detection device 200 generates an image displayed by the image display element 108.
- the video output unit 224 will be described later.
- the image display element 108 can be realized using, for example, a known LCD (Liquid Crystal Display) or an organic EL display (Organic Electro Luminescence Display).
- the hot mirror 112 is disposed between the image display element 108 and the cornea 302 of the user 300 when the user 300 wears the head mounted display 100.
- the hot mirror 112 has a property of transmitting visible light generated by the image display element 108 but reflecting near infrared light.
- the convex lens 114 is disposed on the opposite side of the image display element 108 with respect to the hot mirror 112. In other words, the convex lens 114 is disposed between the hot mirror 112 and the cornea 302 of the user 300 when the user 300 wears the head mounted display 100. That is, the convex lens 114 is disposed at a position facing the cornea 302 of the user 300 when the head mounted display 100 is attached to the user 300.
- the convex lens 114 condenses the image display light transmitted through the hot mirror 112. Therefore, the convex lens 114 functions as an image enlargement unit that enlarges the image generated by the image display element 108 and presents the image to the user 300.
- the convex lens 114 may be a lens group configured by combining various lenses, one having a curvature and the other having a plane. It may be a single convex lens.
- the plurality of infrared light sources 103 are disposed around the convex lens 114.
- the infrared light source 103 emits infrared light toward the cornea 302 of the user 300.
- the image display system 130 of the head mounted display 100 includes two image display elements 108, and an image to be presented to the right eye of the user 300 and an image to be presented to the left eye. Can be generated independently. Therefore, the head mounted display 100 according to the embodiment can present the parallax image for the right eye and the parallax image for the left eye to the right eye and the left eye of the user 300, respectively. Thus, the head mounted display 100 according to the embodiment can present a stereoscopic image with a sense of depth to the user 300.
- the hot mirror 112 transmits visible light and reflects near infrared light. Therefore, the image light emitted by the image display element 108 passes through the hot mirror 112 and reaches the cornea 302 of the user 300. The infrared light emitted from the infrared light source 103 and reflected by the reflection area inside the convex lens 114 reaches the cornea 302 of the user 300.
- the infrared light that has reached the cornea 302 of the user 300 is reflected by the cornea 302 of the user 300 and travels toward the convex lens 114 again.
- the infrared light passes through the convex lens 114 and is reflected by the hot mirror 112.
- the camera 116 includes a filter that blocks visible light, and images near-infrared light reflected by the hot mirror 112. That is, the camera 116 is a near-infrared camera that captures near-infrared light that is emitted from the infrared light source 103 and is corneal-reflected by the eye of the user 300.
- the image display system 130 of the head mounted display 100 includes two cameras 116, that is, a first imaging unit that picks up an image including infrared light reflected by the right eye. And a second imaging unit configured to capture an image including infrared light reflected by the left eye. Thereby, an image for detecting the sight line direction of both the right eye and the left eye of the user 300 can be acquired. Note that if information about the focus coordinates in the depth direction is not required for the line of sight of the user, either of the lines of sight detection suffices.
- the first communication unit 118 outputs the image captured by the camera 116 to the gaze detection apparatus 200 that detects the gaze direction of the user 300. Specifically, the first communication unit 118 transmits the image captured by the camera 116 to the gaze detection apparatus 200.
- the details of the gaze detection unit 221 functioning as the gaze direction detection unit will be described later, but are realized by a gaze detection program executed by a CPU (Central Processing Unit) of the gaze detection apparatus 200.
- the CPU of the head mounted display 100 may execute a program for realizing the line-of-sight direction detection unit.
- the image captured by the camera 116 includes a bright spot caused by the near infrared light reflected by the cornea 302 of the user 300 and the cornea 302 of the user 300 observed in the near infrared wavelength band. And an image of an eye including the
- the configuration for mainly presenting the image to the left eye of the user 300 in the image display system 130 according to the embodiment has been described above, the configuration for presenting the image to the right eye of the user 300 is the same as above. is there.
- the optical configuration for realizing gaze detection in a head mounted display has been described, but in the head mounted display according to the present embodiment, an optical configuration for realizing facial expression recognition for recognizing a user's facial expression Also have.
- the head mounted display 100 includes a camera 181 for imaging around the eyes of the user.
- the camera 181 is a camera that images around the user's eye, and uses a visible light camera or a depth camera.
- a depth camera is used as the camera 181
- the three-dimensional shape of the lower half of the user's face can be specified.
- the camera 181 is a position in the head mount display, at which the user's eye facing the convex lens 114 is imaged through the convex lens 114, and the field of view of the user gazing at the image display element 108 Placed in a position that does not interfere with
- the image is arranged at the top of the image display system 130.
- the camera 181 is not a position at the top where it can capture an image around the user's eye, not a position where it interferes with the user's view. It may be arranged at the bottom or left and right.
- the camera 181 is connected to the first communication unit 118, and the camera 181 transmits the captured image to the first communication unit 118. Then, the first communication unit 118 outputs an image captured by the camera 181 to the gaze detection apparatus 200.
- a second image an image around the eyes of the user captured by the camera 181 will be referred to as a second image.
- FIG. 4 is a block diagram of the head mounted display 100 and the visual axis detection device 200 according to the facial expression recognition system 1.
- the facial expression recognition system 1 includes the head mounted display 100 and the gaze detection apparatus 200 that perform communication with each other.
- the head mounted display 100 includes a first communication unit 118, a display unit 121, an infrared light irradiation unit 122, an image processing unit 123, and an imaging unit 124.
- the first communication unit 118 is a communication interface having a function of performing communication with the second communication unit 220 of the visual axis detection device 200. As described above, the first communication unit 118 communicates with the second communication unit 220 by wired communication or wireless communication. Examples of usable communication standards are as described above.
- the first communication unit 118 transmits, to the second communication unit 220, the image data used for gaze detection transmitted from the camera 116 or the image processing unit 123. In addition, the first communication unit 118 transmits the three-dimensional image data transmitted from the sight line detection device 200 to the display unit 121.
- the first communication unit 118 adds an ID to the first image and the second image so that they can be distinguished from each other, and transmits the image to the facial expression recognition apparatus 200.
- the display unit 121 has a function of displaying the three-dimensional image data transmitted from the first communication unit 118 on the image display element 108.
- the three-dimensional image data includes a right-eye parallax image and a left-eye parallax image, which are a parallax image pair.
- the infrared light irradiation unit 122 controls the infrared light source 103 to emit infrared light to the user's right or left eye.
- the image processing unit 123 performs image processing on the image captured by the camera 116 as necessary, and transmits the image to the first communication unit 118.
- the imaging unit 124 captures an image including near-infrared light reflected by each eye using the camera 116 for the right eye and the camera 117 for the left eye.
- the imaging unit 124 transmits an image obtained by imaging to the first communication unit 118 or the image processing unit 123. Further, the imaging unit 124 transmits an image captured using the camera 180 and an image captured using the camera 181 to the first communication unit 118 or the image processing unit 123.
- the line-of-sight detection apparatus 200 includes a second communication unit 220, a line-of-sight detection unit 221, a combining unit 222, an expression recognition unit 223, an image output unit 224, and a storage unit 225.
- the second communication unit 220 is a communication interface having a function of performing communication with the first communication unit 118 of the head mounted display 100. As described above, the second communication unit 220 communicates with the first communication unit 118 by wired communication or wireless communication. When the second communication unit 220 receives data relating to the left-eye image or the right-eye image for gaze detection, the second communication unit 220 transmits the data to the gaze detection unit 221. In addition, the second communication unit 220 receives data related to the face image of the user (an image around the user's eyes or an image of the lower half of the user's face), that is, data related to the first image or the second image If it does, the data is transmitted to the synthesis unit 222.
- the gaze detection unit 221 receives image data for detecting the gaze of the right eye of the user from the second communication unit 220, and detects the gaze direction of the user's right eye.
- the gaze detection unit 221 calculates a right viewing line vector indicating the gaze direction of the user's right eye using a method described later.
- the gaze detection unit 221 receives image data for gaze detection of the left eye of the user from the second communication unit 220, and detects the gaze direction of the left eye of the user.
- the line-of-sight detection unit 221 calculates a left viewing line vector indicating the line-of-sight direction of the left eye of the user using a method described later.
- the gaze detection unit 221 identifies focus coordinates of the user's gaze, including information in the depth direction, based on the user's right and left viewing line vectors.
- the line-of-sight detection unit 221 identifies focal coordinates at which the user gazes without information in the depth direction.
- the combining unit 222 creates a combined image using the first image transmitted from the second communication unit 220 and the second image.
- the combining unit 222 holds in advance information on a positional relationship that combines the first image and the second image, and combines the first image and the second image so as to match the positional relationship.
- the positional relationship is determined according to the camera angle and the imaging range of each of the cameras 180 and 181, the distance to the user, and the like.
- the combining unit 222 can obtain a simple user's face image by combining the first image and the second image.
- the combining unit 222 transmits the face image of the user obtained by combining to the expression recognition unit 223.
- the facial expression recognition unit 223 executes facial expression recognition processing based on the composite image indicating the user's face transmitted from the composite unit 222.
- the expression recognition process is a process of extracting feature points of a face image for specifying what kind of expression the user is doing, and a process of specifying an emotion to be inferred from the user's expression from that point. May be included.
- As a method of facial expression recognition using a facial image for example, there is a method of estimating facial expression using pattern matching by extracting feature points from the facial image, and this may be used.
- the facial expression recognition unit 223 transmits the estimated facial expression of the user 300 to the video output unit 224.
- the video output unit 224 generates three-dimensional image data to be displayed on the display unit 121 of the head mounted display 100 and transmits the data to the second communication 220. Further, the video output unit 224 generates marker image data to be used for calibration for sight line detection, and transmits the marker image data to the second communication unit 220.
- the image output unit 224 holds information indicating the coordinate system of the three-dimensional image to be output and the three-dimensional position coordinates of the object displayed in the coordinate system.
- the video output unit 224 also has a function of outputting a moving image or a game image to be displayed on the display unit 121 of the head mounted display 100.
- the image output unit 224 has a function of outputting an image (avatar image) of a character operated by the user 300
- the image output unit 224 generates an image of an expression that matches the expression estimated by the expression recognition unit 223.
- Output Alternatively, for example, when the user 300 communicates with the character output by the video output unit 224 and displayed on the head mounted display 100, the video output unit 224 may use the estimated expression of the user 300. A character image showing a corresponding response is generated and output.
- the storage unit 225 is a recording medium for storing various programs and data required for the sight line detection device 200 in operation.
- FIG. 5 is a schematic view for explaining calibration for detection of the gaze direction according to the embodiment.
- the gaze direction of the user 300 is realized by the gaze detection unit 221 in the gaze detection apparatus 200 analyzing an image captured by the camera 116 and output to the gaze detection apparatus 200 by the first communication unit 118.
- the image output unit 224 generates nine points (marker images) from the points Q 1 to Q 9 as shown in FIG. 5 and causes the image display element 108 of the head mounted display 100 to display.
- the gaze detection apparatus 200 causes the user 300 to gaze in order until the points Q 1 to Q 9 are reached. At this time, the user 300 is asked to gaze at each point only with the movement of the eye as much as possible without moving the neck.
- the camera 116 captures an image including the cornea 302 of the user 300 when the user 300 gazes at nine points from the points Q 1 to Q 9 .
- FIG. 6 is a schematic view illustrating position coordinates of the cornea 302 of the user 300.
- the line-of-sight detection unit 221 in the line-of-sight detection apparatus 200 analyzes the image captured by the camera 116 and detects the bright spot 105 derived from infrared light.
- the visual axis detection unit 221 sets a two-dimensional coordinate system 306 in the image captured by the camera 116 based on the detected bright spot 105.
- the gaze detection unit 221 also detects the center P of the cornea 302 of the user 300 by analyzing the image captured by the camera 116. This can be realized by using known image processing such as Hough transform or edge extraction processing, for example. Accordingly, the sight line detection unit 221 can acquire the coordinates of the center P of the cornea 302 of the user 300 in the set two-dimensional coordinate system 306.
- the coordinates of points Q 1 to Q 9 in the two-dimensional coordinate system set on the display screen displayed by the image display element 108 are Q 1 (x 1 , y 1 ) T and Q 2 (x 2 , x 2 , x 2 y 2 ) T ... Q 9 (x 9 , x 9 ) T
- Each coordinate is, for example, the number of a pixel located at the center of each point.
- the center P of the user's 300 cornea 302 when the user 300 is gazing at the points Q 1 to Q 9 is taken as the points P 1 to P 9 , respectively.
- each coordinate of the point P 1 ⁇ P 9 in the two-dimensional coordinate system 306 P 1 (X 1, Y 1) T, P 2 (X 2, Y 2) T, ⁇ , P 9 (Z 9 , Y 9 ) T
- T represents transpose of a vector or a matrix.
- the matrix M is a matrix that projects the direction of the user's 300 gaze on the image plane displayed by the image display element 108.
- the element of the vector y is known because it is the coordinates of the points Q 1 to Q 9 which the visual axis detection unit 221 causes the image display element 108 to display.
- the elements of the matrix A can be acquired because they are the coordinates of the vertex P of the cornea 302 of the user 300. Therefore, the gaze detection unit 221 can acquire the vector y and the matrix A.
- the vector x which is a vector in which the elements of the transformation matrix M are arranged, is unknown. Therefore, the problem of estimating the matrix M becomes the problem of finding the unknown vector x when the vector y and the matrix A are known.
- the number of equations ie, the number of points Q presented to the user 300 at the time of calibration
- the number of unknowns ie, the number of elements of the vector x is 4
- the number of equations is nine, which is an overdetermined problem.
- an optimal vector x opt in the sense of minimizing the sum of squares of elements of the vector e can be obtained by the following equation (6).
- x opt (A T A) -1 A T y (6)
- -1 indicates an inverse matrix.
- the line-of-sight detection unit 221 constructs the matrix M of Expression (1) by using the elements of the determined vector x opt .
- the visual axis detection unit 221 uses the coordinates of the vertex P of the cornea 302 of the user 300 and the matrix M to follow the equation (2), on the moving image displayed by the image display element 108 by the right eye of the user 300 You can estimate where you are looking at in a two-dimensional range.
- the sight line detection unit 221 can calculate a right sighting line vector connecting the fixation point of the right eye on the image display element 108 and the vertex of the cornea of the right eye of the user.
- a left visual line vector connecting the fixation point of the left eye on the image display element 108 and the vertex of the cornea of the left eye of the user is calculated. can do.
- the line-of-sight detection unit 221 can detect the intersection as the focus of the user using the line-of-sight vector of the right eye and the line-of-sight vector of the left eye. If both line-of-sight vectors do not have an intersection point, other methods are used, for example, focusing on the middle point of a line segment connecting points at which the distance between both line-of-sight vectors is closest in both line-of-sight vectors It may be assumed that there is an intersection point between the plane and the line-of-sight vector, and a midpoint of a line connecting the intersection points may be focused. Note that the gaze position ( gaze coordinate position not including depth information) on the plane of the displayed 3D image can be specified by only one of the gaze vectors.
- FIG. 9 is a flowchart showing the operation of the facial expression recognition system 1.
- the imaging unit 124 operates the camera 180 to capture an image of the lower half of the user's face, that is, a first image (step S901).
- a first image An example of an image of the first image 701 obtained by the imaging is shown in FIG.
- the imaging unit 124 transmits the first image obtained by imaging to the first communication unit 118.
- the first communication unit 118 transmits the transmitted first image to the facial expression recognition device 200.
- the imaging unit 124 operates the camera 181 to capture an image of the upper half of the user (around the eyes), that is, a second image (step S902).
- An image example of the second image 702 obtained by the imaging is shown in FIG.
- the imaging unit 124 transmits the second image obtained by imaging to the first communication unit 118.
- the first communication unit 118 transmits the transmitted first image to the facial expression recognition device 200.
- the second communication unit 220 of the facial expression recognition device 200 having received the first image and the second image transmits the first image and the second image to the synthesis unit 222.
- the combining unit 222 combines the transmitted first image 701 and second image 702 according to a predetermined algorithm, and generates a combined image showing the face image of the user 300 (step S903).
- An image example of a composite image 801 obtained by the composition is shown in FIG.
- the combining unit 222 transmits the generated combined image 801 to the expression recognition unit 223.
- the facial expression recognition unit 223 executes facial expression recognition processing on the transmitted composite image 801 according to a predetermined algorithm, and recognizes and estimates the facial expression of the user 300 (step S904).
- the facial expression recognition unit 223 transmits the estimated facial expression information of the user 300 to the video output unit 224.
- the video output unit 224 uses the transmitted facial expression information to reflect the content (step S905).
- the above is the operation relating to the facial expression recognition of the facial expression recognition system 1.
- the recognition of the expression by the expression recognition unit 223 described above makes it possible to recognize the movement of the expression and what kind of emotion the user has. Therefore, the following application methods can be considered.
- Example 1 of application A communication system in which a plurality of head mounted displays and at least one server system are connected by communication is assumed. Then, it is assumed that a virtual reality space in which a plurality of characters operate is provided by the server system. Users wearing the head mounted display create avatars and use the avatars to move between the virtual worlds provided by the virtual reality space.
- the user's 300 expression is reflected on the corresponding avatar by estimating the user's 300 expression using the head mounted display 100 described above. This makes it possible to provide a virtual reality space closer to reality, and to make communication in the virtual reality space more active.
- the head mount display 100 described above is used to estimate the user's 300 expression and notify the server system Then, the reaction based on the user's expression is reflected to the non-player character. For example, if it is recognized that the user is smiling, the non-player character is also made to laugh or is turned, and if the user is recognized to be angry, the non-player character is made to be angry or scared. You may
- Example 3 of application As the application example 3, it is assumed that the video output unit 224 has a function of outputting an avatar image of the user. At this time, the shape of the mouth obtained based on the first image from the camera 180 is reflected as it is in the avatar image, and the shape of the eyes obtained based on the second image from the camera 181 is reflected as it is on the avatar image.
- Real avatar expression can be realized.
- FIG. 13A shows images 1301 and 1302 captured by the camera 180 and the camera 181. As shown in FIG. 13A, it can be recognized from the captured image that the user is surprised. Then, as shown in FIG.
- the video output unit 224 outputs an avatar image 1303 in which a state of being surprised of the user recognized by the sight detection system 1 is reflected.
- the video output unit 224 outputs an avatar image 1303 in which a state of being surprised of the user recognized by the sight detection system 1 is reflected.
- it is particularly effective to generate an avatar image of a three-dimensional shape.
- the application example 4 can be applied to marketing for viewing the user's reaction to the video output from the video output unit 224. That is, the gaze detection system 1 specifies an object displayed ahead of the user's gaze direction detected by the gaze detection apparatus 200 of the gaze detection system 1, and the user's impression of the object is expressed as an expression It estimates based on the user's facial expression which the recognition part 223 detected. For example, when it is recognized that the user's expression has a gentle expression, it can be inferred that the user is feeling favorably with respect to the display object, and the user's expression indicates disgust. When it is recognized that it is, it can be inferred that the display object is disgusted. Thus, for example, when the display object is a product or the like, information as to whether the user likes the product can be collected, and if such information is collected from various users, what product is It can be marketed to become popular.
- the content of the video can be changed based on the expression of the video output by the video output unit 224 indicated by the user. That is, as a video output from the video output unit 224, a branch point is provided in the video, different videos derived from the branch point are prepared, and videos having different endings are prepared as in a so-called multi-ending story. Then, depending on whether or not the user showed a favorable expression on the image, the image to be output to the user may be determined, and the image obtained by branching the story may be output. . This makes it possible to provide a video of a story that is more desirable for the user.
- the degree of difficulty of the game can be dynamically changed based on the user's expression. Specifically, when the head mounted display 100 recognizes that the expression of the user playing the game is steep, the game is difficult for the user, and the video output unit 224 Lowers the level of difficulty, and outputs a game image with a lower level of difficulty. On the contrary, when it is recognized that the user's expression makes the user feel comfortable, the game is easy for the user, so the video output unit 224 raises the difficulty of the game and raises the difficulty. Output game image.
- the video output unit 224 is described as also serving as the game engine, but the game engine is provided separately from the video output unit 224, and the video output unit 224 uses the head transmitted from the game engine as a head. It may be output to the mount display 100.
- Example 7 of application when performing real-time live viewing using the head mounted display 100, interactively changing a user image indicating the head mounted display 100 based on an image captured using the cameras 180 and 181. Can.
- the head mounted display according to the present invention it is possible to acquire the face image of the user by imaging different places with a plurality of cameras and combining them. By this, facial expression recognition can be performed and can be reflected on various contents.
- the image reflected by the hot mirror 112 is imaged.
- the eye of the user 300 may be imaged directly without passing through 112.
- the first image and the second image are captured by the cameras 180 and 181, respectively, to obtain a composite image of the face.
- the method of recognizing the user's expression is not limited to this.
- the head mounted display 100 is provided with a contact sensor, such as a myoelectric sensor, which can specify the user's facial expression at a position where the head mounted display 100 is touched when the user mounts the head mounted display 100. It may be Then, the facial expression recognition unit 223 may recognize the facial expression around the eyes based on the data indicating the movement around the eyes of the user detected by the touch sensor.
- a contact sensor such as a myoelectric sensor
- the expression of the user 300 is limited to recognition. However, depending on the imaging range based on the angle of view of the camera 180, the state of the user 300 with respect to other than the expression can also be recognized and reflected in various contents.
- the camera 180 may be arranged to capture an image so as to show up to the shoulder of the user 300. Then, as shown in FIG. 10, an image that can be recognized up to the state of the shoulder of the user 300 is obtained as a synthesized image 1001 obtained by the synthesizing unit 222 synthesizing the first image and the second image.
- a synthesized image 1001 obtained by the synthesizing unit 222 synthesizing the first image and the second image.
- the composite image 1001 in FIG. 10 since it can be analyzed that the left shoulder of the user 300 is located on the front side of the image, for example, when generating the avatar image of the user 300, the left shoulder of the avatar is inclined forward You may want to generate the
- the image recognition unit 223 analyzes the image 1001 to estimate the posture of the user's body. Then, for example, the posture of the character operated by the user may be controlled based on the estimated posture. Note that a posture estimation unit that estimates the posture of the user from the combined image may be separately provided in the expression recognition device 200.
- human posture estimation technology using conventional image analysis technology, for example, markerless motion capture technology, pattern matching using sample images indicating various postures of the user, and the like.
- the camera 180 is provided on the head mounted display 100.
- the camera 180 may be configured to be removable. An example is shown in FIG.
- FIG. 11A is a perspective view of an example of attaching the camera 180 to the head mounted display 100 as viewed from the top of the head mounted display 100
- FIG. 11B is a perspective view of the head mounted display 100 as viewed from the bottom It is.
- the camera 180 is attached to a U-shaped member 1101. Further, a slide groove 1102 is provided in the head mounted display 100. At both ends of the member 1101, projections are provided so as to fit in the slide grooves 1102. Then, the camera 180 is mounted on the head mount display by sliding the convex portion into the slide groove 1102 and inserting it. At this time, the member 1101 may be configured to be able to be fixed at several places of the slide groove 1102.
- the camera 180 may have a wireless communication function
- the first communication unit 118 of the head mounted display 100 may be configured to receive the first image captured by the camera 180.
- attachment example shown in FIG. 11 is only an example, and other methods may be used to be detachably configured.
- a mortise may be provided on the head mount display, and a tenon fitted to the mortise may be provided on the camera 180 side to be detachable, or may be realized by screwing.
- the camera 180 in the above embodiment may be rotatably provided on the head mounted display 100. That is, the camera 180 may be provided on the head mounted display 100 in the form as shown in FIG.
- FIG. 12 is an enlarged view of a side of the head mounted display 100 where a camera 180 is attached. As shown in FIG. 12, the camera 180 is attached to the head mount display 100 so as to pivot on a pivot shaft 1201 supported by the holding unit 1202. By configuring in this manner, it is possible to perform imaging at an appropriate angle when imaging the first image according to the physical constitution of the user and the like.
- the pivot shaft 1201 may be configured to be fixed at a predetermined pivot angle. By doing this, even if the user 300 moves, it is possible to prevent the angle taken by the camera 180 from changing.
- the rotation shaft 1201 may be provided with a rotation motor, and the imaging unit 124 may control the rotation motor at the time of imaging so as to be able to capture a desired first image.
- the plurality of first images may be photographed at various rotation angles, and the plurality of photographed first and second images may be synthesized by the synthesis unit 222. By doing this, a larger image showing the state of the user 300 can be acquired.
- surroundings of a user's eye was illustrated in the said embodiment, this is not the limitation.
- it may be a full-face type head mounted display, and in that case, a plurality of cameras for capturing the face of the user are provided, and facial expression recognition is performed on a face image obtained by combining images captured by each. You may
- the synthesizing unit 222 is provided, and after the images captured by the camera 180 and the camera 181 are synthesized, recognition of the user's expression is realized.
- the line-of-sight detection system 1 specifies the shape of the user's mouth based on the image captured by the camera 180 without the combining unit 222, and determines the shape of the user's eye based on the image captured by the camera 181
- facial expression recognition may be realized based on the shapes of the eyes and the mouth specified independently.
- the avatar image generation unit provided in the sight line detection system 1 generates the avatar image of the user without recognizing the expression
- the shapes of the eyes and the mouth detected in the partial portions It may be made to reflect for each part. That is, for example, the shape of the user's mouth may be specified based on the image captured by the camera 180, and only the specified shape of the mouth may be reflected in the avatar image.
- the storage unit 225 may realize imaging for gaze detection and expression recognition in advance by the following method of classifying the user's expression. For example, a classification of anger, aversion, fear, happiness, sadness, surprise, etc. is prepared, and the pattern of the face image showing the expression according to each classification (corresponding to the arrangement of each part of the face and the expression according to each emotion) Stores a correspondence table in which the shape patterns of the parts to be Then, the facial expression recognition system identifies a pattern of a face image corresponding to the identified classification based on the classification corresponding to the facial expression recognized by the facial expression recognition unit 223, and generates an avatar image reflecting the identified pattern. An image generation unit may be provided.
- each classification may be associated with a pattern of a face image according to the degree of each expression (emotion). For example, to illustrate the classification of anger, there are five levels from “a bit angry” to "very angry”, and if you are very angry, it will be better than if you are a little angry Also, it is preferable to associate patterns of face images in which the degree of elevation of the eyelid is large, the degree of downward angle of the mouth is large, and the degree of slight swelling is large.
- the facial expression recognition unit 223 also determines the stage of each of the classified facial expressions.
- the facial expression recognition system may realize facial expression recognition and reflect it on the avatar image.
- the camera 116 and the camera 181 are used as separate cameras, but these cameras may be shared cameras.
- the camera 116 is used without using the camera 181, and a visible light camera is adopted as the camera 116.
- the eye captures the eye three-dimensionally with a stereo camera, and three-dimensionally captures the shape of the eyeball. Detect the direction.
- the image as it is is used for facial expression recognition.
- the camera 116 a camera provided with both imaging in the visible light mode and imaging in the infrared mode is used, and the head mounted display 100 detects the line of sight in the infrared mode.
- switching may be performed so as to perform imaging in the visible light mode. The switching can be realized, for example, by switching the filters of the infrared pass filter and the visible light pass filter.
- the camera 116 is used without using the camera 181
- the hot mirror 112 need not be provided.
- the processor of the facial expression recognition apparatus 200 executes the line-of-sight detection program or the like to identify a portion at which the user is gazing.
- This may be realized by a logic circuit (hardware) or a dedicated circuit formed in an integrated circuit (IC (Integrated Circuit) chip, LSI (Large Scale Integration)) or the like in the facial expression recognition apparatus 200.
- these circuits may be realized by one or more integrated circuits, and the functions of the plurality of functional units shown in the above embodiments may be realized by one integrated circuit.
- An LSI may be called a VLSI, a super LSI, an ultra LSI, or the like depending on the degree of integration.
- the line-of-sight detection program may be recorded on a recording medium readable by a processor, and as the recording medium, “non-temporary tangible medium” such as tape, disk, card, semiconductor memory, programmable logic A circuit or the like can be used.
- the search program may be supplied to the processor via any transmission medium (communication network, broadcast wave, etc.) capable of transmitting the search program.
- the present invention can also be realized in the form of a data signal embedded in a carrier wave, in which the gaze detection program is embodied by electronic transmission.
- the gaze detection program uses, for example, a script language such as ActionScript, JavaScript (registered trademark), Python, or Ruby, or a compiler language such as C language, C ++, C #, Objective-C, Java (registered trademark), or the like. Can be implemented.
- a script language such as ActionScript, JavaScript (registered trademark), Python, or Ruby
- a compiler language such as C language, C ++, C #, Objective-C, Java (registered trademark), or the like. Can be implemented.
- Reference Signs List 1 facial expression recognition system 100 head mounted display, 103a infrared light source (second infrared light irradiation unit), 103b infrared light source (first infrared light irradiation unit), 105 bright spot, 108 image display element, 112 hot mirror , 114, 114a, 114b convex lens, 116 camera, 118 first communication unit, 121 display unit, 122 infrared light irradiation unit, 123 image processing unit, 124 imaging unit, 130 image display system, 150 housing, 152a, 152b lens Holding unit, 160 mounting tools, 170 headphones, 180, 181 cameras, 200 expression recognition devices, 220 second communication units, 221 gaze detection units, 222 synthesis units, 223 expression recognition units, 224 video output units, 225 storage units.
- the present invention is applicable to a head mounted display.
Abstract
Description
上述のようなヘッドマウントディスプレイにおいては、ユーザの表情を認識できれば、よりリアルかつアクティブなコンテンツを提供することができる。例えば、ユーザが制御するキャラクタの表情をユーザの表情に応じて変えたり、ヘッドマウントディスプレイに表示するキャラクタの対応を変えたりするといった使用方法が考えられる。
本発明の一態様に係る表情認識システム1は、ユーザの眼を撮像する第1カメラ(181)、ユーザの口元を撮像する第2カメラ(180)、並びに、第1カメラにより撮像された第1画像及び第2カメラにより撮像された第2画像を出力する出力部(118)を有するヘッドマウントディスプレイ(100)と、出力部により出力された第1画像及び第2画像を受信する受信部(220)、受信部により受信された第1画像及び第2画像を合成して合成画像を作成する合成部(222)、並びに、合成部により作成された合成画像に基づいて、ユーザの表情を認識する表情認識部(223)を有する表情認識装置(200)とを備える表情認識システムである。
PN=MQN (N=1,・・・,9) (2)
xopt=(ATA)-1ATy (6)
ここで「-1」は逆行列を示す。
ここから、表情認識システム1における表情認識に係る動作を説明する。図9は、表情認識システム1の動作を示すフローチャートである。
以上が、表情認識システム1の表情認識に係る動作である。
ここで、上記表情認識システムが実行する表情認識のコンテンツの反映方法について説明する。
複数のヘッドマウントディスプレイと少なくとも1台のサーバシステムが通信により接続された通信システムを想定する。そして、当該サーバシステムにより、複数のキャラクタが動作する仮想現実空間が提供されているものとする。ヘッドマウントディスプレイを装着したユーザ達は、各々のアバターを作成し、そのアバターを用いて、仮想現実空間により提供される仮想世界を行き来するものとする。
活用例2においても、活用例1と同様のシステムを想定する。そして、サーバシステムは、ユーザが操作していない、いわゆるノンプレイヤーキャラクターを操作しているものとする。
活用例3としては、映像出力部224が、ユーザのアバター画像を出力する機能を有する場合を想定する。このとき、カメラ180からの第1画像に基づいて得られる口の形状をそのままアバター画像に反映させ、カメラ181からの第2画像に基づいて得られる目の形状をそのままアバター画像に反映させることにより、リアルなアバター表現を実現できる。図13にその一例を示す。図13(a)には、カメラ180とカメラ181とが撮像した画像1301、1302を示している。図13(a)に示すように、撮像された画像からユーザが驚いている様子が認識できる。そして、映像出力部224は、図13(b)に示すように、視線検出システム1が認識したユーザの驚いている様子を反映させたアバター画像1303を出力する。このとき、カメラ180及びカメラ181として深度カメラを用いる場合には、3次元形状のアバター画像を生成するのに特に有効である。
活用例4としては、映像出力部224が出力する映像に対するユーザの反応を見るためのマーケティングに応用することができる。すなわち、視線検出システム1は、視線検出システム1の視線検出装置200が検出したユーザの視線方向の先に表示している対象物を特定し、その対象物に対してユーザが抱いた感想を表情認識部223が検出したユーザの表情に基づいて推定する。例えば、ユーザの表情が優しげな表情をしていることを認識した場合には、表示対象物に対して好意的感情を抱いていることを推定できるし、ユーザの表情が嫌悪感を示すものであることを認識した場合には、表示対象物に対して反感を抱いていることを推定できる。これにより、例えば、表示対象物が何らかの商品などである場合に、その商品をユーザが気に入ったかどうかの情報を収集でき、そのような情報を様々なユーザから収取すれば、どのような商品が人気が高くなるかのマーケティングを行うことができる。
活用例5としては、映像出力部224が出力する映像に対してユーザが示した表情に基づいて、映像の内容を変更することができる。すなわち、映像出力部224が出力する映像として、映像に分岐点を設け、その分岐点から派生する互いに異なる映像を用意し、いわばマルチエンディングストーリーのように、結末が異なるような映像を用意する。そして、ユーザが映像に対して示した表情について、好意的な表情を示したか否かに応じて、ユーザに対して出力する映像を決定し、ストーリーを分岐させた映像を出力することしてもよい。これにより、ユーザにとってより好ましいストーリーの映像を提供することができる。
活用例6としては、映像出力部224がゲーム画像を出力している場合に、ユーザの表情に基づいて、ゲームの難易度を動的に変更することができる。具体的には、ヘッドマウントディスプレイ100を用いてゲームをプレイしているユーザの表情が険しいものであると認識した場合に、ユーザにとってそのゲームは難しいことになるので、映像出力部224はゲームの難易度を下げ、難易度が下がったゲーム画像を出力する。逆に、ユーザの表情が余裕を感じさせるものであると認識した場合に、ユーザにとってそのゲームは容易であることになるので、映像出力部224はゲームの難易度を上げ、難易度が上がったゲーム画像を出力する。なお、ここでは、映像出力部224がゲームエンジンを兼任するものとして記載しているが、ゲームエンジンは映像出力部224とは別に設け、映像出力部224は、ゲームエンジンから伝達される画像をヘッドマウントディスプレイ100に出力することとしてもよい。
活用例7としては、ヘッドマウントディスプレイ100を利用したリアルタイムの実況を行う際に、ヘッドマウントディスプレイ100を示すユーザ画像を、カメラ180、181を用いて撮像した画像に基づいて、インタラクティブに変化させることができる。
上述のように、本発明に係るヘッドマウントディスプレイによれば、ユーザの顔画像を、複数のカメラでそれぞれ別の箇所を撮像して、合成することで、取得することができる。これによって、表情認識を行うことができ、様々なコンテンツに反映することができる。
本発明に係る表情認識システムは、上記実施の形態に限定されるものではなく、その発明の思想を実現するための他の手法により実現されてもよいことは言うまでもない。以下、本発明の思想として含まれる例を説明する。
以下の手法により、視線検出と表情認識のための撮像を実現してもよい。例えば、怒り、嫌悪、恐れ、幸福感、悲しみ、驚きなどの分類を用意し、各分類に応じた表情を示す顔画像のパターン(顔の各パーツの配置やそれぞれの感情に応じた表情に対応するパーツの形状パターン)を対応付けた対応表を記憶する。そして、表情認識システムは、表情認識部223が認識した表情に該当する分類に基づいて、特定した分類に対応する顔画像のパターンを特定し、特定したパターンを反映させたアバター画像を生成するアバター画像生成部を備えることとしてもよい。
Claims (12)
- ユーザの眼を撮像する第1カメラ、
前記ユーザの口元を撮像する第2カメラ、並びに、
前記第1カメラにより撮像された第1画像及び前記第2カメラにより撮像された第2画像を出力する出力部
を有するヘッドマウントディスプレイと、
前記出力部により出力された前記第1画像及び前記第2画像を受信する受信部、並びに、
前記第1画像及び前記第2画像に基づいて、前記ユーザの表情を認識する表情認識部
を有する表情認識装置と
を備える表情認識システム。 - 前記ヘッドマウントディスプレイは、さらに、
前記ユーザの眼に非可視光を照射する光源、及び
前記ユーザの眼により反射された前記非可視光を撮像する第3カメラ
を有し、
前記出力部は、前記第3カメラが撮像した第3画像を出力し、
前記表情認識装置は、さらに、
前記受信部により受信された前記第3画像に基づいて、前記ユーザの視線方向を検出する視線検出部を有する
ことを特徴とする請求項1に記載の表情認識システム。 - 前記表情認識装置は、さらに、
前記受信部により受信された前記第1画像及び前記第2画像を合成して合成画像を作成する合成部を備え、
前記表情認識部は、前記合成画像に基づいて、前記ユーザの表情を認識する
ことを特徴とする請求項1又は2に記載の表情認識システム。 - 前記第2カメラは、前記ヘッドマウントディスプレイに着脱自在に取り付けられることを特徴とする請求項1~3のいずれか1項に記載の表情認識システム。
- 前記第2カメラは、前記ユーザが前記ヘッドマウントディスプレイを装着した場合に、前記ユーザの鼻から肩までの範囲を撮像可能な画角となるように、前記ヘッドマウントディスプレイに取り付けられていることを特徴とする請求項1~3のいずれか1項に記載の表情認識システム。
- 前記表情認識装置は、さらに、
前記受信部により受信された前記第2画像に基づいて、前記ユーザの姿勢を推定する姿勢推定部を有する
ことを特徴とする請求項1~5のいずれか1項に記載の表情認識システム。 - 前記ヘッドマウントディスプレイは、ユーザの眼の周囲を覆い、ユーザの口元を覆わないように構成されていることを特徴とする請求項1~6のいずれか1項に記載の表情認識システム。
- 前記第1カメラ及び前記第2カメラは撮像対象までの距離を示す深度情報を取得するカメラであり、
前記表情認識システムは、さらに、
前記第1カメラが撮像したユーザの眼と、前記第2カメラが撮像したユーザの口元との画像に基づいて、ユーザの眼と口の3次元形状を特定し、特定した3次元形状に基づいて、ユーザのアバターの眼と口の形状に反映させたアバター画像を生成するアバター画像生成部を備える
ことを特徴とする請求項1~7のいずれか1項に記載の表情認識システム。 - 前記表情認識装置は、さらに、複数の表情の分類に応じて、顔画像のパターンを記憶した対応表を記憶する記憶部を備え、
前記表情認識部は、ユーザの表情がいずれの分類に該当するかを前記第2画像に基づいて認識し、
前記表情認識システムは、さらに、
前記対応表を参照して、前記表情認識部が認識した表情に対応する顔画像のパターンを特定し、特定した顔画像のパターンに基づいて前記ユーザのアバター画像を生成するアバター画像生成部を備える
ことを特徴とする請求項1~7のいずれか1項に記載の表情認識システム。 - 前記第1カメラ及び第3カメラは、同一のカメラである
ことを特徴とする請求項2~9のいずれか1項に記載の表情認識システム。 - ヘッドマウントディスプレイと、表情認識装置とを含む表情認識システムによる表情認識方法であって、
前記ヘッドマウントディスプレイによりユーザの眼を示す第1画像を撮像する第1撮像ステップと、
前記ヘッドマウントディスプレイにより前記ユーザの口元を示す第2画像を撮像する第2撮像ステップと、
前記表情認識装置により前記第1画像と前記第2画像とを合成し、合成画像を作成する合成ステップと、
前記合成画像に基づいて、前記ユーザの表情を認識する認識ステップとを含む表情認識方法。 - 表情認識装置のコンピュータに、
ヘッドマウントディスプレイにより撮像されたユーザの眼を示す第1画像を取得する第1取得機能と、
前記ヘッドマウントディスプレイにより撮像された前記ユーザの口元を示す第2画像を取得する第2取得機能と、
前記第1画像と前記第2画像とを合成し、合成画像を作成する合成機能と、
前記合成画像に基づいて、前記ユーザの表情を認識する認識機能とを実現させる表情認識プログラム。
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201680079424.9A CN108604291A (zh) | 2016-01-13 | 2016-01-13 | 表情辨识系统、表情辨识方法及表情辨识程序 |
US16/069,984 US20190012528A1 (en) | 2016-01-13 | 2016-01-13 | Facial expression recognition system, facial expression recognition method, and facial expression recognition program |
PCT/JP2016/050869 WO2017122299A1 (ja) | 2016-01-13 | 2016-01-13 | 表情認識システム、表情認識方法及び表情認識プログラム |
JP2017561444A JP6845982B2 (ja) | 2016-01-13 | 2016-01-13 | 表情認識システム、表情認識方法及び表情認識プログラム |
KR1020187021228A KR20180102591A (ko) | 2016-01-13 | 2016-01-13 | 표정 인식 시스템, 표정 인식 방법 및 표정 인식 프로그램 |
TW106100229A TW201732499A (zh) | 2016-01-13 | 2017-01-05 | 表情辨識系統、表情辨識方法及表情辨識程式 |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/JP2016/050869 WO2017122299A1 (ja) | 2016-01-13 | 2016-01-13 | 表情認識システム、表情認識方法及び表情認識プログラム |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2017122299A1 true WO2017122299A1 (ja) | 2017-07-20 |
Family
ID=59311772
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2016/050869 WO2017122299A1 (ja) | 2016-01-13 | 2016-01-13 | 表情認識システム、表情認識方法及び表情認識プログラム |
Country Status (6)
Country | Link |
---|---|
US (1) | US20190012528A1 (ja) |
JP (1) | JP6845982B2 (ja) |
KR (1) | KR20180102591A (ja) |
CN (1) | CN108604291A (ja) |
TW (1) | TW201732499A (ja) |
WO (1) | WO2017122299A1 (ja) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2019134091A1 (en) * | 2018-01-04 | 2019-07-11 | Microsoft Technology Licensing, Llc | Providing emotional care in a session |
WO2019156651A1 (en) | 2018-02-06 | 2019-08-15 | Hewlett-Packard Development Company, L.P. | Constructing images of users' faces by stitching non-overlapping images |
WO2019177869A1 (en) | 2018-03-16 | 2019-09-19 | Magic Leap, Inc. | Facial expressions from eye-tracking cameras |
JP2020057207A (ja) * | 2018-10-02 | 2020-04-09 | 株式会社スクウェア・エニックス | 描画プログラム、記録媒体、描画制御装置、描画制御方法 |
KR20200060120A (ko) | 2018-11-22 | 2020-05-29 | 연세대학교 산학협력단 | 헤드셋 사용자 표정인식방법 |
JP7339420B1 (ja) | 2022-12-20 | 2023-09-05 | カバー株式会社 | プログラム、方法、情報処理装置 |
US11933974B2 (en) | 2019-02-22 | 2024-03-19 | Semiconductor Energy Laboratory Co., Ltd. | Glasses-type electronic device |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018033137A1 (zh) * | 2016-08-19 | 2018-02-22 | 北京市商汤科技开发有限公司 | 在视频图像中展示业务对象的方法、装置和电子设备 |
JP6913164B2 (ja) | 2016-11-11 | 2021-08-04 | マジック リープ, インコーポレイテッドMagic Leap,Inc. | 完全な顔画像の眼球周囲およびオーディオ合成 |
CN108510542B (zh) * | 2018-02-12 | 2020-09-11 | 北京七鑫易维信息技术有限公司 | 匹配光源与光斑的方法和装置 |
US10509467B1 (en) | 2018-06-01 | 2019-12-17 | Facebook Technologies, Llc | Determining fixation of a user's eyes from images of portions of the user's face enclosed by a head mounted display |
US20200090392A1 (en) * | 2018-09-19 | 2020-03-19 | XRSpace CO., LTD. | Method of Facial Expression Generation with Data Fusion |
JP7361262B2 (ja) * | 2019-03-29 | 2023-10-16 | パナソニックIpマネジメント株式会社 | 精算決済装置および無人店舗システム |
WO2021231900A1 (en) * | 2020-05-15 | 2021-11-18 | Cornell University | Wearable devices for facial expression recognition |
CN111709302B (zh) * | 2020-05-21 | 2023-08-25 | 青岛小鸟看看科技有限公司 | 人脸表情检测方法、装置、头戴显示设备及存储介质 |
GB2621868A (en) * | 2022-08-25 | 2024-02-28 | Sony Interactive Entertainment Inc | An image processing method, device and computer program |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH1196366A (ja) * | 1997-09-19 | 1999-04-09 | Nippon Telegr & Teleph Corp <Ntt> | ヘッドマウントディスプレイを装着した人物の顔画像合成方法およびその装置 |
JP2003015816A (ja) * | 2001-06-29 | 2003-01-17 | Honda Motor Co Ltd | ステレオカメラを使用した顔・視線認識装置 |
JP2008140290A (ja) * | 2006-12-05 | 2008-06-19 | Suzuki Motor Corp | 頭部の位置・姿勢検出装置 |
JP2014021707A (ja) * | 2012-07-18 | 2014-02-03 | Nikon Corp | 情報入出力装置、及び情報入出力方法 |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7106358B2 (en) * | 2002-12-30 | 2006-09-12 | Motorola, Inc. | Method, system and apparatus for telepresence communications |
US7606392B2 (en) * | 2005-08-26 | 2009-10-20 | Sony Corporation | Capturing and processing facial motion data |
US8477175B2 (en) * | 2009-03-09 | 2013-07-02 | Cisco Technology, Inc. | System and method for providing three dimensional imaging in a network environment |
FR2970576B1 (fr) * | 2011-01-19 | 2013-02-08 | Matchic Labs | Procede de determination de la direction du regard et dispositif pour sa mise en oeuvre |
JP2012181704A (ja) * | 2011-03-01 | 2012-09-20 | Sony Computer Entertainment Inc | 情報処理装置および情報処理方法 |
WO2013077076A1 (ja) * | 2011-11-24 | 2013-05-30 | 株式会社エヌ・ティ・ティ・ドコモ | 表情出力装置及び表情出力方法 |
JP5938977B2 (ja) * | 2012-03-23 | 2016-06-22 | ソニー株式会社 | ヘッドマウントディスプレイ及び手術システム |
US9019174B2 (en) * | 2012-10-31 | 2015-04-28 | Microsoft Technology Licensing, Llc | Wearable emotion detection and feedback system |
US10108018B2 (en) * | 2012-11-13 | 2018-10-23 | Sony Corporation | Image display apparatus for displaying an image captured by a mobile apparatus |
WO2014106823A2 (en) * | 2013-01-03 | 2014-07-10 | Meta Company | Extramissive spatial imaging digital eye glass apparatuses, methods and systems for virtual or augmediated vision, manipulation, creation, or interaction with objects, materials, or other entities |
JP5962547B2 (ja) * | 2013-03-08 | 2016-08-03 | ソニー株式会社 | 情報処理装置、情報処理方法およびプログラム |
CA3160567A1 (en) * | 2013-03-15 | 2014-09-18 | Magic Leap, Inc. | Display system and method |
JP6330258B2 (ja) * | 2013-05-15 | 2018-05-30 | セイコーエプソン株式会社 | 虚像表示装置 |
US9256987B2 (en) * | 2013-06-24 | 2016-02-09 | Microsoft Technology Licensing, Llc | Tracking head movement when wearing mobile device |
JP2015092646A (ja) * | 2013-11-08 | 2015-05-14 | ソニー株式会社 | 情報処理装置、制御方法、およびプログラム |
DE102016110903A1 (de) * | 2015-06-14 | 2016-12-15 | Facense Ltd. | Head-Mounted-Devices zur Messung physiologischer Reaktionen |
CN105183147A (zh) * | 2015-08-03 | 2015-12-23 | 众景视界(北京)科技有限公司 | 头戴式智能设备及其建模三维虚拟肢体的方法 |
-
2016
- 2016-01-13 WO PCT/JP2016/050869 patent/WO2017122299A1/ja active Application Filing
- 2016-01-13 US US16/069,984 patent/US20190012528A1/en active Pending
- 2016-01-13 JP JP2017561444A patent/JP6845982B2/ja active Active
- 2016-01-13 CN CN201680079424.9A patent/CN108604291A/zh not_active Withdrawn
- 2016-01-13 KR KR1020187021228A patent/KR20180102591A/ko unknown
-
2017
- 2017-01-05 TW TW106100229A patent/TW201732499A/zh unknown
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH1196366A (ja) * | 1997-09-19 | 1999-04-09 | Nippon Telegr & Teleph Corp <Ntt> | ヘッドマウントディスプレイを装着した人物の顔画像合成方法およびその装置 |
JP2003015816A (ja) * | 2001-06-29 | 2003-01-17 | Honda Motor Co Ltd | ステレオカメラを使用した顔・視線認識装置 |
JP2008140290A (ja) * | 2006-12-05 | 2008-06-19 | Suzuki Motor Corp | 頭部の位置・姿勢検出装置 |
JP2014021707A (ja) * | 2012-07-18 | 2014-02-03 | Nikon Corp | 情報入出力装置、及び情報入出力方法 |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2019134091A1 (en) * | 2018-01-04 | 2019-07-11 | Microsoft Technology Licensing, Llc | Providing emotional care in a session |
US11369297B2 (en) | 2018-01-04 | 2022-06-28 | Microsoft Technology Licensing, Llc | Providing emotional care in a session |
US20220132037A1 (en) * | 2018-02-06 | 2022-04-28 | Hewlett-Packard Development Company, L.P. | Constructing images of users' faces by stitching non-overlapping images |
CN111279390A (zh) * | 2018-02-06 | 2020-06-12 | 惠普发展公司,有限责任合伙企业 | 通过拼接不重叠图像构建用户面部图像 |
EP3682420A4 (en) * | 2018-02-06 | 2021-05-05 | Hewlett-Packard Development Company, L.P. | CONSTRUCTION OF IMAGES OF USERS 'FACES BY ASSEMBLING NON-OVERLAPPING IMAGES |
US11228709B2 (en) * | 2018-02-06 | 2022-01-18 | Hewlett-Packard Development Company, L.P. | Constructing images of users' faces by stitching non-overlapping images |
WO2019156651A1 (en) | 2018-02-06 | 2019-08-15 | Hewlett-Packard Development Company, L.P. | Constructing images of users' faces by stitching non-overlapping images |
US11727544B2 (en) | 2018-02-06 | 2023-08-15 | Hewlett-Packard Development Company, L.P. | Constructing images of users' faces by stitching non-overlapping images |
CN111279390B (zh) * | 2018-02-06 | 2023-11-07 | 惠普发展公司,有限责任合伙企业 | 用于构建用户面部图像的装置和方法以及介质 |
EP3766004A4 (en) * | 2018-03-16 | 2021-12-15 | Magic Leap, Inc. | FACIAL EXPRESSIONS OBTAINED FROM EYE-TRACKING CAMERAS |
WO2019177869A1 (en) | 2018-03-16 | 2019-09-19 | Magic Leap, Inc. | Facial expressions from eye-tracking cameras |
JP2020057207A (ja) * | 2018-10-02 | 2020-04-09 | 株式会社スクウェア・エニックス | 描画プログラム、記録媒体、描画制御装置、描画制御方法 |
KR20200060120A (ko) | 2018-11-22 | 2020-05-29 | 연세대학교 산학협력단 | 헤드셋 사용자 표정인식방법 |
US11933974B2 (en) | 2019-02-22 | 2024-03-19 | Semiconductor Energy Laboratory Co., Ltd. | Glasses-type electronic device |
JP7339420B1 (ja) | 2022-12-20 | 2023-09-05 | カバー株式会社 | プログラム、方法、情報処理装置 |
Also Published As
Publication number | Publication date |
---|---|
JPWO2017122299A1 (ja) | 2018-11-15 |
TW201732499A (zh) | 2017-09-16 |
JP6845982B2 (ja) | 2021-03-24 |
KR20180102591A (ko) | 2018-09-17 |
CN108604291A (zh) | 2018-09-28 |
US20190012528A1 (en) | 2019-01-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6845982B2 (ja) | 表情認識システム、表情認識方法及び表情認識プログラム | |
US11883104B2 (en) | Eye center of rotation determination, depth plane selection, and render camera positioning in display systems | |
KR102417177B1 (ko) | 인사이드-아웃 위치, 사용자 신체 및 환경 추적을 갖는 가상 및 혼합 현실을 위한 머리 장착 디스플레이 | |
US9105210B2 (en) | Multi-node poster location | |
US9035970B2 (en) | Constraint based information inference | |
TW201804314A (zh) | 影像顯示系統、影像顯示方法、影像顯示程式 | |
CN110023814A (zh) | 由可穿戴设备进行面部模型捕获 | |
US11822718B2 (en) | Display systems and methods for determining vertical alignment between left and right displays and a user's eyes | |
KR101892735B1 (ko) | 직관적인 상호작용 장치 및 방법 | |
US11868525B2 (en) | Eye center of rotation determination with one or more eye tracking cameras | |
JP2017102732A (ja) | 表示制御装置及び表示制御方法 | |
US11557020B2 (en) | Eye tracking method and apparatus | |
US20180373414A1 (en) | Method for communicating via virtual space, program for executing the method on computer, and information processing apparatus for executing the program | |
US11743447B2 (en) | Gaze tracking apparatus and systems | |
JP2018000308A (ja) | 映像表示装置システム、心拍特定方法、心拍特定プログラム | |
KR20190038296A (ko) | 화상 표시 시스템, 화상 표시 방법 및 화상 표시 프로그램 | |
US20240040099A1 (en) | Depth of field in video based on gaze | |
D'Angelo et al. | Towards a Low-Cost Augmented Reality Head-Mounted Display with Real-Time Eye Center Location Capability | |
WO2023244267A1 (en) | Systems and methods for human gait analysis, real-time feedback and rehabilitation using an extended-reality device | |
WO2023195995A1 (en) | Systems and methods for performing a motor skills neurological test using augmented or virtual reality | |
JP2019125397A (ja) | 表示制御装置及び表示制御方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 16884902 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2017561444 Country of ref document: JP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 20187021228 Country of ref document: KR Kind code of ref document: A |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 16884902 Country of ref document: EP Kind code of ref document: A1 |