US20200285310A1 - Information processing apparatus, information processing method, and program - Google Patents
Information processing apparatus, information processing method, and program Download PDFInfo
- Publication number
- US20200285310A1 US20200285310A1 US16/753,880 US201816753880A US2020285310A1 US 20200285310 A1 US20200285310 A1 US 20200285310A1 US 201816753880 A US201816753880 A US 201816753880A US 2020285310 A1 US2020285310 A1 US 2020285310A1
- Authority
- US
- United States
- Prior art keywords
- eye
- user
- image
- information
- image display
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000010365 information processing Effects 0.000 title claims abstract description 70
- 238000003672 processing method Methods 0.000 title claims description 15
- 238000001514 detection method Methods 0.000 claims description 21
- 238000003384 imaging method Methods 0.000 claims description 7
- 208000016339 iris pattern Diseases 0.000 claims description 4
- 238000005516 engineering process Methods 0.000 abstract description 34
- 238000012545 processing Methods 0.000 description 31
- 210000003128 head Anatomy 0.000 description 24
- 238000000034 method Methods 0.000 description 17
- 230000002123 temporal effect Effects 0.000 description 16
- 230000001936 parietal effect Effects 0.000 description 15
- 230000006870 function Effects 0.000 description 14
- 230000000007 visual effect Effects 0.000 description 13
- 238000010191 image analysis Methods 0.000 description 10
- 238000004458 analytical method Methods 0.000 description 8
- 208000003464 asthenopia Diseases 0.000 description 6
- 210000004556 brain Anatomy 0.000 description 6
- 238000004364 calculation method Methods 0.000 description 6
- 238000004891 communication Methods 0.000 description 6
- 230000014759 maintenance of location Effects 0.000 description 6
- 239000011521 glass Substances 0.000 description 5
- 230000000694 effects Effects 0.000 description 4
- 210000001747 pupil Anatomy 0.000 description 4
- 230000001133 acceleration Effects 0.000 description 3
- 230000003190 augmentative effect Effects 0.000 description 3
- 238000012937 correction Methods 0.000 description 3
- 238000005401 electroluminescence Methods 0.000 description 3
- 239000004973 liquid crystal related substance Substances 0.000 description 3
- 238000010801 machine learning Methods 0.000 description 3
- 208000001491 myopia Diseases 0.000 description 3
- 238000013528 artificial neural network Methods 0.000 description 2
- 230000017531 blood circulation Effects 0.000 description 2
- 239000000470 constituent Substances 0.000 description 2
- 238000013527 convolutional neural network Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 210000005069 ears Anatomy 0.000 description 2
- 239000003550 marker Substances 0.000 description 2
- 230000003183 myoelectrical effect Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000002040 relaxant effect Effects 0.000 description 2
- 229910001285 shape-memory alloy Inorganic materials 0.000 description 2
- 210000004243 sweat Anatomy 0.000 description 2
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 2
- 206010020675 Hypermetropia Diseases 0.000 description 1
- 206010041349 Somnolence Diseases 0.000 description 1
- 230000032683 aging Effects 0.000 description 1
- 230000004397 blinking Effects 0.000 description 1
- 210000004204 blood vessel Anatomy 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000001186 cumulative effect Effects 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- -1 dimmer sheets Substances 0.000 description 1
- 239000000428 dust Substances 0.000 description 1
- 210000000720 eyelash Anatomy 0.000 description 1
- 210000001061 forehead Anatomy 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 210000003205 muscle Anatomy 0.000 description 1
- 230000001151 other effect Effects 0.000 description 1
- 230000000306 recurrent effect Effects 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000001356 surgical procedure Methods 0.000 description 1
- 230000035900 sweating Effects 0.000 description 1
- 238000002834 transmittance Methods 0.000 description 1
- 239000013585 weight reducing agent Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/0093—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0172—Head mounted characterised by optical features
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0179—Display position adjusting means not related to the information to be displayed
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/327—Calibration thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/332—Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
- H04N13/344—Displays for viewing with the aid of special glasses or head-mounted displays [HMD] with head-mounted left-right displays
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/366—Image reproducers using viewer tracking
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0138—Head-up displays characterised by optical features comprising image capture systems, e.g. camera
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/014—Head-up displays characterised by optical features comprising information/image processing systems
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B2027/0178—Eyeglass type
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0179—Display position adjusting means not related to the information to be displayed
- G02B2027/0187—Display position adjusting means not related to the information to be displayed slaved to motion of at least a part of the body of the user, e.g. head, eye
Definitions
- the present technology relates to an information processing apparatus, an information processing method, and a program applicable to image display apparatuses such as HMDs (Head Mount Displays).
- HMDs Head Mount Displays
- Patent Literature 1 describes a technology to detect user's visual lines to control an image display. For example, user's eyes are shot by cameras installed in an HMD, and a user's visual line direction and an observation point on the HMD are calculated on the basis of shot images. A high-quality partial image is displayed about the observation point, and a low-quality image is displayed on the periphery of the partial image. Thus, it is possible to control an image display on the basis of a region capable of being recognized by a user with high accuracy (paragraphs [0051], [0052], [0060], and [0066], FIGS. 1 and 3 , etc., of the specification of Patent Literature 1).
- Patent Literature 2 a plurality of image data items obtained by shooting a subject at a plurality of different positions is appropriately selected and combined together on the basis of a user's position and a focal position. Combined image data is displayed on a display as a presentation image.
- a blurred image an image having prescribed depth of field
- Patent Literature 1 Japanese Patent Application Laid-open No. 2016-191845
- Patent Literature 2 Japanese Patent Application Laid-open No. 2005-227950
- an information processing apparatus includes an acquisition unit and a movement unit.
- the acquisition unit acquires eye information regarding an eye of a user.
- the movement unit moves, on the basis of the acquired eye information, at least a part of an image display mechanism that emits and guides image light to the eye of the user.
- the information processing apparatus at least a part of the image display mechanism is moved on the basis of eye information regarding an eye of a user. Thus, it is possible to realize high-quality viewing experience for the user.
- the image display mechanism may have a display that emits the image light and a lens system that guides the image light emitted from the display to the eye of the user.
- the movement unit may move at least one of the display or the lens system.
- the eye information may include an eye image obtained by shooting the eye of the user.
- the eye information may include at least one of a shape, a size, a position, an inclination, or an iris pattern of the eye of the user in the eye image.
- the image display mechanism may be provided in an HMD (Head Mount Display) wearable by the user.
- the eye information may include the eye image obtained by shooting the eye of the user with an imaging mechanism of the HMD.
- the information processing apparatus may further include a storage unit that stores reference eye information representing the eye information obtained when the eye of the user is put in a prescribed state with respect to the image display mechanism.
- the movement unit may move at least a part of the image display mechanism on the basis of the acquired eye information and the reference eye information.
- the prescribed state may include a state in which a reference image displayed by the image display mechanism is arranged at a prescribed size and a prescribed position with respect to the eye of the user inside the eye image.
- the reference eye information may include a reference eye image obtained by shooting the eye of the user when the eye of the user is put in the prescribed state with respect to the image display mechanism.
- the movement unit may move at least a part of the image display mechanism on the basis of a difference between the eye image and the reference eye image.
- the movement unit may move at least a part of the image display mechanism such that a state of the eye of the user with respect to the image display mechanism comes close to the prescribed state.
- the movement unit may move at least a part of the image display mechanism such that a state of the eye of the user with respect to the image display mechanism is put in another state different from the prescribed state.
- the information processing apparatus may further include a display control unit that controls an image display by the image display mechanism.
- the display control unit may move, on the basis of the acquired eye information, a display position of an image displayed by the image display mechanism.
- the movement unit may move at least a part of the image display mechanism on the basis of details of content displayed by the image display mechanism.
- the movement unit may move at least a part of the image display mechanism on the basis of a viewing time of the user.
- the information processing apparatus may further include a state acquisition unit that acquires state information regarding a state of the user.
- the movement unit may move at least a part of the image display mechanism on the basis of the acquired state information.
- the information processing apparatus may further include a determination unit that determines reliability of a detection result by a biosensor on the basis of a movement amount by the movement unit with respect to at least a part of the image display mechanism.
- the information processing apparatus may further include an attachment state control unit capable of controlling an attachment state of the HMD on the basis of the acquired eye information.
- the information processing apparatus may further include a notification unit that notifies the user of prescribed information on the basis of the acquired eye information.
- the eye information may include a left-eye image obtained by shooting a left eye of the user and a right-eye image obtained by shooting a right eye of the user.
- the movement unit may move at least a part of the image display mechanism on the basis of the left-eye image or the right-eye image.
- An information processing method is an information processing method performed by a computer system and includes acquiring eye information regarding an eye of a user. On the basis of the acquired eye information, at least a part of an image display mechanism that emits and guides image light to the eye of the user is moved.
- a program according to an embodiment of the present technology causes a computer system to perform the following steps.
- the steps include: a step of acquiring eye information regarding an eye of a user;
- a step of moving on the basis of the acquired eye information, at least a part of an image display mechanism that emits and guides image light to the eye of the user.
- the present technology makes it possible to realize high-quality viewing experience for users. Note that the effect described here should not be interpreted in a limited way, and any effect described in the present specification may be produced.
- FIGS. 1A and 1B are views showing a configuration example of an HMD (Head Mount Display) that is an image display apparatus according to an embodiment of the present technology.
- HMD Head Mount Display
- FIG. 2 is a schematic view for describing a configuration example of the inside of a display unit shown in FIG. 1B .
- FIG. 3 is a block diagram showing a functional configuration example of the HMD according to the present embodiment.
- FIG. 4 is a schematic view for describing an example of eye information regarding the eyes of a user.
- FIG. 5 is a flowchart showing the outline of a basic operation performed by the HMD.
- FIG. 6 is a flowchart showing a processing example performed when the use of the HMD is started.
- FIG. 7 is a flowchart showing a setting example of an initial position.
- FIGS. 8A to 8C are schematic views for describing the setting example of the initial position shown in FIG. 7 .
- FIG. 9 is a flowchart showing an operation example of the HMD after the setting of the initial position.
- FIGS. 10A to 10C are schematic views showing movement examples of a lens system according to a deviation amount.
- FIG. 11 is a flowchart showing the outline of other processing capable of being performed by the HMD.
- FIG. 12 is a perspective view showing the appearance of an HMD according to another embodiment.
- FIGS. 1A and 1B are views showing a configuration example of an HMD (Head Mount Display) that is an image display apparatus according to an embodiment of the present technology.
- FIG. 1A is a perspective view schematically showing the appearance of an HMD 100
- FIG. 1B is a perspective views schematically showing a state in which the HMD 100 is disassembled.
- FIG. 2 is a schematic view for describing a configuration example of the inside of a display unit 13 shown in FIG. 1B .
- HMD 100 functions also as an information processing apparatus according to the present technology.
- the HMD 100 has a base member unit 10 , an attachment band unit 11 , a headphone unit 12 , the display unit 13 , a cover unit 14 , and an imaging mechanism 15 .
- the base member unit 10 is a member arranged in front of left and right eyes 1 ( 1 a and 1 b ) of a user and provided with a front support unit 16 that is brought into contact with the front of the user.
- the attachment band unit 11 is attached to the forehead of the user. As shown in FIGS. 1A and 1B , the attachment band unit 11 has a temporal band 17 and a parietal band 18 .
- the temporal band 17 is connected to the base member unit 10 and attached so as to surround the head of the user from the temporal part to the back part of the head.
- the parietal band 18 is connected to the temporal band 17 and attached so as to surround the head of the user from the temporal part to the parietal part.
- a band adjustment mechanism 25 capable of automatically adjusting the retention force of each of the temporal band 17 and the parietal band 18 is provided.
- the band adjustment mechanism 25 By the operation of the band adjustment mechanism 25 , the length of the temporal band 17 protruding outward from the base member unit 10 is changed. As a result, the retention force for retaining the head is changed. Further, by the operation of the band adjustment mechanism 25 , the length of the parietal band 18 is changed. As a result, the retention force for retaining the head is changed. Note that the angle of the parietal band 18 with respect to the temporal band 17 may be changed.
- the configuration of the band adjustment mechanism 25 is not limited and includes, for example, an arbitrary actuator mechanism using a motor, a piezoelectric element, a wire, a hinge, a solenoid, a shape-memory alloy (SMA), or the like.
- the headphone unit 12 is connected to the base member unit 10 and arranged so as to cover the left and right ears of the user.
- the headphone unit 12 is provided with left and right speakers.
- the position of the headphone unit 12 is manually or automatically controllable. A configuration for controlling the position is not limited, and an arbitrary configuration may be employed.
- the display unit 13 is inserted in the base member unit 10 and arranged in front of the eyes 1 of the user. As shown in FIG. 2 , the display unit 13 includes an image display mechanism 20 that emits and guides image light to the eyes 1 of the user.
- the image display mechanism 20 has a display 21 that emits the image light and a lens system 22 that guides the image light emitted from the display 21 to the eyes 1 of the user.
- a display 21 an arbitrary display device using a liquid crystal, an EL (Electro-Luminescence), or the like may be, for example, used.
- EL Electro-Luminescence
- the image light is emitted from the display so as to be diffused. That is, the image light is emitted as diffused light.
- the image light emitted so as to be diffused is appropriately guided to the eyes 1 of the user by the lens system 22 , an image constituted by the image light is made visually recognizable in a focused state.
- the lens system 22 has a left-eye lens system 22 a arranged in front of a left eye 1 a and a right-eye lens system 22 b arranged in front of a right eye 1 b .
- the configuration of each lens system 22 is arbitrarily and is not limited to a case in which one lens is arranged.
- a plurality of various lenses such as Fresnel lenses or a plurality of arbitrary optical members such as optical filters may be arranged.
- a drive mechanism 26 capable of moving each of the display 21 , the left-eye lens system 22 a, and the right-eye lens system 22 b is provided in the present embodiment.
- the drive mechanism 26 it is possible to arbitrarily change the position or attitude (direction) of the display 21 , the position or attitude (direction) of the left-eye lens system 22 a, and the position or attitude (direction) of the right-eye lens system 22 b independently of each other.
- a vertical direction, a horizontal direction, and a depth direction are an X direction, a Y direction, and a Z direction, respectively, when the display 21 is seen from the eyes 1 of the user.
- the drive mechanism 26 is capable of performing a drive operation such as parallel movements along respective axis directions and rotation based on respective axes with respect to the display 21 or each lens system 22 .
- the specific configuration of the drive mechanism 26 is not limited, and an arbitrary actuator mechanism as described above may be used.
- each of the display 21 , the left-eye lens system 22 a, and the right-eye lens system 22 b corresponds to a part of the image display mechanism 20 . That is, the movement of each of the display 21 , the left-eye lens system 22 a , and the right-eye lens system 22 b corresponds to the movement of at least a part of the image display mechanism 20 .
- the imaging mechanism 15 has a left-eye camera 23 a that shoots the left eye 1 a of the user and a right-eye camera 23 b that shoots the right eye 1 b of the user.
- Each of the left-eye camera 23 a and the right-eye camera 23 b is installed at a prescribed position of the HMD 100 , specifically, at a prescribed position of the base member unit 10 . Accordingly, when the relative position of the base member unit 10 with respect to the eyes 1 of the user changes, the relative positons of the left-eye camera 23 a and the right-eye camera 23 b with respect to the eyes 1 of the user also change.
- the left-eye camera 23 a and the right-eye camera 23 b are arranged at positions at which the left eye 1 a and the right eye 1 b of the user are capable of being shot directly, that is, without the interposition of the left-eye lens system 22 a and the right-eye lens system 22 b.
- the left-eye camera 23 a and the right-eye camera 23 b are arranged so as to be directed obliquely from, for example, a downward side.
- the left and right cameras may be arranged so as to be directed to the eyes 1 of the user from other directions.
- CMOS Complementary Metal-Oxide Semiconductor
- CCD Charge Coupled Device
- the cover unit 14 shown in FIGS. 1A and 1B is attached to the base member unit 10 and configured so as to cover the display unit 13 .
- the HMD 100 thus configured functions as an immersive head mount display configured so as to cover the visual field of the user. By wearing the HMD 100 , the user is allowed to experience, for example, virtual reality (VR) or the like.
- VR virtual reality
- FIG. 3 is a block diagram showing a functional configuration example of the HMD 100 according to the present embodiment.
- the HMD 100 further has an operation button 27 , a communication unit 28 , a connector 29 , a storage unit 30 , a sensor unit 31 , and a controller 32 .
- the operation button 27 is provided at, for example, a prescribed position of the base member unit 10 .
- an operation to turn on/off a power supply or operations related to various functions of the HMD 100 such as a function related to an image display or a sound output and a network communication function may be performed.
- the communication unit 28 is a module for performing network communication, near field communication, or the like with other devices.
- a wireless LAN module such as WiFi or a communication module such as BluetoothTM is, for example, provided.
- the connector 29 is a terminal for connection with other devices.
- a terminal such as a USB (Universal Serial Bus) and a HDMITM (High-Definition Multimedia Interface) is, for example, provided. Further, during charging, the connector 29 is connected to the charging terminal of a charging dock (cradle) to be charged.
- a charging dock cradle
- the sensor unit 31 includes a pressure sensor 34 , a proximity sensor 35 , a nine-axis sensor 36 , a GPS 37 , and a biosensor 38 .
- the pressure sensor 34 is provided at, for example, prescribed positions of the temporal band 17 and the parietal band 18 shown in FIGS. 1A and 1B . By the pressure sensor 34 , it is possible to measure pressure applied from the temporal band 17 and the parietal band 18 to the head.
- the proximity sensor 35 is provided at a prescribed position on the inner periphery side of the attachment band unit 11 , and its detection result is used to determine the attachment/detachment of the HMD 100 .
- the nine-axis sensor 36 includes a three-axis acceleration sensor, a three-axis gyro sensor, and a three-axis compass sensor. By the nine-axis sensor 36 , it is possible to detect acceleration, angular speed, and directions in three axes.
- the GPS 37 acquires information regarding the current place of the HMD 100 . These sensors are provided at, for example, prescribed positions of the base member unit 10 . Of course, these sensors may be provided at other positions.
- the biosensor 38 acquires biological information regarding the user.
- a brain wave sensor, a myoelectric sensor, a pulse sensor, a sweat sensor, a temperature sensor, a blood flow sensor, a body motion sensor, or the like is, for example, provided. These sensors are provided at prescribed positions of the HMD 100 so that their detection terminal parts are in contact with prescribed positions of the body.
- the brain wave sensor is provided to be capable of being in contact with a prescribed position of the head.
- the pulse sensor is provided at a position at which the pulse sensor is capable of being in contact with a blood vessel of the neck.
- the types of sensors provided as the sensor unit 31 are not limited, and arbitrary sensors may be provided.
- a temperature sensor, a humidity sensor, or the like capable of measuring temperature, humidity, or the like of an environment in which the HMD 100 is used may be, for example, provided.
- the storage unit 30 is a non-volatile storage device, and a HDD (Hard Disk Drive) or the like is, for example, used as such.
- the storage unit 30 stores a control program 40 for controlling the entire operation of the HMD 100 .
- the storage unit 30 stores a mechanism drive table 41 .
- the mechanism drive table 41 is a table referred to when at least a part of the image display mechanism 20 is moved on the basis of eye information regarding the user, and will be described in detail later.
- a method for installing the control program 40 and the mechanism drive table 41 in the HMD 100 is not limited.
- the controller 32 controls the operations of the respective blocks of the HMD 100 .
- the controller 32 has, for example, hardware configurations for a computer such as a CPU and a memory (RAM, ROM). Various processing is performed when the CPU loads the control program 40 stored in the storage unit 30 into the RAM and performs the same.
- a PLD Programmable Logic Device
- FPGA Field Programmable Gate Array
- ASIC Application Specific Integrated Circuit
- an image analysis unit 43 a deviation calculation unit 44 , a mechanism drive unit 45 , a display control unit 46 , a band adjustment unit 47 , a state analysis unit 48 , a reliability determination unit 49 , and a notification unit 50 are realized when the CPU of the controller 32 performs a program according to the present embodiment. Further, an information processing method according to the present embodiment is performed by these function blocks. Note that dedicated hardware such as an IC (Integrated Circuit) may be appropriately used to realize the respective function blocks.
- IC Integrated Circuit
- the image analysis unit 43 analyzes eye images of the user shot by the imaging mechanism 15 . That is, the image analysis unit 43 acquires a left-eye image shot by the left-eye camera 23 a and a right-eye image shot by the right-eye camera 23 b and analyzes these images. All the left-eye image and the right-eye image input to the image analysis unit 43 and various characteristic parameters detected as a result of the analysis of the image analysis unit 43 are included in eye information regarding the eyes 1 of the user in the present embodiment. On the basis of these eye information items, it is possible to dynamically estimate, for example, the positions of the eyes 1 of the user. In the present embodiment, the image analysis unit 43 functions as an acquisition unit.
- FIG. 4 is a schematic view for describing an example of eye information regarding the eyes 1 of the user.
- eye information it is possible to detect, for example, the shape, size, position, inclination, or the like of an eye 1 P of the user in an eye image 55 .
- an eye region 56 of the user is detected, and its shape, size, position, and inclination are detected.
- a size t of a long axis 57 of the eye 1 P of the user may be used as the size of the eye region 56 of the user.
- the inclination of the long axis 57 of the eye 1 P of the user with respect to a horizontal direction may be used as the inclination of the eve region 56 of the user.
- the long axis 57 of the eye 1 P or the user is an axis connecting the position of the right end and the position of the left end of each of the eyes 1 of the user to each other in a substantially horizontal direction when the vertical direction of the face (for example, a direction from the parietal part to the jaw) of the user becomes a substantially vertical direction and the horizontal direction of the face (for example, a direction in which positions at the same height of both ears are connected to each other) becomes a substantially horizontal direction.
- information such as information as to whether the entire eye region 56 of the user is included in the eye image 55 and information as to which of the vertical and horizontal directions is a direction in which the eye region 56 is missing is also included in the eye information regarding the eyes 1 of the user.
- arbitrary information such as a visual line direction, a pupil size, an iris pattern, and an eyelash shape may be detected as the eye information.
- a method for analyzing an eye image to detect characteristic parameters is not limited.
- An arbitrary segmentation technology or an arbitrary image analysis technology may be used.
- a machine learning algorithm using a DNN such as a RNN (Recurrent Neural Network), a CNN (Convolutional Neural Network), and a MLP (Multilayer Perceptron) may be used.
- DNN Deep Neural Network
- RNN Recurrent Neural Network
- CNN Convolutional Neural Network
- MLP Multilayer Perceptron
- the deviation calculation unit 44 calculates a deviation amount from the initial state of the eyes 1 of the user on the basis of eye information regarding the user acquired by the image analysis unit 43 .
- the initial state of the eyes 1 of the user and a deviation from the initial state will be described later.
- the mechanism drive unit 45 outputs a control signal to the drive mechanism 26 .
- each of the display 21 , the left-eye lens system 22 a , and the right-eye lens system 22 b is moved.
- the deviation calculation unit 44 and the mechanism drive unit 45 function as a movement unit.
- the display control unit 46 controls an image display by the image display mechanism 20 .
- arbitrary image processing and display control such as correcting a displayed image and moving the display position of a displayed image are, for example, performed.
- the band adjustment unit 47 outputs a control signal to the band adjustment mechanism 25 . On the basis of the control signal output from the band adjustment unit 47 , the retention force of each of the temporal band 17 and the parietal band 18 is adjusted.
- the band adjustment unit 47 functions as an attachment state control unit capable of controlling the attachment state of the HMD 100 .
- the state analysis unit 48 acquires various state information items on the states of the user or the HMD 100 on the basis of detection results from the sensor unit 31 .
- the state information regarding the user information as to whether the user is in a sleeping state or biological information such as body heat, a pulse rate, a brain wave state, muscle movements, a sweat amount, and a concentration state is, for example, acquired.
- information regarding the current place of the user more specifically, information as to whether the user is indoors, outdoors, or is in conference, or the like is acquirable.
- motion information regarding motion done by the user is acquired.
- Information during walking, running, moving by train, driving, or the like is, for example, acquired.
- Information on the types of sports being played or the like is also acquirable.
- information regarding a posture as to whether the user is sitting, standing, stooping, turning sideways, or turning up is acquired.
- These state information items on the user are acquirable by, for example, an arbitrary motion analysis technology such as a motion analysis using parameters obtained by machine learning or the like.
- various information items such as an active function, an operation mode, the attachment/detachment of the HMD 100 to/from the user, the attachment position of the HMD 100 with respect to the user, a battery remaining amount, connection with a charging dock, and apparatus temperature are, for example, acquired.
- various information items such as temperature, humidity, a current place, weather, date and time are also acquirable. Sensors, devices, or the like for acquiring these information items only have to be appropriately provided in the HMD 100 .
- a method for analyzing state is not limited, and a machine learning algorithm may be, for example, used.
- the state analysis unit 48 functions as a state acquisition unit.
- the reliability determination unit 49 determines the reliability of a detection result by the sensor unit 31 . The determination will be described later.
- the notification unit 50 notifies the user of various information items.
- the notification of prescribed information is realized by, for example, an arbitrary method such as displaying an image or text on the display 21 , outputting a sound from the headphone unit 12 , shining light on (blinking) a prescribed position at the outer part of the base member unit 10 or the cover unit 14 , and vibrating the base member unit 10 or the like with an arbitrary vibration mechanism. It is possible to make, for example, a notification of arbitrary information such as information regarding the use of the HMD 100 , information regarding content, and information regarding a detection result by the sensor unit 31 .
- FIG. 5 is a flowchart showing the outline of a basic operation performed by the HMD 100 .
- eye information regarding the eyes 1 a and 1 b of the user is acquired by the image analysis unit 43 (step 201 ).
- a change in viewing state and the notification of prescribed information are performed (step 102 ).
- the change in viewing state includes, for example, arbitrary processing to change a state in which content is viewed, such as moving the left and right lens systems 22 a and 22 b, moving the display 21 , moving the display position of as image, and adjusting the attachment band unit 11 . Further, the notification itself of a prescribed alert or the like is also included in the change in viewing state.
- FIG. 6 is a flowchart showing a processing example performed when the use of the HMD 100 is started.
- a determination is first made as to whether the HMD 100 has been attached to the head of the user (step 201 ). The determination is made by the state analysis unit 48 shown in FIG. 3 on the basis of a detection result by the proximity sensor 24 .
- the user is notified of the fact by the notification unit 50 shown in FIG. 3 (step 202 ).
- a notification method is not limited. For example, a sound may be output from the headphone unit 12 , or prescribed colored light may be output to the outside of the base member unit 10 or the cover unit 14 . Alternatively, processing to make the entire display 21 blink may be performed. Of course, the notification of the fact to the user may be performed by the vibration of the base member unit 10 or the like based on a vibration mechanism.
- step 203 the search of the eye regions of the user is started. Specifically, the detection of the eye region 56 of the left eye 1 a and the eye region 56 of the right eye 1 b is performed by the image analysis unit 43 shown in FIG. 3 . Then, a determination is made as to whether the left and right eye regions 56 are not completely detectable (step 204 ).
- the determination is performed by the state analysis unit 48 on the basis of, for example, a detection result by the nine-axis sensor.
- step 205 it is also possible to provide a proximity camera for finger print shooting or the like in the sensor unit 31 and shoot the skin or the like of the user with the proximity camera to measure the movement direction and distance of a skin pattern. For example, it is also possible to estimate the movement distance on the basis of a cumulative movement distance since calibration. Further, the determination as to whether the HMD 100 is physically moving may be made on the basis of an image shot by a camera (including IoT or the like) provided at a distant place. For example, face detection is performed with respect to the face wearing the HMD 100 to detect the position of a LED marker attached to the HMD. It is possible to determine whether the HMD has physically moved from the result of the face detection and the positional relationship of the LED marker of the HMD.
- step 206 the processing is on standby until the HMD 100 stops (step 206 ). That is, the processing is on standby until the state of the user or the HMD 100 is settled. After being on standby for a prescribed time, the processing returns to step 203 .
- the HMD 100 is not moving (No in step 205 )
- the user is notified of an instruction to change an attachment state so that the eyes 1 of the user are capable of being shot by the left-eye camera 23 a and the right-eye camera 23 b (step 207 ). For example, announcement or the like urging the user to manually correct the position of the HMD 100 is output. Then, the processing returns to step 203 .
- step 208 a determination is made as to whether all the left and right eye regions 56 are detectable.
- a direction in which the HMD 100 is deviated and the degree to which the HMD 100 is deviated are estimated on the basis of the missing region (step 209 ).
- a control signal is generated on the basis of the deviation amount estimated in step 209 and output to the band adjustment mechanism 25 by the band adjustment unit 47 . That is, the control signal is output to the band adjustment mechanism 25 so as to create an attachment state in which the eyes 1 of the user are capable of being shot by the left-eye camera 23 a and the right-eye camera 23 b. Then, the lengths or positions of the temporal band 17 and the parietal band 18 of the attachment band unit 11 are changed by the band adjustment mechanism 25 (step 210 ).
- the length or the like of the temporal band 17 is adjusted.
- the length or the like of the parietal band 18 is adjusted.
- the adjustment is not limited to such processing. Note that the band on the temporal part side and the band on the parietal part side shown in FIG. 6 correspond to the temporal band 17 and the parietal band 18 , respectively.
- step 208 When all the left and right eye regions 56 are detectable (Yes in step 208 ), the setting of an initial position is performed (step 211 ).
- FIG. 7 is a flowchart showing a setting example of an initial position.
- FIGS. 8A to 8C are schematic views for describing the setting example of the initial position shown in FIG. 7 .
- the setting of the initial position refers to the setting of the states of the left and right eyes 1 of the user with respect to the image display mechanism 20 to an initial state in which content is capable of being appropriately viewed when the use of the HMD 100 is started.
- the initial state in which the content is capable of being appropriately viewed corresponds to “a case in which the eyes of the user are put in a prescribed state.” Further, the initial state refers also to a prescribed viewing state.
- a reference image 60 is displayed at a prescribed position of the display 21 (step 301 ). As shown in, for example, FIG. 8A , a point image colored in green or the like is displayed as the reference image 60 at a substantially central area of the display 21 . Then, left and right eye images 55 shot in this state are analyzed (step 302 ).
- This state corresponds to a state in which an image of the diffused and emitted image light of the reference image 60 is formed minimally and a state in which the left and right eyes 1 of the user are focused with respect to the reference image 60 displayed on the display 21 . That is, this state is a state in which the above initial state is realized.
- the left and right lens systems 22 a and 22 b are moved (step 304 ). That is, a control signal is generated and output to the drive mechanism 26 by the mechanism drive unit 45 . On the basis of the input control signal, the drive mechanism 26 moves the left and right lens systems 22 a and 22 b.
- a reference image 60 PA When a reference image 60 PA is deviated rightward with respect to the center of the pupil and expanded in size as shown in, for example, FIG. 8C , it is determined that focus is not achieved. Then, the lens system 22 is appropriately moved on the basis of XYZ axes so as to create a focused state (a state shown in FIG. 8B ). Similarly, when a reference image 60 PB is deviated leftward with respect to the center of the pupil and expanded in size as shown in FIG. 8C , the lens system 22 is appropriately moved on the basis of the XYZ axes so as to create a focused state.
- the prescribed state that is, the initial state
- the prescribed state that is, the initial state
- prescribed allowable ranges are set with respect to the position and the size and the size and position of the reference image 60 P are included in the allowable ranges, it may be determined that the prescribed state (that is, the initial state) in step 303 is realized.
- the size, position, and inclination of each of left and right eye regions 56 are stored in the storage unit 30 (step 305 ).
- the size (for example, the number of pixels) of the long axis of the eye region 56 shown in FIG. 4 is stored as the size of each of the eye regions 56 .
- the position and inclination (for example, coordinate values of pixels) of the long axis 57 are stored as the position and inclination of each of the eye regions 56 .
- the sizes and positions of the entire eye regions 56 may be stored as they are.
- eye information acquired when the initial state is realized that is, eye images shot when the initial state is realized or various characteristic parameters calculated from the eye images correspond to reference eye information. Further, the eye images shot when the initial state is realized correspond to reference eye images.
- the shape, size, color, or the like of the reference image 60 is not limited and may be arbitrarily designed.
- an image of a building, a vehicle, or a balloon, a prescribed logo mark, or the like may be displayed as the reference image 60 .
- a condition which is set with respect to the reference image 60 P in the eye images 55 and on which the realization of the state is determined is not also limited. That is, a condition on the size or position of the reference image 60 P may be arbitrarily set, and a parameter or the like (such as a brightness value and a blurring degree) other than the size or position may be employed as the condition on the realization of the initial state.
- the reference image 60 having a simple shape like the point image shown in FIG. 8A a determination is made as to whether the initial state is realized on the basis of the size of the reference image 60 P in the eye images 55 .
- a reference image 60 having a complicated shape such as a balloon it is possible to determine whether the initial state is realized on the basis of not only the size of the reference image 60 P in the eye images 55 but also focus (a blurring degree).
- a condition for determining the realization of the initial state is not limited to the above example.
- the display 21 may be moved to realize the initial state.
- the display position of the reference image 60 may be moved.
- a content image or the like only has to be displayed at a position corresponding to the position of the reference image 60 obtained when the initial state is realized.
- the setting of the initial position refers to processing to set focus or a display position before content is viewed.
- FIG. 9 is a flowchart showing an operation example of the HMD 100 after the setting of the initial position.
- Steps 401 to 410 are substantially the same as steps 201 to 210 shown in FIG. 6 .
- the processing proceeds from Yes in step 408 to step 411 in most cases.
- the user is highly likely to walk or run in VR (Virtual Reality) during the viewing of content.
- the HMD 100 could be deviated.
- step 411 a determination is made by the deviation calculation unit 44 shown in FIG. 3 as to whether there is any change in the inclinations of the long axes 57 of the eye regions 56 in comparison with the inclinations of the long axes 57 in the initial state.
- the display position of an image (a content image) displayed on the display 21 is moved by the display control unit 46 .
- a display coordinate axis with respect to the display 21 is rotated at a rotation angle corresponding to the change in the inclinations of the long axes 57 (step 412 ).
- the processing returns to step 403 without changing the viewing state.
- a control signal corresponding to the change is generated and output to the drive mechanism 26 by the mechanism drive unit 45 .
- the respective positions or angles of the display 21 , the left-eye lens system 22 a, and the right-eye lens system 22 b are adjusted. (step 414 ). Specifically, the respective positions or angles of the display 21 , the left-eye lens system 22 a, and the right-eye lens system 22 b are adjusted so that the state of the eyes 1 of the user with respect to the image display mechanism 20 comes close to the initial state.
- the mechanism drive table 41 stored in the storage unit 30 is referred to.
- the mechanism drive table 41 is a table in which deviation amounts calculated by the deviation calculation unit 44 and movement amounts of the display 21 , the left-eye lens system 22 a, and the right-eye lens system 22 b are associated with each other.
- the mechanism drive table 41 is created in advance by calibration or the like and stored in the storage unit 30 .
- a two-dimensional lookup table showing the relationships between deviation amounts and the combinations of movement amounts of the display and the lens system with respect to each of the left and right eyes is created.
- a three-dimensional lookup table showing the relationships between deviation amounts, movement amounts of the display, and movement amounts of the lens system may be created.
- a condition for moving the display, a condition for moving the lens system, or the like is set, and a device to be moved may be appropriately determined. In this case, a plurality of table information items used according to a determined result may be created.
- movement of the display 21 or the lens system 22 may be performed without using table information.
- movement amounts may be calculated by computation as occasion demands on the basis of deviation amounts to appropriately control the operation of the drive mechanism 26 .
- FIGS. 10A to 10C are schematic views showing movement examples of the lens system 22 according to a deviation amount.
- FIG. 10A it is assumed that the size of an eye region 56 in an eye image 55 becomes smaller than that of the eye region 56 in the initial state.
- the base member unit 10 in which the left-eye camera 23 a and the right-eye camera 23 b are installed is moved in a direction away from the eyes 1 of the user.
- the left and right lens systems 22 are caused to come close to the eyes of the user.
- the base member unit 10 in which the left-eye camera 23 a and the right-eye camera 23 b are installed is moved in a direction close to the eyes 1 of the user.
- the left and right lens systems 22 are caused to get away from the eyes of the user.
- the base member unit 10 in which the left-eye camera 23 a and the right-eye camera 23 b are installed is moved in the XY plane.
- the left and right lens systems 22 are caused to move in a direction in which the deviation is corrected.
- At least a part of the image display mechanism 20 is moved on the basis of the differences between reference eye images shot in the initial state and eye images shot during the viewing of content.
- a method for moving at least a part of the image display mechanism 20 according to a deviation is not limited to the methods shown in the examples of FIGS. 10A to 10C and an arbitrary drive method (correction method) may be performed.
- a change in viewing state and a prescribed notification are performed on the basis of eye information regarding the eyes 1 of the user in the HMD 100 according to the present embodiment. That is, the movement of at least a part of the image display mechanism 20 , the adjustment of the attachment band unit 11 , and the notification of prescribed information are performed on the basis of eye information.
- the movement of at least a part of the image display mechanism 20 , the adjustment of the attachment band unit 11 , and the notification of prescribed information are performed on the basis of eye information.
- the movement or the like of the lens system 22 is automatically performed on the basis of eye information regarding the eyes 1 of the user even during the viewing of content.
- Examples of a notification method for the user will be described. For example, when it is determined that a deviation is beyond an automatic adjustment, the user is notified of the occurrence of the deviation and urged to make an adjustment. Thus, it is possible to prevent an unintended attachment state from being kept. Specifically, the display of an image or text on the display 21 , the output of a sound from the headphone unit 12 , the vibration of the base member unit 10 or the like with a vibration mechanism, or the like is performed.
- a notification or alert is displayed using, for example, text at a size at which the text is seeable even in a blurred state or an image, a symbol, a mark, or the like that is recognizable even in a blurred state.
- a notification or alert is performed by a sound, vibration described above, or the like.
- a short notification that is not so emphasized is, for example, performed.
- the notification is emphasized stepwise or other notification methods are performed. For example, when the user does not notice a sound, a vibration function is used in combination with the sound.
- the awakened degree of the user is measured, and a notification method is appropriately changed on the basis of a measurement result. For example, processing such as emphasizing the notification to make the user awakened, cancelling the content contrary, and turning off the power supply of the HMD 100 itself is assumed.
- the HMD 100 When it is determined that the HMD 100 is deviated horizontally, it is also possible to give guidance to the user through content to move the head instantaneously to correct the deviation. For example, a sound is used and quickly transmitted from the right ear to the left ear. Alternatively, an image that quickly moves horizontally is displayed on the display 21 . The user is caused to feel the movement of any object on the screen through this content and then move the head instantaneously to follow the object. Thus, it is also possible to give guidance to the user to move the head. In addition, it is also possible to give guidance to the user to move the head instantaneously using both a sound and the movement of an image.
- a prescribed button operation the movement of the head (an increase in acceleration), or a sound input from the user as a trigger, it is also possible to loosen one side and tighten the other side of the attachment band unit 11 to move and restore the entire HMD 100 itself to a normal position.
- FIG. 11 is a flowchart showing the outline of other processing capable of being performed by the HMD 100 .
- step 501 information regarding the state of the user, the details of displayed content, and a viewing time is acquired.
- step 502 a change in viewing state and a prescribed notification are performed on the basis of these information items.
- processing or the like to move at least a part of the image display mechanism 20 may be performed on the basis of information different from eye information.
- these information items and eye information may be combined together to perform a change in viewing state and a prescribed notification.
- the lens system 22 or the like is moved so that the state of the eyes 1 of the user comes close to the initial state when a deviation occurs.
- the movement of the display 21 or the lens system 22 may be performed so that the state of the eyes 1 of the user with respect to the image display mechanism 20 is put in another state different from the initial state. That is, the viewing state may be changed to a state in which focus is not achieved intentionally, or may be changed so as to make the display position of an image deviated.
- the blurring of a view is performed in, for example, a case in which the user is covered with dust from a forward-traveling vehicle, a case in which the user dives into water, or the like in a racing game. Further, it is also possible to adjust the blurred degree of a view according to the attachment/detachment of an appropriate item (for example, glasses or the like) inside VR.
- an appropriate item for example, glasses or the like
- the blurred degree of a view according to the physical condition or awakened state of an avatar in content.
- the blurring of a view or the like is performed in a case in which the user is damaged or is in a critical condition in, for example, a battle game, a shooting game, or the like. It is also possible to make only the right eye blurred when the right eye of the user gets hit. Further, it is also possible to make a view blurred when the user feels strong sleepy.
- the viewing state is changed according to the details of content, whereby it is possible to enhance a sense of realism of the content and realize high-quality viewing experience.
- the lens system 22 or the like is moved on the basis of eye information so as to realize the visual field.
- the lens system 22 or the like is moved so as to make a focal distance get close to the user when the user selects the visual field of a foresighted person and moved so as to make the local distance get away from the user when the user selects the visual field of a farsighted person.
- the lens system 22 or the like is appropriately moved according to the distance of an object. Note that it is also possible to automatically adjust visibility during the attachment of the HMD 100 by moving the lens system 22 or the like.
- the viewing state is changed on the basis of the viewing time of content. For example, when content is used over a specified time, it is possible to make a view blurred to cause the user to be aware of fatigue and have a rest. For example, as the used time of content comes close to a content setting time (90 minutes or the like), the user is caused to be gradually put in a near-sighted state (caused to have a shorter focal distance) and feel eyestrain. Such processing corresponds to the power mode of a TV. Further, on this occasion, it is also possible to determine whether the user gets tired on the basis of biological information (brain waves, pulses, blood flows, or the like). Further, it is also possible to display a message such as “enter an eyestrain mode” at a corner of the display 21 .
- the HMD 100 in, for example, a case in which the HMD 100 is used over a specified time, a case in which it is determined that the user gets tired on the basis of biological information, or the like, it is also possible to offer content such as an image or music for relaxing eyestrain to the user.
- the viewing state is changed on the basis of state information regarding the state of the user.
- the processing is performed according to an awakened degree described above or performed according to biological information.
- a display state may be changed using a monitoring result. For example, when it is determined that the user gets tired on the basis of biological information, the user is caused to be gradually put in a near-sighted state (caused to have a shorter focal distance) and feel eyestrain. Further, it is also possible to display a message such as “enter an eyestrain mode” at a corner of the display 21 . Further, it is also possible to offer content such as an image or music for relaxing eyestrain to the user.
- the reliability determination unit 49 shown in FIG. 3 will be described.
- the reliability determination unit 49 determines the reliability of detection result by the biosensor 38 on the basis of a movement amount (correction amount) from the initial state in the display 21 or the left and right lens systems 22 . For example, it is determined that the HMD 100 is deviated from the head as movement amounts of the lens system 22 are greater. That is, with the assumption that the detection terminal part of a brain wave sensor, a myoelectric sensor, or the like is not highly likely to be substantially in contact with a prescribed position of the body, the reliability of a detected signal is set to be low. Thus, it is possible to effectively use a detection result by the biosensor 38 .
- the HMD 100 it is possible to detect the swinging or the like of the HMD 100 with the nine-axis sensor 36 but is not possible to determine whether the detection terminal part of the biosensor 38 is substantially in contact with the skin. In the present embodiment, it is possible to determine whether a detection terminal part is substantially in contact with the skin and determine the reliability of a detection result on the basis of a movement amount by the mechanism drive unit 145 with respect to at least a part of the image display mechanism 20 .
- the reliability of a detection result by the biosensor 38 may be determined on the basis of, for example, biological information obtained from the biosensors of other wearable devices in the form of a watch, a neckless, a ring, or the like or determination results about contacting the skin with biosensors determined by other wearable devices.
- the reliability of a detection result by the biosensor 38 may be determined on the basis of biological information or the like acquired by other terminals not classified into wearable devices.
- the used information may be weighted.
- a weighting method is not limited and may be arbitrarily set.
- each of the display 21 , the left-eye lens system 22 a, and the right-eye lens system 22 b is movable as an example of moving at least a part of the image display mechanism 20 .
- the display 21 may be movable, or only the left and right lens systems 22 may be movable.
- other members constituting the image display mechanism 20 may be movable.
- only a part of the image display mechanism 20 may be movable to change a viewing state.
- the attachment band unit 11 When the lens system 22 or the like is frequently moved by the mechanism drive unit 45 , it may be determined that the fastening degree of the attachment band unit 11 is loose. When it is determined that the fastening degree is loose, the attachment band unit 11 is automatically adjusted on the basis of, for example, eye information. Alternatively, a notification about urging the adjustment of the attachment band unit 11 is made.
- FIG. 12 is a perspective view showing the appearance of an HMD according to another embodiment.
- the HMD 200 is a glasses-type apparatus including a see-through display and attached to the head of the user to be used.
- the HMD 200 has a frame 210 , a left lens 222 a, a right lens 222 b, a left-eye display 221 a , a right-eye display 221 b, a left--eye camera 223 a, and a right-eye camera 223 b.
- a controller, a sensor unit, or the like substantially the same as that shown in FIG. 3 is constituted inside or at a prescribed position of the frame 210 .
- the left lens 222 a and the right lens 222 b are arranged in front of the left eye and right eye of the user, respectively.
- the left-eye display 221 a and the right-eye display 221 b are provided in the left lens 222 a and the right lens 222 b , respectively, so as to cover the visual field of the user.
- the left-eye display 221 a and the right-eye display 221 b are see-through displays and have a left-eye image and a right-eye image or the like displayed thereon, respectively.
- the user wearing the HMD 200 is allowed to visually recognize images displayed on the respective displays while visually recognizing an actual scene.
- the user is allowed to experience AR (Augmented Reality) or the like.
- dimmer elements may be provided on the outside (a side opposite to the eyes of the user) of the left-eye display 221 a and the right-eye display 221 b, respectively.
- the dimmer elements are elements capable of adjusting an amount of light passing through the elements. With the provision of the dimmer elements, it is possible to control, for example, an actual scene passing through the respective displays and visually recognized by the user and cause the user to visually recognize images displayed on the respective displays in an emphasized manner. Thus, the user is allowed to experience VR (Virtual Reality) or the like.
- the left-eye display 221 a and the right-eye display 221 b see-through organic EL displays, LCDs (Liquid Crystal Displays), or the like are, for example, used. Further, as the dimmer elements, dimmer glasses, dimmer sheets, liquid crystal shutters, or the like capable of electrically controlling transmittance are, for example, used.
- an image display mechanism is realized by the left lens 222 a, the right lens 222 b, the left-eye display 221 a , and the right-eye display 221 b.
- the left-eye camera 223 a and the right-eye camera 223 b are provided at arbitrary positions at which the left eye and right eye of the user are capable of being shot. On the basis of, for example, images of the left eye and the right eye shot by the left-eye camera 223 a and the right-eye camera 223 b, eye information regarding the eyes of the user is acquired.
- each of the left lens 222 a and the right lens 222 b is configured to be movable with respect to the frame 210 and moved by a drive mechanism. Further, the frame 210 itself is also configured to be movable and capable of changing its retention force. By appropriately changing the position or inclination of each of the left lens 222 a and the right lens 222 b on the basis of eye information, it is possible to realize high-quality viewing experience like the above embodiments.
- the present technology is also applicable to a see-through HMD such as AR glasses.
- a see-through HMD such as AR glasses.
- the display positions of content images displayed on see-through displays on the basis of eye information during the experience of AR, it is also possible to correct the deviation between the real world and the content images with high accuracy. That is, in the AR, it is possible to change coordinate axes in a virtual world overlapping a real scene with high accuracy.
- the present technology is effective mainly for correcting the deviation of focus with respect to content in VR and is effective mainly for correcting the deviation of the position of content in AR.
- the present technology is not limited such objects.
- Fields to which the present technology is applicable are not limited.
- the present technology is applicable to, for example, arbitrary fields in which information is displayed in the view of the user in operation, such as video games, medical treatment (HMDs for surgeries or the like), motor sports (drivers' helmets or the like), sports including golf or the like, and examinations in factories or the like.
- the above description refers to an HMD that is an image display apparatus.
- the information processing apparatus according to the present technology may be realized by an arbitrary computer that is constituted separately from an HMD and connected to the HMD in a wired or wireless fashion.
- An information processing method according to the present technology may be performed by, for example, a cloud server.
- an HMD and other computers may work together to perform the information processing method according to the present technology.
- the information processing method and a program according to the present technology are capable of being performed not only by a computer system constituted by a single computer but also by a computer system in which a plurality of computers works together to operate.
- a system refers to the aggregate of a plurality of constituents (such as apparatuses and modules (components)) and all the constituents are not necessarily accommodated in the same housing in the present disclosure. Accordingly, both of a plurality of apparatuses that is accommodated in separate housings and connected to each other via a network and one apparatus in which a plurality of modules is accommodated in one housing are systems.
- the execution of the information processing method and the program according to the present technology based on a computer system includes both a case in which the control or the like of, for example, the acquisition of eye information or the movement of at least a part of an image display mechanism is executed by a single computer and a case in which respective processing is executed by different computers. Further, the execution of respective processing by a prescribed computer includes causing other computers to execute a part or all of the processing and acquiring results.
- the information processing method and the program according to the present technology is also applicable to the configuration of cloud computing in which one function is shared and cooperatively processed by a plurality of apparatuses via a network.
- At least two characteristic parts may be combined together. That is, the various characteristic parts described in the respective embodiments may be arbitrarily combined together without being distinguished between the respective embodiments. Further, the various effects described above are given only for an example and should not be interpreted in a limited way. Further, other effects may be produced.
- An information processing apparatus including:
- an acquisition unit that acquires eye information regarding an eye of a user
- a movement unit that moves, on the basis of the acquired eye information, at least a part of an image display mechanism that emits and guides image light to the eye of the user.
- the image display mechanism has a display that emits the image light and a lens system that guides the image light emitted from the display to the eye of the user, and
- the movement unit moves at least one of the display or the lens system.
- the eye information includes an eye image obtained by shooting the eye of the user.
- the eye information includes at least one of a shape, a size, a position, an inclination, or an iris pattern of the eye of the user in the eye image.
- the image display mechanism is provided in an HMD (Head Mount Display) wearable by the user, and
- the eye information includes the eye image obtained by shooting the eye of the user with an imaging mechanism of the HMD.
- a storage unit that stores reference eye information representing the eye information obtained when the eye of the user is put in a prescribed state with respect to the image display mechanism
- the movement unit moves at least a part of the image display mechanism on the basis of the acquired eye information and the reference eye information.
- the prescribed state includes a state in which a reference image displayed by the image display mechanism is arranged at a prescribed size and a prescribed position with respect to the eye of the user inside the eye image.
- the eye information includes an eye image obtained by shooting the eye of the user
- the reference eye information includes a reference eye image obtained by shooting the eye of the user when the eye of the user is put in the prescribed state with respect to the image display mechanism, and
- the movement unit moves at least a part of the image display mechanism on the basis of a difference between the eye image and the reference eye image.
- the movement unit moves at least a part of the image display mechanism such that a state of the eye of the user with respect to the image display mechanism comes close to the prescribed state.
- the movement unit moves at least a part of the image display mechanism such that a state of the eye of the user with respect to the image display mechanism is put in another state different from the prescribed state.
- the display control unit moves, on the basis of the acquired eye information, a display position of an image displayed by the image display mechanism.
- the movement unit moves at least a part of the image display mechanism on the basis of details of content displayed by the image display mechanism.
- the movement unit moves at least a part of the image display mechanism on the basis of a viewing time of the user.
- a state acquisition unit that acquires state information regarding a state of the user, in which
- the movement unit moves at least a part of the image display mechanism on the basis of the acquired state information.
- the image display mechanism is provided in an HMD (Head Mount Display) wearable by the user, and
- the information processing apparatus further includes a determination unit that determines reliability of a detection result by a biosensor on the basis of a movement amount by the movement unit with respect to at least a part of the image display mechanism.
- the image display mechanism is provided in an HMD (Head Mount Display) wearable by the user, and
- the information processing apparatus further includes an attachment state control unit capable of controlling an attachment state of the HMD on the basis of the acquired eye information.
- a notification unit that notifies the user of prescribed information on the basis of the acquired eye information.
- the eye information includes a left-eye image obtained by shooting a left eye of the user and a right-eye image obtained by shooting a right eye of the user, and
- the movement unit moves at least a part of the image display mechanism on the basis of the left-eye image or the right-eye image.
- a step of moving on the basis of the acquired eye information, at least a part of as image display mechanism that emits and guides image light to the eye of the user.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Optics & Photonics (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
Abstract
Description
- The present technology relates to an information processing apparatus, an information processing method, and a program applicable to image display apparatuses such as HMDs (Head Mount Displays).
-
Patent Literature 1 describes a technology to detect user's visual lines to control an image display. For example, user's eyes are shot by cameras installed in an HMD, and a user's visual line direction and an observation point on the HMD are calculated on the basis of shot images. A high-quality partial image is displayed about the observation point, and a low-quality image is displayed on the periphery of the partial image. Thus, it is possible to control an image display on the basis of a region capable of being recognized by a user with high accuracy (paragraphs [0051], [0052], [0060], and [0066],FIGS. 1 and 3 , etc., of the specification of Patent Literature 1). - In Patent Literature 2, a plurality of image data items obtained by shooting a subject at a plurality of different positions is appropriately selected and combined together on the basis of a user's position and a focal position. Combined image data is displayed on a display as a presentation image. Thus, it is possible to display a blurred image (an image having prescribed depth of field) which is a natural (more realistic) image for a human being and in which the periphery of a notice point is blurred (paragraphs [0059], [0074], [0075], and [0098],
FIGS. 10 to 13 , etc., of the specification of Patent Literature 2). - Patent Literature 1: Japanese Patent Application Laid-open No. 2016-191845
- Patent Literature 2: Japanese Patent Application Laid-open No. 2005-227950
- In the future as well, it is expected that the experience of VR (Virtual Reality) or AR (Augmented Reality) using HMDs or the like will become pervasive, and technologies making it possible to realize high-quality viewing experience have been demanded.
- In view of the above circumstances, it is an object of the present technology to provide an information processing apparatus, an information processing method, and a program making it possible to realize high-quality viewing experience for users.
- In order to achieve the above object, an information processing apparatus according to an embodiment of the present technology includes an acquisition unit and a movement unit.
- The acquisition unit acquires eye information regarding an eye of a user.
- The movement unit moves, on the basis of the acquired eye information, at least a part of an image display mechanism that emits and guides image light to the eye of the user.
- In the information processing apparatus, at least a part of the image display mechanism is moved on the basis of eye information regarding an eye of a user. Thus, it is possible to realize high-quality viewing experience for the user.
- The image display mechanism may have a display that emits the image light and a lens system that guides the image light emitted from the display to the eye of the user. In this case, the movement unit may move at least one of the display or the lens system.
- The eye information may include an eye image obtained by shooting the eye of the user.
- The eye information may include at least one of a shape, a size, a position, an inclination, or an iris pattern of the eye of the user in the eye image.
- The image display mechanism may be provided in an HMD (Head Mount Display) wearable by the user. In this case, the eye information may include the eye image obtained by shooting the eye of the user with an imaging mechanism of the HMD.
- The information processing apparatus may further include a storage unit that stores reference eye information representing the eye information obtained when the eye of the user is put in a prescribed state with respect to the image display mechanism. In this case, the movement unit may move at least a part of the image display mechanism on the basis of the acquired eye information and the reference eye information.
- The prescribed state may include a state in which a reference image displayed by the image display mechanism is arranged at a prescribed size and a prescribed position with respect to the eye of the user inside the eye image.
- The reference eye information may include a reference eye image obtained by shooting the eye of the user when the eye of the user is put in the prescribed state with respect to the image display mechanism. In this case, the movement unit may move at least a part of the image display mechanism on the basis of a difference between the eye image and the reference eye image.
- The movement unit may move at least a part of the image display mechanism such that a state of the eye of the user with respect to the image display mechanism comes close to the prescribed state.
- The movement unit may move at least a part of the image display mechanism such that a state of the eye of the user with respect to the image display mechanism is put in another state different from the prescribed state.
- The information processing apparatus may further include a display control unit that controls an image display by the image display mechanism. In this case, the display control unit may move, on the basis of the acquired eye information, a display position of an image displayed by the image display mechanism.
- The movement unit may move at least a part of the image display mechanism on the basis of details of content displayed by the image display mechanism.
- The movement unit may move at least a part of the image display mechanism on the basis of a viewing time of the user.
- The information processing apparatus may further include a state acquisition unit that acquires state information regarding a state of the user. In this case, the movement unit may move at least a part of the image display mechanism on the basis of the acquired state information.
- The information processing apparatus may further include a determination unit that determines reliability of a detection result by a biosensor on the basis of a movement amount by the movement unit with respect to at least a part of the image display mechanism.
- The information processing apparatus may further include an attachment state control unit capable of controlling an attachment state of the HMD on the basis of the acquired eye information.
- The information processing apparatus may further include a notification unit that notifies the user of prescribed information on the basis of the acquired eye information.
- The eye information may include a left-eye image obtained by shooting a left eye of the user and a right-eye image obtained by shooting a right eye of the user. In this case, the movement unit may move at least a part of the image display mechanism on the basis of the left-eye image or the right-eye image.
- An information processing method according to an embodiment of the present technology is an information processing method performed by a computer system and includes acquiring eye information regarding an eye of a user. On the basis of the acquired eye information, at least a part of an image display mechanism that emits and guides image light to the eye of the user is moved.
- A program according to an embodiment of the present technology causes a computer system to perform the following steps.
- The steps include: a step of acquiring eye information regarding an eye of a user; and
- a step of moving, on the basis of the acquired eye information, at least a part of an image display mechanism that emits and guides image light to the eye of the user.
- As described above, the present technology makes it possible to realize high-quality viewing experience for users. Note that the effect described here should not be interpreted in a limited way, and any effect described in the present specification may be produced.
-
FIGS. 1A and 1B are views showing a configuration example of an HMD (Head Mount Display) that is an image display apparatus according to an embodiment of the present technology. -
FIG. 2 is a schematic view for describing a configuration example of the inside of a display unit shown inFIG. 1B . -
FIG. 3 is a block diagram showing a functional configuration example of the HMD according to the present embodiment. -
FIG. 4 is a schematic view for describing an example of eye information regarding the eyes of a user. -
FIG. 5 is a flowchart showing the outline of a basic operation performed by the HMD. -
FIG. 6 is a flowchart showing a processing example performed when the use of the HMD is started. -
FIG. 7 is a flowchart showing a setting example of an initial position. -
FIGS. 8A to 8C are schematic views for describing the setting example of the initial position shown inFIG. 7 . -
FIG. 9 is a flowchart showing an operation example of the HMD after the setting of the initial position. -
FIGS. 10A to 10C are schematic views showing movement examples of a lens system according to a deviation amount. -
FIG. 11 is a flowchart showing the outline of other processing capable of being performed by the HMD. -
FIG. 12 is a perspective view showing the appearance of an HMD according to another embodiment. - Hereinafter, embodiments according to the present technology will be described with reference to the drawings.
-
FIGS. 1A and 1B are views showing a configuration example of an HMD (Head Mount Display) that is an image display apparatus according to an embodiment of the present technology.FIG. 1A is a perspective view schematically showing the appearance of anHMD 100, andFIG. 1B is a perspective views schematically showing a state in which theHMD 100 is disassembled.FIG. 2 is a schematic view for describing a configuration example of the inside of adisplay unit 13 shown inFIG. 1B . Note thatHMD 100 functions also as an information processing apparatus according to the present technology. - The
HMD 100 has abase member unit 10, anattachment band unit 11, aheadphone unit 12, thedisplay unit 13, acover unit 14, and animaging mechanism 15. - The
base member unit 10 is a member arranged in front of left and right eyes 1 (1 a and 1 b) of a user and provided with afront support unit 16 that is brought into contact with the front of the user. - The
attachment band unit 11 is attached to the forehead of the user. As shown inFIGS. 1A and 1B , theattachment band unit 11 has atemporal band 17 and aparietal band 18. Thetemporal band 17 is connected to thebase member unit 10 and attached so as to surround the head of the user from the temporal part to the back part of the head. Theparietal band 18 is connected to thetemporal band 17 and attached so as to surround the head of the user from the temporal part to the parietal part. - In the present embodiment, a band adjustment mechanism 25 (see
FIG. 3 ) capable of automatically adjusting the retention force of each of thetemporal band 17 and theparietal band 18 is provided. By the operation of theband adjustment mechanism 25, the length of thetemporal band 17 protruding outward from thebase member unit 10 is changed. As a result, the retention force for retaining the head is changed. Further, by the operation of theband adjustment mechanism 25, the length of theparietal band 18 is changed. As a result, the retention force for retaining the head is changed. Note that the angle of theparietal band 18 with respect to thetemporal band 17 may be changed. - The configuration of the
band adjustment mechanism 25 is not limited and includes, for example, an arbitrary actuator mechanism using a motor, a piezoelectric element, a wire, a hinge, a solenoid, a shape-memory alloy (SMA), or the like. - The
headphone unit 12 is connected to thebase member unit 10 and arranged so as to cover the left and right ears of the user. Theheadphone unit 12 is provided with left and right speakers. The position of theheadphone unit 12 is manually or automatically controllable. A configuration for controlling the position is not limited, and an arbitrary configuration may be employed. - The
display unit 13 is inserted in thebase member unit 10 and arranged in front of theeyes 1 of the user. As shown inFIG. 2 , thedisplay unit 13 includes animage display mechanism 20 that emits and guides image light to theeyes 1 of the user. - In the present embodiment, the
image display mechanism 20 has adisplay 21 that emits the image light and alens system 22 that guides the image light emitted from thedisplay 21 to theeyes 1 of the user. As thedisplay 21, an arbitrary display device using a liquid crystal, an EL (Electro-Luminescence), or the like may be, for example, used. When an image is displayed on thedisplay 21, light emitted from thedisplay 21 corresponds to the image light. - Note that the image light is emitted from the display so as to be diffused. That is, the image light is emitted as diffused light. For example, when the image light emitted so as to be diffused is appropriately guided to the
eyes 1 of the user by thelens system 22, an image constituted by the image light is made visually recognizable in a focused state. - The
lens system 22 has a left-eye lens system 22 a arranged in front of aleft eye 1 a and a right-eye lens system 22 b arranged in front of aright eye 1 b. The configuration of eachlens system 22 is arbitrarily and is not limited to a case in which one lens is arranged. A plurality of various lenses such as Fresnel lenses or a plurality of arbitrary optical members such as optical filters may be arranged. - As schematically shown by arrows in
FIG. 2 , a drive mechanism 26 (seeFIG. 3 ) capable of moving each of thedisplay 21, the left-eye lens system 22 a, and the right-eye lens system 22 b is provided in the present embodiment. By the operation of thedrive mechanism 26, it is possible to arbitrarily change the position or attitude (direction) of thedisplay 21, the position or attitude (direction) of the left-eye lens system 22 a, and the position or attitude (direction) of the right-eye lens system 22 b independently of each other. - As shown in, for example,
FIG. 2 , it is assumed that a vertical direction, a horizontal direction, and a depth direction (a direction toward the display 21) are an X direction, a Y direction, and a Z direction, respectively, when thedisplay 21 is seen from theeyes 1 of the user. Thedrive mechanism 26 is capable of performing a drive operation such as parallel movements along respective axis directions and rotation based on respective axes with respect to thedisplay 21 or eachlens system 22. The specific configuration of thedrive mechanism 26 is not limited, and an arbitrary actuator mechanism as described above may be used. - In the present embodiment, each of the
display 21, the left-eye lens system 22 a, and the right-eye lens system 22 b corresponds to a part of theimage display mechanism 20. That is, the movement of each of thedisplay 21, the left-eye lens system 22 a, and the right-eye lens system 22 b corresponds to the movement of at least a part of theimage display mechanism 20. - The
imaging mechanism 15 has a left-eye camera 23 a that shoots theleft eye 1 a of the user and a right-eye camera 23 b that shoots theright eye 1 b of the user. Each of the left-eye camera 23 a and the right-eye camera 23 b is installed at a prescribed position of theHMD 100, specifically, at a prescribed position of thebase member unit 10. Accordingly, when the relative position of thebase member unit 10 with respect to theeyes 1 of the user changes, the relative positons of the left-eye camera 23 a and the right-eye camera 23 b with respect to theeyes 1 of the user also change. - Note that the left-
eye camera 23 a and the right-eye camera 23 b are arranged at positions at which theleft eye 1 a and theright eye 1 b of the user are capable of being shot directly, that is, without the interposition of the left-eye lens system 22 a and the right-eye lens system 22 b. The left-eye camera 23 a and the right-eye camera 23 b are arranged so as to be directed obliquely from, for example, a downward side. Of course, the left and right cameras may be arranged so as to be directed to theeyes 1 of the user from other directions. - As the left-
eye camera 23 a and the right-eye camera 23 b, digital cameras including image sensors such as CMOS (Complementary Metal-Oxide Semiconductor) sensors and CCD (Charge Coupled Device) sensors are, for example, used. Further, infrared cameras installing infrared illumination such as an infrared LED may be, for example, used. - The
cover unit 14 shown inFIGS. 1A and 1B is attached to thebase member unit 10 and configured so as to cover thedisplay unit 13. TheHMD 100 thus configured functions as an immersive head mount display configured so as to cover the visual field of the user. By wearing theHMD 100, the user is allowed to experience, for example, virtual reality (VR) or the like. -
FIG. 3 is a block diagram showing a functional configuration example of theHMD 100 according to the present embodiment. TheHMD 100 further has anoperation button 27, acommunication unit 28, aconnector 29, astorage unit 30, asensor unit 31, and acontroller 32. - The
operation button 27 is provided at, for example, a prescribed position of thebase member unit 10. By theoperation button 27, an operation to turn on/off a power supply or operations related to various functions of theHMD 100 such as a function related to an image display or a sound output and a network communication function may be performed. - The
communication unit 28 is a module for performing network communication, near field communication, or the like with other devices. A wireless LAN module such as WiFi or a communication module such as Bluetooth™ is, for example, provided. - The
connector 29 is a terminal for connection with other devices. A terminal such as a USB (Universal Serial Bus) and a HDMI™ (High-Definition Multimedia Interface) is, for example, provided. Further, during charging, theconnector 29 is connected to the charging terminal of a charging dock (cradle) to be charged. - The
sensor unit 31 includes apressure sensor 34, aproximity sensor 35, a nine-axis sensor 36, aGPS 37, and abiosensor 38. Thepressure sensor 34 is provided at, for example, prescribed positions of thetemporal band 17 and theparietal band 18 shown inFIGS. 1A and 1B . By thepressure sensor 34, it is possible to measure pressure applied from thetemporal band 17 and theparietal band 18 to the head. - The
proximity sensor 35 is provided at a prescribed position on the inner periphery side of theattachment band unit 11, and its detection result is used to determine the attachment/detachment of theHMD 100. The nine-axis sensor 36 includes a three-axis acceleration sensor, a three-axis gyro sensor, and a three-axis compass sensor. By the nine-axis sensor 36, it is possible to detect acceleration, angular speed, and directions in three axes. TheGPS 37 acquires information regarding the current place of theHMD 100. These sensors are provided at, for example, prescribed positions of thebase member unit 10. Of course, these sensors may be provided at other positions. - The
biosensor 38 acquires biological information regarding the user. As thebiosensor 38, a brain wave sensor, a myoelectric sensor, a pulse sensor, a sweat sensor, a temperature sensor, a blood flow sensor, a body motion sensor, or the like is, for example, provided. These sensors are provided at prescribed positions of theHMD 100 so that their detection terminal parts are in contact with prescribed positions of the body. For example, the brain wave sensor is provided to be capable of being in contact with a prescribed position of the head. The pulse sensor is provided at a position at which the pulse sensor is capable of being in contact with a blood vessel of the neck. - The types of sensors provided as the
sensor unit 31 are not limited, and arbitrary sensors may be provided. A temperature sensor, a humidity sensor, or the like capable of measuring temperature, humidity, or the like of an environment in which theHMD 100 is used may be, for example, provided. - The
storage unit 30 is a non-volatile storage device, and a HDD (Hard Disk Drive) or the like is, for example, used as such. Thestorage unit 30 stores acontrol program 40 for controlling the entire operation of theHMD 100. Further, thestorage unit 30 stores a mechanism drive table 41. The mechanism drive table 41 is a table referred to when at least a part of theimage display mechanism 20 is moved on the basis of eye information regarding the user, and will be described in detail later. A method for installing thecontrol program 40 and the mechanism drive table 41 in theHMD 100 is not limited. - The
controller 32 controls the operations of the respective blocks of theHMD 100. Thecontroller 32 has, for example, hardware configurations for a computer such as a CPU and a memory (RAM, ROM). Various processing is performed when the CPU loads thecontrol program 40 stored in thestorage unit 30 into the RAM and performs the same. - As the
controller 32, a PLD (Programmable Logic Device) such as a FPGA (Field Programmable Gate Array), any device such as an ASIC (Application Specific Integrated Circuit), or the like may be, for example, used. - In the present embodiment, an
image analysis unit 43, adeviation calculation unit 44, amechanism drive unit 45, adisplay control unit 46, aband adjustment unit 47, astate analysis unit 48, areliability determination unit 49, and anotification unit 50 are realized when the CPU of thecontroller 32 performs a program according to the present embodiment. Further, an information processing method according to the present embodiment is performed by these function blocks. Note that dedicated hardware such as an IC (Integrated Circuit) may be appropriately used to realize the respective function blocks. - The
image analysis unit 43 analyzes eye images of the user shot by theimaging mechanism 15. That is, theimage analysis unit 43 acquires a left-eye image shot by the left-eye camera 23 a and a right-eye image shot by the right-eye camera 23 b and analyzes these images. All the left-eye image and the right-eye image input to theimage analysis unit 43 and various characteristic parameters detected as a result of the analysis of theimage analysis unit 43 are included in eye information regarding theeyes 1 of the user in the present embodiment. On the basis of these eye information items, it is possible to dynamically estimate, for example, the positions of theeyes 1 of the user. In the present embodiment, theimage analysis unit 43 functions as an acquisition unit. -
FIG. 4 is a schematic view for describing an example of eye information regarding theeyes 1 of the user. As eye information, it is possible to detect, for example, the shape, size, position, inclination, or the like of aneye 1P of the user in aneye image 55. Specifically, aneye region 56 of the user is detected, and its shape, size, position, and inclination are detected. As shown in, for example,FIG. 4 , a size t of along axis 57 of theeye 1P of the user may be used as the size of theeye region 56 of the user. Further, the inclination of thelong axis 57 of theeye 1P of the user with respect to a horizontal direction may be used as the inclination of theeve region 56 of the user. - Note that the
long axis 57 of theeye 1P or the user is an axis connecting the position of the right end and the position of the left end of each of theeyes 1 of the user to each other in a substantially horizontal direction when the vertical direction of the face (for example, a direction from the parietal part to the jaw) of the user becomes a substantially vertical direction and the horizontal direction of the face (for example, a direction in which positions at the same height of both ears are connected to each other) becomes a substantially horizontal direction. - Further, information such as information as to whether the
entire eye region 56 of the user is included in theeye image 55 and information as to which of the vertical and horizontal directions is a direction in which theeye region 56 is missing is also included in the eye information regarding theeyes 1 of the user. Besides, arbitrary information such as a visual line direction, a pupil size, an iris pattern, and an eyelash shape may be detected as the eye information. - Note that a method for analyzing an eye image to detect characteristic parameters is not limited. An arbitrary segmentation technology or an arbitrary image analysis technology may be used. Further, a machine learning algorithm using a DNN (Deep Neural Network) such as a RNN (Recurrent Neural Network), a CNN (Convolutional Neural Network), and a MLP (Multilayer Perceptron) may be used.
- The
deviation calculation unit 44 calculates a deviation amount from the initial state of theeyes 1 of the user on the basis of eye information regarding the user acquired by theimage analysis unit 43. The initial state of theeyes 1 of the user and a deviation from the initial state will be described later. - The
mechanism drive unit 45 outputs a control signal to thedrive mechanism 26. On the basis of the control signal output from themechanism drive unit 45, each of thedisplay 21, the left-eye lens system 22 a, and the right-eye lens system 22 b is moved. In the present embodiment, thedeviation calculation unit 44 and themechanism drive unit 45 function as a movement unit. - The
display control unit 46 controls an image display by theimage display mechanism 20. By thedisplay control unit 46, arbitrary image processing and display control such as correcting a displayed image and moving the display position of a displayed image are, for example, performed. - The
band adjustment unit 47 outputs a control signal to theband adjustment mechanism 25. On the basis of the control signal output from theband adjustment unit 47, the retention force of each of thetemporal band 17 and theparietal band 18 is adjusted. In the present embodiment, theband adjustment unit 47 functions as an attachment state control unit capable of controlling the attachment state of theHMD 100. - The
state analysis unit 48 acquires various state information items on the states of the user or theHMD 100 on the basis of detection results from thesensor unit 31. As the state information regarding the user, information as to whether the user is in a sleeping state or biological information such as body heat, a pulse rate, a brain wave state, muscle movements, a sweat amount, and a concentration state is, for example, acquired. Further, information regarding the current place of the user, more specifically, information as to whether the user is indoors, outdoors, or is in conference, or the like is acquirable. - Further, motion information regarding motion done by the user is acquired. Information during walking, running, moving by train, driving, or the like is, for example, acquired. Information on the types of sports being played or the like is also acquirable. Further, information regarding a posture as to whether the user is sitting, standing, stooping, turning sideways, or turning up is acquired. These state information items on the user are acquirable by, for example, an arbitrary motion analysis technology such as a motion analysis using parameters obtained by machine learning or the like.
- Further, as information regarding the apparatus states of the
HMD 100, various information items such as an active function, an operation mode, the attachment/detachment of theHMD 100 to/from the user, the attachment position of theHMD 100 with respect to the user, a battery remaining amount, connection with a charging dock, and apparatus temperature are, for example, acquired. Further, as information regarding use environments, various information items such as temperature, humidity, a current place, weather, date and time are also acquirable. Sensors, devices, or the like for acquiring these information items only have to be appropriately provided in theHMD 100. Further, a method for analyzing state is not limited, and a machine learning algorithm may be, for example, used. In the present embodiment, thestate analysis unit 48 functions as a state acquisition unit. - The
reliability determination unit 49 determines the reliability of a detection result by thesensor unit 31. The determination will be described later. - The
notification unit 50 notifies the user of various information items. The notification of prescribed information is realized by, for example, an arbitrary method such as displaying an image or text on thedisplay 21, outputting a sound from theheadphone unit 12, shining light on (blinking) a prescribed position at the outer part of thebase member unit 10 or thecover unit 14, and vibrating thebase member unit 10 or the like with an arbitrary vibration mechanism. It is possible to make, for example, a notification of arbitrary information such as information regarding the use of theHMD 100, information regarding content, and information regarding a detection result by thesensor unit 31. - The operation of the
HMD 100 that is an image display apparatus according to the present embodiment will be described.FIG. 5 is a flowchart showing the outline of a basic operation performed by theHMD 100. First, eye information regarding theeyes - The change in viewing state includes, for example, arbitrary processing to change a state in which content is viewed, such as moving the left and
right lens systems display 21, moving the display position of as image, and adjusting theattachment band unit 11. Further, the notification itself of a prescribed alert or the like is also included in the change in viewing state. - That is, in the
HMD 100 of the present embodiment, various processing is performed on the basis of eye information regarding theeyes 1 of the user. Thus, it is possible to realize high-quality viewing experience for the user. Hereinafter, a specific example to which the basic operation shown inFIG. 5 is applied will be described. -
FIG. 6 is a flowchart showing a processing example performed when the use of theHMD 100 is started. When the user turns on the power supply of theHMD 100, a determination is first made as to whether theHMD 100 has been attached to the head of the user (step 201). The determination is made by thestate analysis unit 48 shown inFIG. 3 on the basis of a detection result by the proximity sensor 24. - When the
HMD 100 has not been attached (No in step 201), the user is notified of the fact by thenotification unit 50 shown inFIG. 3 (step 202). A notification method is not limited. For example, a sound may be output from theheadphone unit 12, or prescribed colored light may be output to the outside of thebase member unit 10 or thecover unit 14. Alternatively, processing to make theentire display 21 blink may be performed. Of course, the notification of the fact to the user may be performed by the vibration of thebase member unit 10 or the like based on a vibration mechanism. - When the
HMD 100 has been attached (Yes in step 201), the search of the eye regions of the user is started (step 203). Specifically, the detection of theeye region 56 of theleft eye 1 a and theeye region 56 of theright eye 1 b is performed by theimage analysis unit 43 shown inFIG. 3 . Then, a determination is made as to whether the left andright eye regions 56 are not completely detectable (step 204). - When the left and
right eye regions 56 are not completely detectable (Yes in step 204), a determination is made as to whether theHMD 100 is physically moving (step 205). The determination is performed by thestate analysis unit 48 on the basis of, for example, a detection result by the nine-axis sensor. - Besides, as determination processing in step 205, it is also possible to provide a proximity camera for finger print shooting or the like in the
sensor unit 31 and shoot the skin or the like of the user with the proximity camera to measure the movement direction and distance of a skin pattern. For example, it is also possible to estimate the movement distance on the basis of a cumulative movement distance since calibration. Further, the determination as to whether theHMD 100 is physically moving may be made on the basis of an image shot by a camera (including IoT or the like) provided at a distant place. For example, face detection is performed with respect to the face wearing theHMD 100 to detect the position of a LED marker attached to the HMD. It is possible to determine whether the HMD has physically moved from the result of the face detection and the positional relationship of the LED marker of the HMD. - When the
HMD 100 is moving (Yes in step 205), the processing is on standby until theHMD 100 stops (step 206). That is, the processing is on standby until the state of the user or theHMD 100 is settled. After being on standby for a prescribed time, the processing returns to step 203. When theHMD 100 is not moving (No in step 205), the user is notified of an instruction to change an attachment state so that theeyes 1 of the user are capable of being shot by the left-eye camera 23 a and the right-eye camera 23 b (step 207). For example, announcement or the like urging the user to manually correct the position of theHMD 100 is output. Then, the processing returns to step 203. - When the left and
right eye regions 56 are detectable (No in step 204), a determination is made as to whether all the left andright eye regions 56 are detectable (step 208). When there is any missing region in the left and right eye regions 56 (No in step 208), a direction in which theHMD 100 is deviated and the degree to which theHMD 100 is deviated are estimated on the basis of the missing region (step 209). - A control signal is generated on the basis of the deviation amount estimated in step 209 and output to the
band adjustment mechanism 25 by theband adjustment unit 47. That is, the control signal is output to theband adjustment mechanism 25 so as to create an attachment state in which theeyes 1 of the user are capable of being shot by the left-eye camera 23 a and the right-eye camera 23 b. Then, the lengths or positions of thetemporal band 17 and theparietal band 18 of theattachment band unit 11 are changed by the band adjustment mechanism 25 (step 210). - When the fact that the
HMD 100 is deviated in the horizontal direction is, for example, estimated in step 209, the length or the like of thetemporal band 17 is adjusted. When the fact that theHMD 100 is deviated in the vertical direction is estimated, the length or the like of theparietal band 18 is adjusted. Of course, the adjustment is not limited to such processing. Note that the band on the temporal part side and the band on the parietal part side shown inFIG. 6 correspond to thetemporal band 17 and theparietal band 18, respectively. When the adjustment of theattachment band unit 11 is completed, the processing returns to step 203. - When all the left and
right eye regions 56 are detectable (Yes in step 208), the setting of an initial position is performed (step 211). -
FIG. 7 is a flowchart showing a setting example of an initial position.FIGS. 8A to 8C are schematic views for describing the setting example of the initial position shown inFIG. 7 . The setting of the initial position refers to the setting of the states of the left andright eyes 1 of the user with respect to theimage display mechanism 20 to an initial state in which content is capable of being appropriately viewed when the use of theHMD 100 is started. In the present embodiment, the initial state in which the content is capable of being appropriately viewed corresponds to “a case in which the eyes of the user are put in a prescribed state.” Further, the initial state refers also to a prescribed viewing state. - First, a
reference image 60 is displayed at a prescribed position of the display 21 (step 301). As shown in, for example,FIG. 8A , a point image colored in green or the like is displayed as thereference image 60 at a substantially central area of thedisplay 21. Then, left andright eye images 55 shot in this state are analyzed (step 302). - A determination is made as to whether a
reference image 60P is arranged at a prescribed size and a prescribed position with respect to the left andright eyes 1P of the user in the left andright eye images 55. Specifically, a determination is made as to whether the size and position of thereference image 60P are put in a prescribed state (step 303). In the present embodiment, a determination is made as to whether thereference image 60P is displayed minimally and clearly at the center of the pupil in the left andright eye images 55. This state corresponds to a state in which an image of the diffused and emitted image light of thereference image 60 is formed minimally and a state in which the left andright eyes 1 of the user are focused with respect to thereference image 60 displayed on thedisplay 21. That is, this state is a state in which the above initial state is realized. - When the size and the position of the
reference image 60P are not put in the prescribed state in the left andright eye images 55, the left andright lens systems drive mechanism 26 by themechanism drive unit 45. On the basis of the input control signal, thedrive mechanism 26 moves the left andright lens systems - When a reference image 60PA is deviated rightward with respect to the center of the pupil and expanded in size as shown in, for example,
FIG. 8C , it is determined that focus is not achieved. Then, thelens system 22 is appropriately moved on the basis of XYZ axes so as to create a focused state (a state shown inFIG. 8B ). Similarly, when a reference image 60PB is deviated leftward with respect to the center of the pupil and expanded in size as shown inFIG. 8C , thelens system 22 is appropriately moved on the basis of the XYZ axes so as to create a focused state. - By, for example, moving the
lens system 22 in a Z direction, it is possible to control the size of thereference image 60P. Further, by moving thelens system 22 in parallel in an XY plane direction or by inclining thelens system 22 along the respective axes, it is possible to control the position of thereference image 60P. Besides, an arbitrary drive method capable of controlling the size and position of thereference image 60P may be performed. - Note that when the
reference image 60P becomes minimum at a central position after the processing of steps 303 and 304 is repeatedly performed by a prescribed number of times, it may be determined that the prescribed state (that is, the initial state) in step 303 is realized. Alternatively, when prescribed allowable ranges are set with respect to the position and the size and the size and position of thereference image 60P are included in the allowable ranges, it may be determined that the prescribed state (that is, the initial state) in step 303 is realized. - When it is determined that the
reference image 60P is put in the prescribed state, that is, when the initial state is realized (Yes in step 303), the size, position, and inclination of each of left andright eye regions 56 are stored in the storage unit 30 (step 305). For example, the size (for example, the number of pixels) of the long axis of theeye region 56 shown inFIG. 4 is stored as the size of each of theeye regions 56. Further, the position and inclination (for example, coordinate values of pixels) of thelong axis 57 are stored as the position and inclination of each of theeye regions 56. Besides this, the sizes and positions of theentire eye regions 56 may be stored as they are. - In the present embodiment, eye information acquired when the initial state is realized, that is, eye images shot when the initial state is realized or various characteristic parameters calculated from the eye images correspond to reference eye information. Further, the eye images shot when the initial state is realized correspond to reference eye images.
- By performing the setting of the initial position as described above, it is possible to substantially prevent influence caused by a personal difference in the shape of the face or head of the user or a personal difference in the distance between left and right eyes and realize a high-quality viewing system for each user who uses the
HMD 100. - Note that the shape, size, color, or the like of the
reference image 60 is not limited and may be arbitrarily designed. For example, an image of a building, a vehicle, or a balloon, a prescribed logo mark, or the like may be displayed as thereference image 60. Further, a condition which is set with respect to thereference image 60P in theeye images 55 and on which the realization of the state is determined is not also limited. That is, a condition on the size or position of thereference image 60P may be arbitrarily set, and a parameter or the like (such as a brightness value and a blurring degree) other than the size or position may be employed as the condition on the realization of the initial state. - In the case of, for example, the
reference image 60 having a simple shape like the point image shown inFIG. 8A , a determination is made as to whether the initial state is realized on the basis of the size of thereference image 60P in theeye images 55. On the other hand, in a case in which areference image 60 having a complicated shape such as a balloon is used, it is possible to determine whether the initial state is realized on the basis of not only the size of thereference image 60P in theeye images 55 but also focus (a blurring degree). Of course, a condition for determining the realization of the initial state is not limited to the above example. - Further, in step 304, the
display 21 may be moved to realize the initial state. Alternatively, the display position of thereference image 60 may be moved. For example, in a case in which the display position of thereference image 60 is changed to perform the setting of the initial position, a content image or the like only has to be displayed at a position corresponding to the position of thereference image 60 obtained when the initial state is realized. - The setting of the initial position refers to processing to set focus or a display position before content is viewed. In the present technology, it is possible to dynamically change the viewing state not only at the starting of content but also at the running (viewing) of the content as will be described below.
-
FIG. 9 is a flowchart showing an operation example of theHMD 100 after the setting of the initial position. Steps 401 to 410 are substantially the same as steps 201 to 210 shown inFIG. 6 . Immediately after the setting of the initial position is completed, the processing proceeds from Yes in step 408 to step 411 in most cases. Note that the user is highly likely to walk or run in VR (Virtual Reality) during the viewing of content. In this case, theHMD 100 could be deviated. However, it is possible to cause the processing to automatically proceed to step 411 by performing the processing of steps 401 to 410. - In step 411, a determination is made by the
deviation calculation unit 44 shown inFIG. 3 as to whether there is any change in the inclinations of thelong axes 57 of theeye regions 56 in comparison with the inclinations of thelong axes 57 in the initial state. When there is a change in the inclinations of thelong axes 57, the display position of an image (a content image) displayed on thedisplay 21 is moved by thedisplay control unit 46. Specifically, a display coordinate axis with respect to thedisplay 21 is rotated at a rotation angle corresponding to the change in the inclinations of the long axes 57 (step 412). Thus, it is possible to correct the inclination of an image caused by the rotation deviation or the like of theHMD 100. - When there is no change in the inclinations of the
long axes 57, a determination is made by thedeviation calculation unit 44 as to whether there is any change in the shapes, sizes, and positions of theeye regions 56 in comparison with the shapes, sizes, and positions of theeye regions 56 in the initial state (step 413). When there is no change in the shapes, sizes, and positions of theeye regions 56, the processing returns to step 403 without changing the viewing state. - When there is a change in the shapes, sizes, and positions of the
eye regions 56, a control signal corresponding to the change is generated and output to thedrive mechanism 26 by themechanism drive unit 45. Thus, the respective positions or angles of thedisplay 21, the left-eye lens system 22 a, and the right-eye lens system 22 b are adjusted. (step 414). Specifically, the respective positions or angles of thedisplay 21, the left-eye lens system 22 a, and the right-eye lens system 22 b are adjusted so that the state of theeyes 1 of the user with respect to theimage display mechanism 20 comes close to the initial state. - Therefore, in the present embodiment, the mechanism drive table 41 stored in the
storage unit 30 is referred to. The mechanism drive table 41 is a table in which deviation amounts calculated by thedeviation calculation unit 44 and movement amounts of thedisplay 21, the left-eye lens system 22 a, and the right-eye lens system 22 b are associated with each other. The mechanism drive table 41 is created in advance by calibration or the like and stored in thestorage unit 30. - For example, a two-dimensional lookup table showing the relationships between deviation amounts and the combinations of movement amounts of the display and the lens system with respect to each of the left and right eyes is created. Alternatively, a three-dimensional lookup table showing the relationships between deviation amounts, movement amounts of the display, and movement amounts of the lens system may be created. Further, a condition for moving the display, a condition for moving the lens system, or the like is set, and a device to be moved may be appropriately determined. In this case, a plurality of table information items used according to a determined result may be created.
- Of course, the movement of the
display 21 or thelens system 22 may be performed without using table information. For example, movement amounts may be calculated by computation as occasion demands on the basis of deviation amounts to appropriately control the operation of thedrive mechanism 26. -
FIGS. 10A to 10C are schematic views showing movement examples of thelens system 22 according to a deviation amount. As shown in, for example,FIG. 10A , it is assumed that the size of aneye region 56 in aneye image 55 becomes smaller than that of theeye region 56 in the initial state. In this state, thebase member unit 10 in which the left-eye camera 23 a and the right-eye camera 23 b are installed is moved in a direction away from theeyes 1 of the user. In order to correct influence caused by this deviation, that is, in order to bring this state close to the initial state, the left andright lens systems 22 are caused to come close to the eyes of the user. - As shown in
FIG. 10B , it is assumed that the size of aneye region 56 in aneye image 55 becomes greater than that of theeye region 56 in the initial state. In this state, thebase member unit 10 in which the left-eye camera 23 a and the right-eye camera 23 b are installed is moved in a direction close to theeyes 1 of the user. In order to correct influence caused by this deviation, that is, in order to bring this state close to the initial state, the left andright lens systems 22 are caused to get away from the eyes of the user. - As shown in
FIG. 10C , it is assumed that the position of aneye region 56 in aneye image 55 is deviated from the position of theeye region 56 in the initial state. In this state, thebase member unit 10 in which the left-eye camera 23 a and the right-eye camera 23 b are installed is moved in the XY plane. In order to correct influence caused by this deviation, that is, in order to bring this state close to the initial state, the left andright lens systems 22 are caused to move in a direction in which the deviation is corrected. Alternatively, it is also possible to adjust the angles of the left andright lens systems 22 to correct the positional deviation of theeye region 56. - That is, in the present embodiment, at least a part of the
image display mechanism 20 is moved on the basis of the differences between reference eye images shot in the initial state and eye images shot during the viewing of content. Note that a method for moving at least a part of theimage display mechanism 20 according to a deviation is not limited to the methods shown in the examples ofFIGS. 10A to 10C and an arbitrary drive method (correction method) may be performed. For example, it is also possible to perform processing to move thedisplay 12 or thelens system 22 so that the sizes of the left andright eye regions 56 become substantially the same. - As described above, a change in viewing state and a prescribed notification are performed on the basis of eye information regarding the
eyes 1 of the user in theHMD 100 according to the present embodiment. That is, the movement of at least a part of theimage display mechanism 20, the adjustment of theattachment band unit 11, and the notification of prescribed information are performed on the basis of eye information. Thus, it is possible to realize high-quality viewing experience for, for example, the user who uses theHMD 100. - In the future as well, it is expected that the experience of VR (Virtual Reality) or AR (Augmented Reality) using an HMD or the like will become pervasive. Further, it is also expected that the weight reduction of a used HMD will be advanced, and it will be really important to effectively correct influence caused by the deviation of an HMD. According to, for example, an analog method in which the user manually moves an HMD to be adjusted, it is difficult to adjust a position at which focus is achieved and an image is appropriately viewed and also difficult to determine the most appropriate position for the user himself/herself. As a result, the user struggles to adjust the position of the HMD, and it is highly likely that the viewing of content is disturbed.
- In the present embodiment, the movement or the like of the
lens system 22 is automatically performed on the basis of eye information regarding theeyes 1 of the user even during the viewing of content. Thus, it is also possible to correct influence caused by a deviation before, for example, being noticed by the user without disturbing the viewing of content. As a result, it is possible to realize extremely high-quality viewing experience. Further, in a case in which a manual adjustment may be needed, it is also possible to determine a deviation amount on the basis of eye information and notify the user of an appropriate adjustment method. Further, it is also possible to notify the user of whether a position is most appropriate. That is, it is possible to give effective guidance to the user. - Examples of a notification method for the user will be described. For example, when it is determined that a deviation is beyond an automatic adjustment, the user is notified of the occurrence of the deviation and urged to make an adjustment. Thus, it is possible to prevent an unintended attachment state from being kept. Specifically, the display of an image or text on the
display 21, the output of a sound from theheadphone unit 12, the vibration of thebase member unit 10 or the like with a vibration mechanism, or the like is performed. - When an image becomes blurred during, for example, the viewing of content, the user notices the blurring. Accordingly, it is also effective to notify the user of the blurring straightforwardly. A notification or alert is displayed using, for example, text at a size at which the text is seeable even in a blurred state or an image, a symbol, a mark, or the like that is recognizable even in a blurred state. Alternatively, a notification or alert is performed by a sound, vibration described above, or the like.
- It is also important to naturally perform a notification without disturbing the viewing of the user. A short notification that is not so emphasized is, for example, performed. Alternatively, it is also possible to determine the state of the user from a sensor result by the
sensor unit 31 and reflect the determined state of the user on a notification method. On the basis of, for example, the state of sweating, brain waves, or the like, a determination is made as to whether the user notices a notification. When the user does not notice the notification, the notification is emphasized stepwise or other notification methods are performed. For example, when the user does not notice a sound, a vibration function is used in combination with the sound. - When it is determined that the user does not completely notice a notification, it is likely that the user is not viewing content at all. In this case, the awakened degree of the user is measured, and a notification method is appropriately changed on the basis of a measurement result. For example, processing such as emphasizing the notification to make the user awakened, cancelling the content contrary, and turning off the power supply of the
HMD 100 itself is assumed. - When it is determined that the
HMD 100 is deviated horizontally, it is also possible to give guidance to the user through content to move the head instantaneously to correct the deviation. For example, a sound is used and quickly transmitted from the right ear to the left ear. Alternatively, an image that quickly moves horizontally is displayed on thedisplay 21. The user is caused to feel the movement of any object on the screen through this content and then move the head instantaneously to follow the object. Thus, it is also possible to give guidance to the user to move the head. In addition, it is also possible to give guidance to the user to move the head instantaneously using both a sound and the movement of an image. - It is also possible to change the retention force of the
attachment band unit 11 to correct the deviation of theHMD 100. Using, for example, a prescribed button operation, the movement of the head (an increase in acceleration), or a sound input from the user as a trigger, it is also possible to loosen one side and tighten the other side of theattachment band unit 11 to move and restore theentire HMD 100 itself to a normal position. -
FIG. 11 is a flowchart showing the outline of other processing capable of being performed by theHMD 100. As shown in step 501, information regarding the state of the user, the details of displayed content, and a viewing time is acquired. Then, as shown in step 502, a change in viewing state and a prescribed notification are performed on the basis of these information items. - That is, processing or the like to move at least a part of the
image display mechanism 20 may be performed on the basis of information different from eye information. Of course, these information items and eye information may be combined together to perform a change in viewing state and a prescribed notification. Thus, it is possible to realize high-quality viewing experience. - In the above description, the
lens system 22 or the like is moved so that the state of theeyes 1 of the user comes close to the initial state when a deviation occurs. Besides this case, the movement of thedisplay 21 or thelens system 22 may be performed so that the state of theeyes 1 of the user with respect to theimage display mechanism 20 is put in another state different from the initial state. That is, the viewing state may be changed to a state in which focus is not achieved intentionally, or may be changed so as to make the display position of an image deviated. - For example, when content refers to a video game or the like, it is also possible to cause defocusing intentionally to make a view blurred in water or a dusty environment. The blurring of a view is performed in, for example, a case in which the user is covered with dust from a forward-traveling vehicle, a case in which the user dives into water, or the like in a racing game. Further, it is also possible to adjust the blurred degree of a view according to the attachment/detachment of an appropriate item (for example, glasses or the like) inside VR.
- Further, it is also possible to adjust the blurred degree of a view according to the physical condition or awakened state of an avatar in content. For example, it is also possible to sense the state of the user and reflect the sensed state of the user on an avatar, or is possible to adjust a blurred degree according to the state of an avatar in content. The blurring of a view or the like is performed in a case in which the user is damaged or is in a critical condition in, for example, a battle game, a shooting game, or the like. It is also possible to make only the right eye blurred when the right eye of the user gets hit. Further, it is also possible to make a view blurred when the user feels strong sleepy.
- As described above, the viewing state is changed according to the details of content, whereby it is possible to enhance a sense of realism of the content and realize high-quality viewing experience.
- Further, in medical fields or the like, it is also possible to realize the experience of the visual fields of foresighted persons, near-sighted persons, astigmatic persons, and persons with visual and hearing disabilities, or the like. For example, when the user wearing the
HMD 100 selects a visual field that he/she wants to experience, thelens system 22 or the like is moved on the basis of eye information so as to realize the visual field. For example, thelens system 22 or the like is moved so as to make a focal distance get close to the user when the user selects the visual field of a foresighted person and moved so as to make the local distance get away from the user when the user selects the visual field of a farsighted person. When the user selects the visual field of a foresighted person, thelens system 22 or the like is appropriately moved according to the distance of an object. Note that it is also possible to automatically adjust visibility during the attachment of theHMD 100 by moving thelens system 22 or the like. - Further, it is also possible to apply the present technology to ordinary glasses. For example, it is possible to accurately change a correction frequency, a visual field, or the like by moving lenses or the like provided in glasses on the basis of eye information.
- The viewing state is changed on the basis of the viewing time of content. For example, when content is used over a specified time, it is possible to make a view blurred to cause the user to be aware of fatigue and have a rest. For example, as the used time of content comes close to a content setting time (90 minutes or the like), the user is caused to be gradually put in a near-sighted state (caused to have a shorter focal distance) and feel eyestrain. Such processing corresponds to the power mode of a TV. Further, on this occasion, it is also possible to determine whether the user gets tired on the basis of biological information (brain waves, pulses, blood flows, or the like). Further, it is also possible to display a message such as “enter an eyestrain mode” at a corner of the
display 21. Further, in, for example, a case in which theHMD 100 is used over a specified time, a case in which it is determined that the user gets tired on the basis of biological information, or the like, it is also possible to offer content such as an image or music for relaxing eyestrain to the user. - The viewing state is changed on the basis of state information regarding the state of the user. The processing is performed according to an awakened degree described above or performed according to biological information. For example, when a mechanism or the like that monitors the physical condition of the user wearing the
HMD 100 is provided, a display state may be changed using a monitoring result. For example, when it is determined that the user gets tired on the basis of biological information, the user is caused to be gradually put in a near-sighted state (caused to have a shorter focal distance) and feel eyestrain. Further, it is also possible to display a message such as “enter an eyestrain mode” at a corner of thedisplay 21. Further, it is also possible to offer content such as an image or music for relaxing eyestrain to the user. - The
reliability determination unit 49 shown inFIG. 3 will be described. Thereliability determination unit 49 determines the reliability of detection result by thebiosensor 38 on the basis of a movement amount (correction amount) from the initial state in thedisplay 21 or the left andright lens systems 22. For example, it is determined that theHMD 100 is deviated from the head as movement amounts of thelens system 22 are greater. That is, with the assumption that the detection terminal part of a brain wave sensor, a myoelectric sensor, or the like is not highly likely to be substantially in contact with a prescribed position of the body, the reliability of a detected signal is set to be low. Thus, it is possible to effectively use a detection result by thebiosensor 38. - For example, it is possible to detect the swinging or the like of the
HMD 100 with the nine-axis sensor 36 but is not possible to determine whether the detection terminal part of thebiosensor 38 is substantially in contact with the skin. In the present embodiment, it is possible to determine whether a detection terminal part is substantially in contact with the skin and determine the reliability of a detection result on the basis of a movement amount by the mechanism drive unit 145 with respect to at least a part of theimage display mechanism 20. - Note that the reliability of a detection result by the
biosensor 38 may be determined on the basis of, for example, biological information obtained from the biosensors of other wearable devices in the form of a watch, a neckless, a ring, or the like or determination results about contacting the skin with biosensors determined by other wearable devices. Of course, the reliability of a detection result by thebiosensor 38 may be determined on the basis of biological information or the like acquired by other terminals not classified into wearable devices. - When eye information and other information are used in combination, the used information may be weighted. A weighting method is not limited and may be arbitrarily set.
- The present technology is not limited to the embodiment described above, and various other embodiments may be realized.
- The above description refers to a case in which each of the
display 21, the left-eye lens system 22 a, and the right-eye lens system 22 b is movable as an example of moving at least a part of theimage display mechanism 20. Besides this, only thedisplay 21 may be movable, or only the left andright lens systems 22 may be movable. Further, other members constituting theimage display mechanism 20 may be movable. Further, only a part of theimage display mechanism 20 may be movable to change a viewing state. - When the
lens system 22 or the like is frequently moved by themechanism drive unit 45, it may be determined that the fastening degree of theattachment band unit 11 is loose. When it is determined that the fastening degree is loose, theattachment band unit 11 is automatically adjusted on the basis of, for example, eye information. Alternatively, a notification about urging the adjustment of theattachment band unit 11 is made. - By using the present technology, it is also possible to view content in a state in which the fastening degree of the
attachment band unit 11 is made loose on purpose. By, for example, performing the processing shown inFIG. 9 at a high speed, it is possible to prevent influence caused by the deviation of theHMD 100 when the fastening degree is loose. Thus, since there is no need to tightly fasten theattachment band unit 11, it is possible to substantially reduce an attachment load on the user. Further, since a mechanism for finely adjusting theattachment band unit 11 is eliminated, the simplification, miniaturization, and lightweight of the apparatus are realized. Further, time and trouble to attach the apparatus is saved. - Further, it is also possible to estimate the aging degradation or the like of the
drive mechanism 26 or the like on the basis of a time for adjusting focus, that is, a time until the apparatus is restored to the initial state. -
FIG. 12 is a perspective view showing the appearance of an HMD according to another embodiment. TheHMD 200 is a glasses-type apparatus including a see-through display and attached to the head of the user to be used. TheHMD 200 has aframe 210, aleft lens 222 a, aright lens 222 b, a left-eye display 221 a, a right-eye display 221 b, a left--eye camera 223 a, and a right-eye camera 223 b. Further, a controller, a sensor unit, or the like substantially the same as that shown inFIG. 3 is constituted inside or at a prescribed position of theframe 210. - The
left lens 222 a and theright lens 222 b are arranged in front of the left eye and right eye of the user, respectively. The left-eye display 221 a and the right-eye display 221 b are provided in theleft lens 222 a and theright lens 222 b, respectively, so as to cover the visual field of the user. - The left-
eye display 221 a and the right-eye display 221 b are see-through displays and have a left-eye image and a right-eye image or the like displayed thereon, respectively. The user wearing theHMD 200 is allowed to visually recognize images displayed on the respective displays while visually recognizing an actual scene. Thus, the user is allowed to experience AR (Augmented Reality) or the like. - Note that dimmer elements (not shown) or the like may be provided on the outside (a side opposite to the eyes of the user) of the left-
eye display 221 a and the right-eye display 221 b, respectively. The dimmer elements are elements capable of adjusting an amount of light passing through the elements. With the provision of the dimmer elements, it is possible to control, for example, an actual scene passing through the respective displays and visually recognized by the user and cause the user to visually recognize images displayed on the respective displays in an emphasized manner. Thus, the user is allowed to experience VR (Virtual Reality) or the like. - As the left-
eye display 221 a and the right-eye display 221 b, see-through organic EL displays, LCDs (Liquid Crystal Displays), or the like are, for example, used. Further, as the dimmer elements, dimmer glasses, dimmer sheets, liquid crystal shutters, or the like capable of electrically controlling transmittance are, for example, used. In the present embodiment, an image display mechanism is realized by theleft lens 222 a, theright lens 222 b, the left-eye display 221 a, and the right-eye display 221 b. - The left-
eye camera 223 a and the right-eye camera 223 b are provided at arbitrary positions at which the left eye and right eye of the user are capable of being shot. On the basis of, for example, images of the left eye and the right eye shot by the left-eye camera 223 a and the right-eye camera 223 b, eye information regarding the eyes of the user is acquired. - In the
HMD 200 shown inFIG. 12 , each of theleft lens 222 a and theright lens 222 b is configured to be movable with respect to theframe 210 and moved by a drive mechanism. Further, theframe 210 itself is also configured to be movable and capable of changing its retention force. By appropriately changing the position or inclination of each of theleft lens 222 a and theright lens 222 b on the basis of eye information, it is possible to realize high-quality viewing experience like the above embodiments. - As described above, the present technology is also applicable to a see-through HMD such as AR glasses. For example, by changing the display positions of content images displayed on see-through displays on the basis of eye information during the experience of AR, it is also possible to correct the deviation between the real world and the content images with high accuracy. That is, in the AR, it is possible to change coordinate axes in a virtual world overlapping a real scene with high accuracy.
- With attention paid to this point, it may be said that the present technology is effective mainly for correcting the deviation of focus with respect to content in VR and is effective mainly for correcting the deviation of the position of content in AR. Of course, the present technology is not limited such objects.
- Fields to which the present technology is applicable are not limited. The present technology is applicable to, for example, arbitrary fields in which information is displayed in the view of the user in operation, such as video games, medical treatment (HMDs for surgeries or the like), motor sports (drivers' helmets or the like), sports including golf or the like, and examinations in factories or the like.
- As an embodiment of the information processing apparatus according to the present technology, the above description refers to an HMD that is an image display apparatus. However, the information processing apparatus according to the present technology may be realized by an arbitrary computer that is constituted separately from an HMD and connected to the HMD in a wired or wireless fashion. An information processing method according to the present technology may be performed by, for example, a cloud server. Alternatively, an HMD and other computers may work together to perform the information processing method according to the present technology.
- That is, the information processing method and a program according to the present technology are capable of being performed not only by a computer system constituted by a single computer but also by a computer system in which a plurality of computers works together to operate. Note that, a system refers to the aggregate of a plurality of constituents (such as apparatuses and modules (components)) and all the constituents are not necessarily accommodated in the same housing in the present disclosure. Accordingly, both of a plurality of apparatuses that is accommodated in separate housings and connected to each other via a network and one apparatus in which a plurality of modules is accommodated in one housing are systems.
- The execution of the information processing method and the program according to the present technology based on a computer system includes both a case in which the control or the like of, for example, the acquisition of eye information or the movement of at least a part of an image display mechanism is executed by a single computer and a case in which respective processing is executed by different computers. Further, the execution of respective processing by a prescribed computer includes causing other computers to execute a part or all of the processing and acquiring results.
- That is, the information processing method and the program according to the present technology is also applicable to the configuration of cloud computing in which one function is shared and cooperatively processed by a plurality of apparatuses via a network.
- Among the characteristic parts according to the present technology described above, at least two characteristic parts may be combined together. That is, the various characteristic parts described in the respective embodiments may be arbitrarily combined together without being distinguished between the respective embodiments. Further, the various effects described above are given only for an example and should not be interpreted in a limited way. Further, other effects may be produced.
- Note that the present technology may employ the following configurations.
- (1) An information processing apparatus including:
- an acquisition unit that acquires eye information regarding an eye of a user; and
- a movement unit that moves, on the basis of the acquired eye information, at least a part of an image display mechanism that emits and guides image light to the eye of the user.
- (2) The information processing apparatus according to (1), in which
- the image display mechanism has a display that emits the image light and a lens system that guides the image light emitted from the display to the eye of the user, and
- the movement unit moves at least one of the display or the lens system.
- (3) The information processing apparatus according to (1) or (2), in which
- the eye information includes an eye image obtained by shooting the eye of the user.
- (4) The information processing apparatus according to (3), in which
- the eye information includes at least one of a shape, a size, a position, an inclination, or an iris pattern of the eye of the user in the eye image.
- (5) The information processing apparatus according to (3) or (4), in which
- the image display mechanism is provided in an HMD (Head Mount Display) wearable by the user, and
- the eye information includes the eye image obtained by shooting the eye of the user with an imaging mechanism of the HMD.
- (6) The information processing apparatus according to any one of (1) to (5), further including:
- a storage unit that stores reference eye information representing the eye information obtained when the eye of the user is put in a prescribed state with respect to the image display mechanism, and
- the movement unit moves at least a part of the image display mechanism on the basis of the acquired eye information and the reference eye information.
- (7) The information processing apparatus according to (6), in which
- the prescribed state includes a state in which a reference image displayed by the image display mechanism is arranged at a prescribed size and a prescribed position with respect to the eye of the user inside the eye image.
- (8) The information processing apparatus according to (6) or (7), in which
- the eye information includes an eye image obtained by shooting the eye of the user,
- the reference eye information includes a reference eye image obtained by shooting the eye of the user when the eye of the user is put in the prescribed state with respect to the image display mechanism, and
- the movement unit moves at least a part of the image display mechanism on the basis of a difference between the eye image and the reference eye image.
- (9) The information processing apparatus according to (8), in which
- the movement unit moves at least a part of the image display mechanism such that a state of the eye of the user with respect to the image display mechanism comes close to the prescribed state.
- (10) The information processing apparatus according to (8) or (9), in which
- the movement unit moves at least a part of the image display mechanism such that a state of the eye of the user with respect to the image display mechanism is put in another state different from the prescribed state.
- (11) The information processing apparatus according to any one of (1) to (10), further including
- a display control unit that controls an image display by the image display mechanism, in which
- the display control unit moves, on the basis of the acquired eye information, a display position of an image displayed by the image display mechanism.
- (12) The information processing apparatus according to any one of (1) to (11), in which
- the movement unit moves at least a part of the image display mechanism on the basis of details of content displayed by the image display mechanism.
- (13) The information processing apparatus according to any one of (1) to (12), in which
- the movement unit moves at least a part of the image display mechanism on the basis of a viewing time of the user.
- (14) The information processing apparatus according to any one of (1) to (13), further including
- a state acquisition unit that acquires state information regarding a state of the user, in which
- the movement unit moves at least a part of the image display mechanism on the basis of the acquired state information.
- (15) The information processing apparatus according to any one of (1) to (14), in which
- the image display mechanism is provided in an HMD (Head Mount Display) wearable by the user, and
- the information processing apparatus further includes a determination unit that determines reliability of a detection result by a biosensor on the basis of a movement amount by the movement unit with respect to at least a part of the image display mechanism.
- (16) The information processing apparatus according to any one of (1) to (15), in which
- the image display mechanism is provided in an HMD (Head Mount Display) wearable by the user, and
- the information processing apparatus further includes an attachment state control unit capable of controlling an attachment state of the HMD on the basis of the acquired eye information.
- (17) The information processing apparatus according to any one of (1) to (16), further including
- a notification unit that notifies the user of prescribed information on the basis of the acquired eye information.
- (18) The information processing apparatus according to any one of (1) to (17), in which
- the eye information includes a left-eye image obtained by shooting a left eye of the user and a right-eye image obtained by shooting a right eye of the user, and
- the movement unit moves at least a part of the image display mechanism on the basis of the left-eye image or the right-eye image.
- (19) An information processing method performed by a computer system, the information processing method including:
- acquiring eye information regarding an eye of a user; and
- moving, on the basis of the acquired eye information, at least a part of an image display mechanism that emits and guides image light to the eye of the user.
- (20) A program causing a computer system to perform:
- a step of acquiring eye information regarding an eye of a user; and
- a step of moving, on the basis of the acquired eye information, at least a part of as image display mechanism that emits and guides image light to the eye of the user.
-
- 1(1 a, 1 b) eye of user
- 11 attachment band unit
- 12 display
- 13 display unit
- 15 imaging mechanism
- 17 temporal band
- 18 parietal band
- 20 image display mechanism
- 21 display
- 22(22 a, 22 b) lens system
- 23 a, 223 a left-eye camera
- 23 b, 223 b right-eye camera
- 25 band adjustment mechanism
- 26 drive mechanism
- 30 storage unit
- 31 sensor unit
- 32 controller
- 38 biosensor
- 55 eye image
- 56 eye region
- 57 long axis
- 60 reference image
- 100, 200 HMD
- 221 a left-eye display
- 221 b right-eye display
- 222 a left lens
- 222 b right lens
Claims (20)
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JPJP2017-199080 | 2017-10-13 | ||
JP2017199080A JP2019074582A (en) | 2017-10-13 | 2017-10-13 | Information processing apparatus, information processing method, and program |
JP2017-199080 | 2017-10-13 | ||
PCT/JP2018/030554 WO2019073689A1 (en) | 2017-10-13 | 2018-08-17 | Information processing device, information processing method, and program |
Publications (2)
Publication Number | Publication Date |
---|---|
US20200285310A1 true US20200285310A1 (en) | 2020-09-10 |
US11157078B2 US11157078B2 (en) | 2021-10-26 |
Family
ID=66101352
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/753,880 Active US11157078B2 (en) | 2017-10-13 | 2018-08-17 | Information processing apparatus, information processing method, and program |
Country Status (5)
Country | Link |
---|---|
US (1) | US11157078B2 (en) |
EP (1) | EP3697086A4 (en) |
JP (1) | JP2019074582A (en) |
CN (1) | CN111213375B (en) |
WO (1) | WO2019073689A1 (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11030926B2 (en) * | 2019-07-18 | 2021-06-08 | Samsung Electronics Co., Ltd. | Image display apparatus capable of multi-depth expression |
CN114280778A (en) * | 2021-11-24 | 2022-04-05 | 歌尔光学科技有限公司 | Intelligent glasses and image distance adjusting method thereof |
TWI828418B (en) * | 2022-11-11 | 2024-01-01 | 宏達國際電子股份有限公司 | Head-mounted display device and adjustment module |
CN117452644A (en) * | 2023-11-01 | 2024-01-26 | 华瀚智盛(北京)科技有限公司 | Head-mounted equipment |
WO2024176154A3 (en) * | 2023-02-22 | 2024-10-10 | Augmedics Ltd. | Head-mounted stereoscopic display device with digital loupes and associated methods |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR102709979B1 (en) * | 2021-10-26 | 2024-09-26 | 옵티시스 주식회사 | Head mounted display |
Family Cites Families (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0749464A (en) * | 1993-08-03 | 1995-02-21 | Keibunshiya:Kk | Head installation type display device |
JPH07154829A (en) * | 1993-11-25 | 1995-06-16 | Matsushita Electric Ind Co Ltd | Spectacles video display device |
JP3348956B2 (en) | 1994-03-24 | 2002-11-20 | ソニー株式会社 | Display device |
JP2000131643A (en) * | 1998-10-28 | 2000-05-12 | Sony Corp | Video display device |
JP2001189902A (en) * | 1999-12-28 | 2001-07-10 | Nec Corp | Method for controlling head-mounted display and head- mounted display device |
JP2002031776A (en) * | 2000-07-14 | 2002-01-31 | Canon Inc | Display device |
JP2004304296A (en) * | 2003-03-28 | 2004-10-28 | Nikon Corp | Head mount display |
JP4298455B2 (en) * | 2003-09-30 | 2009-07-22 | キヤノン株式会社 | Scanning image display device |
JP4517664B2 (en) | 2004-02-12 | 2010-08-04 | ソニー株式会社 | Image processing apparatus and method, program, and recording medium |
JP4560368B2 (en) * | 2004-10-08 | 2010-10-13 | キヤノン株式会社 | Eye detection device and image display device |
JP2012194501A (en) * | 2011-03-18 | 2012-10-11 | Brother Ind Ltd | Head-mounted display and virtual image presentation method |
US20150097772A1 (en) * | 2012-01-06 | 2015-04-09 | Thad Eugene Starner | Gaze Signal Based on Physical Characteristics of the Eye |
US9122321B2 (en) * | 2012-05-04 | 2015-09-01 | Microsoft Technology Licensing, Llc | Collaboration environment using see through displays |
JP2014106445A (en) * | 2012-11-29 | 2014-06-09 | Toshiba Corp | Electronic apparatus, and display control method |
US10345903B2 (en) | 2013-07-30 | 2019-07-09 | Microsoft Technology Licensing, Llc | Feedback for optic positioning in display devices |
US20150145977A1 (en) | 2013-11-22 | 2015-05-28 | Samsung Display Co., Ltd. | Compensation technique for viewer position in autostereoscopic displays |
US10591735B2 (en) * | 2015-01-15 | 2020-03-17 | Sony Interactive Entertainment Inc. | Head-mounted display device and image display system |
JP2016191845A (en) | 2015-03-31 | 2016-11-10 | ソニー株式会社 | Information processor, information processing method and program |
JP6572600B2 (en) * | 2015-04-09 | 2019-09-11 | セイコーエプソン株式会社 | Information processing apparatus, information processing apparatus control method, and computer program |
WO2019171719A1 (en) * | 2018-03-05 | 2019-09-12 | 株式会社ワコム | Input device employing electronic pen |
-
2017
- 2017-10-13 JP JP2017199080A patent/JP2019074582A/en active Pending
-
2018
- 2018-08-17 CN CN201880065099.XA patent/CN111213375B/en active Active
- 2018-08-17 US US16/753,880 patent/US11157078B2/en active Active
- 2018-08-17 EP EP18865388.5A patent/EP3697086A4/en active Pending
- 2018-08-17 WO PCT/JP2018/030554 patent/WO2019073689A1/en unknown
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11030926B2 (en) * | 2019-07-18 | 2021-06-08 | Samsung Electronics Co., Ltd. | Image display apparatus capable of multi-depth expression |
CN114280778A (en) * | 2021-11-24 | 2022-04-05 | 歌尔光学科技有限公司 | Intelligent glasses and image distance adjusting method thereof |
TWI828418B (en) * | 2022-11-11 | 2024-01-01 | 宏達國際電子股份有限公司 | Head-mounted display device and adjustment module |
WO2024176154A3 (en) * | 2023-02-22 | 2024-10-10 | Augmedics Ltd. | Head-mounted stereoscopic display device with digital loupes and associated methods |
CN117452644A (en) * | 2023-11-01 | 2024-01-26 | 华瀚智盛(北京)科技有限公司 | Head-mounted equipment |
Also Published As
Publication number | Publication date |
---|---|
JP2019074582A (en) | 2019-05-16 |
US11157078B2 (en) | 2021-10-26 |
WO2019073689A1 (en) | 2019-04-18 |
CN111213375A (en) | 2020-05-29 |
EP3697086A1 (en) | 2020-08-19 |
EP3697086A4 (en) | 2020-12-09 |
CN111213375B (en) | 2022-08-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11157078B2 (en) | Information processing apparatus, information processing method, and program | |
US11061240B2 (en) | Head-mountable apparatus and methods | |
US10495885B2 (en) | Apparatus and method for a bioptic real time video system | |
US10310595B2 (en) | Information processing apparatus, information processing method, computer program, and image processing system | |
CA2984147C (en) | Privacy-sensitive consumer cameras coupled to augmented reality systems | |
US20210350762A1 (en) | Image processing device and image processing method | |
JP5580855B2 (en) | Obstacle avoidance device and obstacle avoidance method | |
US20190238827A1 (en) | Eye gaze responsive virtual reality headset | |
CA3040218C (en) | Apparatus and method for a bioptic real time video system | |
JP2008256946A (en) | Sickness prevention device for image display device | |
JPWO2019187808A1 (en) | Information processing equipment, information processing methods, and programs | |
US20160091717A1 (en) | Head-mounted display system and operation method thereof | |
US20210392318A1 (en) | Gaze tracking apparatus and systems | |
US11314327B2 (en) | Head mounted display and control method thereof | |
US11747897B2 (en) | Data processing apparatus and method of using gaze data to generate images | |
US11023041B1 (en) | System and method for producing images based on gaze direction and field of view | |
JP6103743B2 (en) | Display device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
AS | Assignment |
Owner name: SONY CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SAZUKA, NAOYA;YOSHIKAWA, KIYOSHI;YOKOYAMA, MASAYUKI;SIGNING DATES FROM 20200403 TO 20200622;REEL/FRAME:053124/0345 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |