US20160252956A1 - Imaging Method - Google Patents
Imaging Method Download PDFInfo
- Publication number
- US20160252956A1 US20160252956A1 US15/096,181 US201615096181A US2016252956A1 US 20160252956 A1 US20160252956 A1 US 20160252956A1 US 201615096181 A US201615096181 A US 201615096181A US 2016252956 A1 US2016252956 A1 US 2016252956A1
- Authority
- US
- United States
- Prior art keywords
- hmd
- wearer
- eye
- particular direction
- gaze
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000003384 imaging method Methods 0.000 title description 8
- 210000001747 pupil Anatomy 0.000 claims abstract description 42
- 230000033001 locomotion Effects 0.000 claims abstract description 37
- 238000000034 method Methods 0.000 claims description 38
- 230000006870 function Effects 0.000 claims description 16
- 230000004044 response Effects 0.000 claims description 4
- 238000001514 detection method Methods 0.000 abstract description 4
- 230000003287 optical effect Effects 0.000 description 41
- 238000004891 communication Methods 0.000 description 11
- 210000003128 head Anatomy 0.000 description 10
- 230000003247 decreasing effect Effects 0.000 description 5
- 230000002093 peripheral effect Effects 0.000 description 5
- 230000000694 effects Effects 0.000 description 4
- 239000000463 material Substances 0.000 description 4
- 238000004091 panning Methods 0.000 description 4
- 239000004033 plastic Substances 0.000 description 4
- 229920003023 plastic Polymers 0.000 description 4
- 230000010287 polarization Effects 0.000 description 4
- 238000006073 displacement reaction Methods 0.000 description 3
- 238000003708 edge detection Methods 0.000 description 3
- 230000003993 interaction Effects 0.000 description 3
- 230000004434 saccadic eye movement Effects 0.000 description 3
- 239000010409 thin film Substances 0.000 description 3
- 230000003190 augmentative effect Effects 0.000 description 2
- 230000010267 cellular communication Effects 0.000 description 2
- 238000013500 data storage Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000004424 eye movement Effects 0.000 description 2
- 239000011521 glass Substances 0.000 description 2
- 239000007788 liquid Substances 0.000 description 2
- 239000002184 metal Substances 0.000 description 2
- 229920003229 poly(methyl methacrylate) Polymers 0.000 description 2
- 239000004926 polymethyl methacrylate Substances 0.000 description 2
- 210000003786 sclera Anatomy 0.000 description 2
- 229910052710 silicon Inorganic materials 0.000 description 2
- 239000010703 silicon Substances 0.000 description 2
- 238000012935 Averaging Methods 0.000 description 1
- HBBGRARXTFLTSG-UHFFFAOYSA-N Lithium ion Chemical compound [Li+] HBBGRARXTFLTSG-UHFFFAOYSA-N 0.000 description 1
- 102220616555 S-phase kinase-associated protein 2_E48R_mutation Human genes 0.000 description 1
- 206010053694 Saccadic eye movement Diseases 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 230000000454 anti-cipatory effect Effects 0.000 description 1
- 238000000576 coating method Methods 0.000 description 1
- 150000001925 cycloalkenes Chemical class 0.000 description 1
- 210000005069 ears Anatomy 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 230000000193 eyeblink Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 239000010408 film Substances 0.000 description 1
- 238000009501 film coating Methods 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 229910001416 lithium ion Inorganic materials 0.000 description 1
- 230000003340 mental effect Effects 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 229920000642 polymer Polymers 0.000 description 1
- 230000001711 saccadic effect Effects 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 230000002463 transducing effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/012—Head tracking input arrangements
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/0093—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0172—Head mounted characterised by optical features
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0176—Head mounted characterised by mechanical features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0346—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04817—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance using icons
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/0482—Interaction with lists of selectable items, e.g. menus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0138—Head-up displays characterised by optical features comprising image capture systems, e.g. camera
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/014—Head-up displays characterised by optical features comprising information/image processing systems
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B2027/0178—Eyeglass type
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0179—Display position adjusting means not related to the information to be displayed
- G02B2027/0187—Display position adjusting means not related to the information to be displayed slaved to motion of at least a part of the body of the user, e.g. head, eye
Definitions
- Wearable systems can integrate various elements, such as miniaturized computers, input devices, sensors, detectors, image displays, wireless communication devices as well as image and audio processors, into a device that can be worn by a user.
- Such devices provide a mobile and lightweight solution to communicating, computing and interacting with one's environment.
- wearable compact optical displays that augment the wearer's experience of the real world.
- an artificial image By placing an image display element close to the wearer's eye(s), an artificial image can be made to overlay the wearer's view of the real world.
- image display elements are incorporated into systems also referred to as “near-eye displays”, “head-mounted displays” (HMDs) or “heads-up displays” (HUDs).
- HMDs head-mounted displays
- HUDs heads-up displays
- the artificial image may fill or nearly fill the wearer's field of view.
- a head-mounted display In a first aspect, a head-mounted display (HMD) is provided.
- the HMD includes a head-mounted support, an optical system, an infrared light source, a camera, and a computer.
- the optical system is attached to the head-mounted support and includes a display panel configured to generate a virtual image, wherein the virtual image is viewable from a viewing location.
- the infrared light source is configured to illuminate the viewing location with infrared light such that infrared light is reflected from the viewing location as reflected infrared light and the camera is configured to image the viewing location by collecting the reflected infrared light.
- the computer is configured to determine a gaze axis based on one or more images of the viewing location obtained by the camera and control the display panel to move the virtual images within a field of view based on the gaze axis, a reference axis related to the HMD, and a tracking rate.
- a method in a second aspect, includes determining a gaze axis within a field of view of a head-mounted display (HMD), wherein the HMD is configured to display virtual images within the field of view.
- the method further includes determining a reference axis related to the HMD, adjusting a tracking rate based on the gaze axis and the reference axis, and moving the virtual images within the field of view based on the gaze axis, the reference axis and the tracking rate.
- HMD head-mounted display
- a non-transitory computer readable medium includes instructions executable by a computing device to cause the computing device to perform functions including, receiving eye-tracking images from a head-mounted display (HMD), wherein the HMD is configured to display virtual images within a field of view.
- the non-transitory computer readable medium further includes determining a gaze axis from the eye-tracking images, determining a reference axis related to the HMD, calculating an angle difference between the gaze axis and the reference axis, and adjusting a tracking rate based on the angle difference.
- the non-transitory computer readable medium further includes controlling the HMD to display the virtual images based upon the gaze axis, the reference axis, and the tracking rate.
- FIG. 1 is schematic diagram of a wearable computing device, in accordance with an example embodiment.
- FIG. 2 is a top view of an optical system, in accordance with an example embodiment.
- FIG. 3A is a front view of a head-mounted display, in accordance with an example embodiment.
- FIG. 3B is a top view of the head-mounted display of FIG. 3A , in accordance with an example embodiment.
- FIG. 3C is a side view of the head-mounted display of FIG. 3A and FIG. 3B , in accordance with an example embodiment.
- FIG. 4A is a side view of a head-mounted display with a forward gaze axis, in accordance with an example embodiment.
- FIG. 4B is a side view of the head-mounted display of FIG. 4A with an upward gaze axis, in accordance with an example embodiment.
- FIG. 5 is a flowchart of a method, in accordance with an example embodiment.
- FIG. 6A is a field of view of a head-mounted display showing scrolling text, in accordance with an example embodiment.
- FIG. 6B is a field of view of a head-mounted display showing scrolling text, in accordance with an example embodiment.
- FIG. 6C is a field of view of a head-mounted display showing scrolling text, in accordance with an example embodiment.
- FIG. 6D is a field of view of a head-mounted display showing scrolling text, in accordance with an example embodiment.
- FIG. 7A is an overhead view of HMD user traveling on a subway, in accordance with an example embodiment.
- FIG. 7B is a field of view of a HMD user interface, in accordance with an example embodiment.
- FIG. 7C is a field of view of a HMD user interface, in accordance with an example embodiment.
- FIG. 7D is a field of view of a HMD user interface, in accordance with an example embodiment.
- FIG. 7E is a field of view of a HMD user interface, in accordance with an example embodiment.
- a head-mounted display may enable its wearer to observe the wearer's real-world surroundings and also view a displayed image, such as a computer-generated image.
- the displayed image may overlay a portion of the wearer's field of view of the real world.
- the wearer of the HMD is going about his or her daily activities, such as walking, driving, exercising, etc., the wearer may be able to see a displayed image generated by the HMD at the same time that the wearer is looking out at his or her real-world surroundings.
- the displayed image also known as a virtual image, might include, for example, graphics, text, and/or video.
- the content of the displayed image could relate to any number of contexts, including but not limited to the wearer's current environment, an activity in which the wearer is currently engaged, the biometric status of the wearer, and any audio, video, or textual communications that have been directed to the wearer.
- the images displayed by the HMD may also be part of an interactive user interface.
- the HMD could be part of a wearable computing device.
- the images displayed by the HMD could include menus, selection boxes, navigation icons, or other user interface features that enable the wearer to invoke functions of the wearable computing device or otherwise interact with the wearable computing device.
- the images displayed by the HMD could appear anywhere in the wearer's field of view.
- the displayed image might occur at or near the center of the wearer's field of view, or the displayed image might be confined to the top, bottom, or a corner of the wearer's field of view.
- the displayed image might be at the periphery of or entirely outside of the wearer's normal field of view.
- the displayed image might be positioned such that it is not visible when the wearer looks straight ahead but is visible when the wearer looks in a specific direction, such as up, down, or to one side.
- the displayed image might overlay only a small portion of the wearer's field of view, or the displayed image might fill most or all of the wearer's field of view.
- the displayed image could be displayed continuously or only at certain times (e.g., only when the wearer is engaged in certain activities).
- the displayed images may appear fixed relative to the wearer's environment.
- the virtual images may appear anchored to a particular object or location within the wearer's environment.
- displayed images may appear fixed relative to the wearer's field of view.
- the HMD may include a graphical user interface that may stay substantially anchored to the wearer's field of view regardless of the HMD orientation. In practice, both types of virtual imagery may be implemented together in an HMD.
- an optical system in the HMD may include a light source, such as a light-emitting diode (LED), that is configured to illuminate a display panel, such as a liquid crystal-on-silicon (LCOS) display.
- a light source such as a light-emitting diode (LED)
- LCOS liquid crystal-on-silicon
- the display panel generates light patterns by spatially modulating the light from the light source, and the light patterns may be viewable as virtual images at a viewing location.
- the HMD may obtain data from the wearer in order to perform certain functions, for instance to provide context-sensitive information to the wearer.
- the HMD may obtain information regarding the wearer and the wearer's environment and respond accordingly.
- the HMD may use a pupil position recognition technique, wherein if the HMD recognizes that the wearer's pupil location, and thus a corresponding gaze axis, is inclined with respect to a reference axis, the HMD may display virtual images related to objects located above the wearer.
- the HMD may recognize, by a similar pupil position recognition technique, that the wearer is looking downward. Accordingly, the HMD may display virtual images related to objects located below a reference axis of the wearer.
- the wearer's pupil may be illuminated by an infrared light source or multiple infrared light sources.
- An infrared camera may image the pupil.
- the infrared light source(s) could be located in the HMD optical path, or could alternatively be located off-axis.
- the infrared camera could also be located in the HMD optical path or off-axis.
- Possible eye tracking modalities that could be used include dark pupil imaging and dual-glint Purkinje image tracking, among other techniques known in the art.
- a processor may implement an image processing algorithm to find the edges or extents of the imaged pupil.
- the image processing algorithms may include pattern recognition, Canny edge detection, thresholding, contrast detection, or differential edge detection, to name a few. Those skilled in the art will understand that a variety of different image processing techniques could be used individually or in combination with other methods in order to obtain pupil location.
- the processor may determine a gaze axis, which may be defined as an axis extending from a viewing location and through a gaze point located within the wearer's field of view.
- the processor may also determine a reference axis, which may be defined as an axis extending from a viewing location and through a point in space.
- the point in space may include the apparent center of the display of the HMD or a target object, among other possibilities.
- the processor may act to adjust various components of the displayed virtual image based on an angle difference between the gaze axis and the reference axis and a tracking rate. For example, an upward scrolling list of text information may be presented to a user of an HMD, similar to the traditional display of credits at the end of a movie.
- the reference axis may include an axis that extends through the apparent center of the HMD display. If a gaze axis is determined that indicates the user is gazing down near the bottom of the display with respect to the reference axis, the tracking rate may be increased such that the rate of upward scrolling is increased, providing more text to the user.
- the tracking rate may be decreased such that the rate of upward scrolling is slowed, allowing the user to catch up in reading the text.
- the tracking rate may be decreased or set to zero (eliminating scrolling of the virtual image). In this respect, a user may be able to focus his/her attention on specific text.
- FIG. 1 is schematic diagram of a wearable computing device or a head-mounted display (HMD) 100 that may include several different components and subsystems.
- HMD 100 includes an eye-tracking system 102 , a HMD-tracking system 104 , an optical system 106 , peripherals, a power supply 110 , a processor 112 , a memory 114 , and a user interface 115 .
- the eye-tracking system 102 may include hardware such as an infrared camera 116 and at least one infrared light source 118 .
- the HMD-tracking system 104 may include a gyroscope 120 , a global positioning system (GPS) 122 , and an accelerometer 124 .
- GPS global positioning system
- the optical system 106 may include, in one embodiment, a display panel 126 , a display light source 128 , and optics 130 .
- Peripherals 108 may include, for example, a wireless communication interface 134 , a touchpad 136 , a microphone 138 , a camera 140 , and a speaker 142 .
- HMD 100 includes a see-through display.
- the wearer of HMD 100 may observe a portion of the real-world environment, i.e., in a particular field of view provided by the optical system 106 .
- HMD 100 is operable to display virtual images that are superimposed on the field of view, for example, to provide an “augmented reality” experience. Some of the virtual images displayed by HMD 100 may be superimposed over particular objects in the field of view. HMD 100 may also display images that appear to hover within the field of view instead of being associated with particular objects in the field of view.
- Components of the HMD 100 may be configured to work in an interconnected fashion with other components within or outside their respective systems.
- the infrared camera 116 may image one or both of the HMD wearer's eyes.
- the infrared camera 116 may deliver image information to the processor 112 , which may access the memory 112 and make a determination regarding the direction of the HMD wearer's gaze, also termed a gaze axis.
- the processor 112 may further accept input from the GPS unit 122 , the gyroscope 120 , and/or the accelerometer 124 to determine the location and orientation of the HMD 100 .
- the processor 112 may control the user interface 115 and the display panel 126 to display virtual images to the HMD wearer that may include context-specific information based on the HMD location and orientation as well as the HMD wearer's gaze axis.
- HMD 100 could be configured as, for example, eyeglasses, goggles, a helmet, a hat, a visor, a headband, or in some other form that can be supported on or from the wearer's head. Further, HMD 100 may be configured to display images to both of the wearer's eyes, for example, using two see-through displays. Alternatively, HMD 100 may include only a single see-through display and may display images to only one of the wearer's eyes, either the left eye or the right eye. In other embodiments, HMD 100 may include an opaque display configured to display images to a first eye or both eyes of the HMD wearer. In embodiments where an opaque display is presented to the first eye of the HMD wearer, a view of the real-world environment could be available to a second eye of the HMD wearer.
- a power supply 110 may provide power to various HMD components and could represent, for example, a rechargeable lithium-ion battery.
- Various other power supply materials and types known in the art are possible.
- the function of the HMD 100 may be controlled by a processor 112 that executes instructions stored in a non-transitory computer readable medium, such as the memory 114 .
- processor 112 in combination with instructions stored in the memory 114 may function as a controller of HMD 100 .
- processor 112 may control the user interface 115 to adjust what images are displayed by HMD 100 .
- the processor 112 may also control the wireless communication interface 134 and various other components of the HMD 100 .
- the processor 112 may additionally represent a plurality of computing devices that may serve to control individual components or subsystems of the HMD 100 .
- the memory 114 may store data that may include a set of calibrated wearer eye pupil positions and a collection of past eye pupil positions.
- the memory 114 may function as a database of information related to gaze direction. Such information may be used by HMD 100 to anticipate where the user will look and determine what images are to be displayed to the wearer.
- Calibrated wearer eye pupil positions may include, for instance, information regarding the extents or range of the wearer's eye pupil movement (right/left and upwards/downwards) as well as wearer eye pupil positions that may relate to various reference axes.
- Reference axes could represent, for example, an axis extending from a viewing location and through a target object or the apparent center of a field of view (i.e. the reference axis may correspond to a center line of the field of view). Other possibilities for reference axes exist. Thus, a reference axis may further represent a basis for determining dynamic gaze direction.
- control instructions may be enacted using eye movements. For instance, two consecutive wearer eye blinks may represent a control instruction directing the HMD 100 to capture an image with a peripheral camera 140 .
- Control instructions could also include the dwell-based selection of a target object. For instance, if a wearer fixates visually upon a particular virtual image or real-world object for longer than a predetermined time period, a control instruction may be generated to select the virtual image or real-world object as a target object. Many other control instructions are possible.
- memory 114 could store various recorded data from previous HMD/user interactions. For instance, multiple images of a HMD wearer's eye(s) could be averaged to obtain an averaged eye gaze axis. This could lessen the effect of saccadic eye movements or saccades, in which the eye moves in a rapid and somewhat random manner around an eye gaze axis. These saccades help humans build up a mental image of a field of view with better resolution than if the eye remained static, and by averaging a number of eye images within a particular time period, an average gaze axis could be determined with less saccadic ‘noise’.
- memory 114 could store recorded data regarding recent eye gaze axes for various application-based functions.
- the recent variance of the eye gaze axis could be coupled to scrolling images generated by the HMD 100 .
- the images e.g. text or other images
- the eye gaze axis variance is low, the images may scroll slower or stop altogether.
- a lower variance in eye gaze axis could indicate the HMD wearer is concentrating on one particular gaze location, whereas a higher eye gaze axis variance means the opposite—the HMD wearer may be quickly scanning a document and desire a faster scrolling speed.
- the variance may differ depending on the axis along which it is measured. For example, the horizontal variance of a HMD wearer's eye gaze may be high while the vertical variance may be relatively low. This could indicate to the HMD 100 that the wearer is reading text. Accordingly, text scrolling/tracking could be adjusted in a different or more controlled fashion compared to ‘non-reading’ scrolling/panning/pagination situations.
- the HMD 100 may include a user interface 115 for providing information to the wearer or receiving input from the wearer.
- the user interface 115 could be associated with, for example, the displayed virtual images, a touchpad, a keypad, buttons, a microphone, and/or other peripheral input devices.
- the processor 112 may control the functioning of the HMD 100 based on input received through the user interface 115 . For example, the processor 112 may utilize user input from the user interface 115 to control how the HMD 100 displays images within a field of view or determine what images the HMD 100 displays.
- the infrared camera 116 may be utilized by the eye-tracking system 102 to capture images of a viewing location associated with the HMD 100 .
- the infrared camera 116 may image the eye of a HMD wearer that may be located at the viewing location.
- the images could be either video images or still images.
- the images obtained by the infrared camera 116 regarding the HMD wearer's eye may help determine where the wearer is looking within the HMD field of view, for instance by ascertaining the location of the HMD wearer's eye pupil.
- Analysis of the images obtained by the infrared camera 116 could be performed by the processor 112 in conjunction with the memory 114 .
- the imaging of the viewing location could occur continuously or at discrete times depending upon, for instance, user interactions with the user interface 115 .
- the infrared camera 116 could be integrated into the optical system 106 .
- the infrared camera 116 could be mounted separately from the optical system 106 and/or HMD 100 .
- the infrared camera 116 could additionally represent a visible light camera with sensing capabilities in the infrared wavelengths.
- the infrared light source 118 could represent one or more infrared light-emitting diodes (LEDs) or infrared laser diodes that may illuminate a viewing location. Thus, one or both eyes of a wearer of the HMD 100 may be illuminated by the infrared light source 118 .
- the infrared light source 118 may be positioned along an optical axis common to the infrared camera, and/or the infrared light source 118 may be positioned elsewhere.
- the infrared light source 118 could be mounted separately from the optical system 106 and/or HMD 100 .
- the infrared light source 118 may illuminate the viewing location continuously or may be turned on at discrete times. Additionally, when illuminated, the infrared light source 118 may be modulated at a particular frequency.
- the HMD-tracking system 104 could be configured to provide a HMD position and HMD orientation to the processor 112 .
- This position and orientation data may help determine a reference axis to which a gaze axis is compared.
- the reference axis may correspond to the orientation of the HMD.
- the gyroscope 120 could be a microelectromechanical system (MEMS) gyroscope or a fiber optic gyroscope.
- the gyroscope 120 may be configured to provide orientation information to the processor 112 .
- the GPS unit 122 could be a receiver that obtains clock and other signals from GPS satellites and may be configured to provide real-time location information to the processor 112 .
- the HMD-tracking system 104 could further include an accelerometer 124 configured to provide motion input data to the processor 112 .
- the optical system 106 could represent components configured to provide virtual images to a viewing location.
- An example of optical system 106 is described in detail below.
- the HMD 100 may include a wireless communication interface 134 for wirelessly communicating with one or more devices directly or via a communication network.
- wireless communication interface 134 could use 3G cellular communication, such as CDMA, EVDO, GSM/GPRS, or 4G cellular communication, such as WiMAX or LTE.
- wireless communication interface 134 could communicate with a wireless local area network (WLAN), for example, using WiFi.
- WLAN wireless local area network
- wireless communication interface 134 could communicate directly with a device, for example, using an infrared link, Bluetooth, or ZigBee.
- FIG. 1 shows various components of the HMD 100 (i.e., wireless communication interface 134 , processor 112 , memory 114 , infrared camera 116 , display panel 126 , GPS 122 , and user interface 115 ) as being integrated into HMD 100
- one or more of these components could be physically separate from HMD 100 .
- infrared camera 116 could be mounted on the wearer separate from HMD 100 .
- the HMD 100 could be part of a wearable computing device in the form of separate devices that can be worn on or carried by the wearer.
- the separate components that make up the wearable computing device could be communicatively coupled together in either a wired or wireless fashion.
- FIG. 2 illustrates a top view of an optical system 200 that is configured to display a virtual image superimposed upon a real-world scene viewable along a viewing axis 204 .
- a distal portion 232 and a proximal portion 234 represent optically-coupled portions of the optical system 200 that may or may not be physically separated.
- An example embodiment includes a display panel 206 that may be illuminated by a light source 208 . Light emitted from the light source 208 is incident upon the distal beam splitter 210 .
- the light source 208 may include one or more light-emitting diodes (LEDs) and/or laser diodes.
- the light source 208 may further include a linear polarizer that acts to pass one particular polarization to the rest of the optical system.
- the distal beam splitter 210 is a polarizing beam splitter that reflects light depending upon the polarization of light incident upon the beam splitter.
- s-polarized light from the light source 208 may be preferentially reflected by a distal beam-splitting interface 212 towards the display panel 206 .
- the display panel 206 in the example embodiment is a liquid crystal-on-silicon (LCOS) display, but could also be a digital light projector (DLP) micro-mirror display, or other type of reflective display panel.
- the display panel 206 acts to spatially-modulate the incident light to generate a light pattern.
- the display panel 206 may be an emissive-type display such as an organic light-emitting diode (OLED) display.
- OLED organic light-emitting diode
- the display panel 206 In the example in which the display panel 206 is a LCOS display panel, the display panel 206 generates a light pattern with a polarization perpendicular to the polarization of light initially incident upon the panel. In this example embodiment, the display panel 206 converts incident s-polarized light into a light pattern with p-polarization. The generated light pattern from the display panel 206 is directed towards the distal beam splitter 210 . The p-polarized light pattern passes through the distal beam splitter 210 and is directed along an optical axis 214 towards the proximal region of the optical system 200 . In an example embodiment, the proximal beam splitter 216 is also a polarizing beam splitter.
- image former 218 includes a concave mirror 230 and a proximal quarter-wave plate 228 .
- the light pattern passes through the proximal quarter-wave plate 228 and is reflected by the concave mirror 230 .
- the reflected light pattern passes back through proximal quarter-wave plate 228 .
- the light patterns are converted to the s-polarization and are formed into a viewable image.
- This viewable image is incident upon the proximal beam splitter 216 and the viewable image is reflected from proximal beam splitting interface 220 towards a viewing location 222 along a viewing axis 204 .
- a real-world scene is viewable through a viewing window 224 .
- the viewing window 224 may include a linear polarizer in order to reduce stray light within the optical system. Light from the viewing window 224 is at least partially transmitted through the proximal beam splitter 216 .
- both a virtual image and a real-world image are viewable to the viewing location 222 through the proximal beam splitter 216 .
- FIG. 2 depicts the distal portion 232 of the optical system housing as to the left of the proximal portion 234 of the optical system housing when viewed from above, it is understood that other embodiments are possible to physically realize the optical system 200 , including the distal portion 232 being configured to be to the right, below and above with respect to the proximal portion 234 .
- an image former 218 as comprising a concave mirror 230 , it is understood by those skilled in the art that the image former 218 may comprise a different optical element, such as an optical lens or a diffractive optic element.
- the proximal beam splitter 216 , the distal beam splitter 210 , and other components of optical system 200 are made of glass. Alternatively, some or all of such optical components may be partially or entirely plastic, which can also function to reduce the weight of optical system 200 .
- a suitable plastic material is Zeonex® E48R cyclo olefin optical grade polymer which is available from Zeon Chemicals L.P., Louisville, Ky. Another suitable plastic material is polymethyl methacrylate (PMMA).
- An example embodiment may include an infrared light source 226 that is configured to illuminate the viewing location 222 .
- FIG. 2 depicts the infrared light source 226 as adjacent to viewing window 224 , those skilled in the art will understand that the infrared light source 226 could be located elsewhere, such as on the side of the proximal beam splitter 216 that is adjacent to the viewing location 222 or in the distal portion 232 of the optical system 200 .
- the infrared light source 226 may represent, for example, one or more infrared light-emitting diodes (LEDs). Infrared LEDs with a small size may be implemented, such as the Vishay Technology TSML 1000 product.
- LEDs infrared light-emitting diodes
- the infrared light source 226 may include one or more LEDs located at different locations in the optical system 200 .
- Infrared light generated from the infrared light source 226 is configured to be incident upon the viewing location 222 .
- the wearer's eye pupil may be illuminated with the infrared light.
- the infrared light may be reflected from the wearer's eye back along the viewing axis 204 towards the proximal beam splitter 216 .
- a portion of the reflected infrared light may be reflected from the beam splitting interface 220 towards the image former 218 .
- the image former 218 may include a dichroic thin film configured to selectively reflect or transmit incident light depending upon the wavelength of the incident light.
- the dichroic thin film may be configured to pass infrared light while reflecting visible light.
- the visible light pattern generated by the display panel 206 may be reflected by the concave mirror 230 and the visible light pattern may be formed into a viewable image.
- the infrared light may thus be preferably transmitted through the concave mirror 230 to infrared camera 202 .
- Dichroic thin film coatings are available commercially from companies such as JML Optical Industries and Precision Glass & Optics (PG&O) and comprise multiple layers of dielectric and/or metal films. These dichroic coatings are also called ‘cold mirrors’.
- a small aperture or apertures may be introduced into the image former 218 , which may be realized by one or more pinholes in the concave mirror 230 .
- most of the visible and infrared light is reflected off of and formed by the image former 218 into an image viewable by the HMD wearer. Some of the visible and infrared light passes through the aperture and is incident upon the infrared camera 202 .
- the infrared camera 202 may selectively filter and detect the infrared light from the combination of visible and infrared light to obtain information regarding the wearer's eye pupil location.
- the infrared light source 226 may be modulated to provide a frequency reference for a lock-in amplifier or phase-locked loop in order that the infrared light signal is obtained efficiently.
- the visible light source 208 may be modulated and infrared light detection could be performed when the visible light source 208 is off, for example.
- FIG. 3A presents a front view of a head-mounted display (HMD) 300 in an example embodiment that includes a head-mounted support 309 .
- FIGS. 3B and 3C present the top and side views, respectively, of the HMD in FIG. 3A .
- the head-mounted support 309 includes lens frames 314 and 316 , a center frame support 318 , lens elements 310 and 312 , and extending side-arms 320 and 322 .
- the center frame support 318 and side-arms 320 and 322 are configured to secure the head-mounted support 309 to the wearer's head via the wearer's nose and ears, respectively.
- Each of the frame elements 314 , 316 , and 318 and the extending side-arms 320 and 322 may be formed of a solid structure of plastic or metal, or may be formed of a hollow structure of similar material so as to allow wiring and component interconnects to be internally routed through the head-mounted support 309 .
- head-mounted support 309 may support external wiring.
- Lens elements 310 and 312 are at least partially transparent so as to allow the wearer to look through them.
- the wearer's left eye 308 may look through left lens 312 and the wearer's right eye 306 may look through right lens 310 .
- Optical systems 302 and 304 which may be configured as shown in FIG. 2 , may be positioned in front of lenses 310 and 312 , respectively, as shown in FIGS. 3A, 3B, and 3C .
- Optical systems 302 and 304 may be attached to the head-mounted support 309 using support mounts 324 and 326 , respectively.
- optical systems 302 and 304 may be integrated partially or completely into lens elements 310 and 312 , respectively.
- a HMD might include an optical system for only one of the wearer's eyes (either left eye 308 or right eye 306 ).
- the HMD wearer may simultaneously observe from optical systems 302 and 304 a real-world image with an overlaid virtual image.
- the HMD 300 may include various elements such as a processor 340 , a touchpad 342 , a microphone 344 , and a button 346 .
- the computer 340 may use data from, among other sources, various sensors and cameras to determine the virtual image that should be displayed to the user.
- an infrared light source or sources may illuminate the viewing position(s) 308 and 306 , i.e. the wearer's eye(s), and the reflected infrared light may be preferentially collected with an infrared camera.
- FIGS. 4A and 4B depict side and front views of an eye as well as schematic drawings of pupil location information under different conditions.
- One way to determine a gaze axis of a person is to ascertain the position of the person's eye pupil with respect to a reference point, such as a viewing location.
- a reference point such as a viewing location.
- infrared light is generally reflected off of a person's eye. The reflected light may be collected and detected with an infrared detector.
- image processing can be conducted with a processor 112 in order to determine, for instance, the extents and centroid location of the person's pupil.
- the other known means and methods of eye-tracking, including the use of visible light illumination and/or imaging techniques are possible.
- a person may be looking directly forward as depicted in FIG. 4A .
- the eye 412 is open and the pupil 418 is located along a reference axis 410 .
- image processing which may include edge detection
- the position of the pupil may be determined to be at pupil location 422 .
- the processor 112 may subsequently determine that the gaze axis based on the pupil location 422 coincides with a reference axis 410 .
- Virtual image display position and movement may be adjusted due to the determined pupil location 422 .
- the processor 112 may adjust a tracking rate to zero when the gaze axis and the reference axis are equivalent or nearly equivalent. This may allow a user to slowly read critical text or closely examine a virtual image, for example.
- a person may be looking upwards with respect to a reference axis 428 .
- the eye 434 is open and the pupil location is generally higher than a reference point 440 .
- imaging the person's pupil 438 with infrared light may result in a determined pupil position 442 .
- the processor 112 may determine that the gaze axis 430 that is above the reference axis 428 .
- the angle difference 432 may represent the absolute difference in angle between the reference axis 428 and the gaze axis 430 .
- the processor 112 may calculate the angle difference 432 and, based on the angle difference 432 , adjust a tracking rate. For instance, a large angle difference 432 could represent an adjustment in tracking rate such that the tracking rate is higher, for instance to scroll a virtual image across a field of view at a faster rate.
- eye gaze determination techniques could include the use of different eye gaze determination techniques. For instance, instead of using the eye pupil to determine gaze axis, it is possible to track eye motions using the boundary between the sclera and iris ( 416 and 436 in FIGS. 4A and 4B ). For the purposes of determining an eye gaze axis, finding the centroid of the sclera/iris boundary may be equivalent to finding the centroid of a pupil.
- a method 500 is provided for adjusting virtual images within a field of view based on a gaze axis, a reference axis and a tracking rate.
- Method 500 could be performed using an HMD that is configured as shown in any of FIGS. 1-3C or configured in some other way.
- FIG. 5 illustrates the steps in an example method, however, it is understood that in other embodiments, the steps may appear in different order and steps may be added or subtracted.
- a gaze axis is determined within a field of view of a head-mounted display (HMD) (Step 502 ).
- the HMD is generally configured to display virtual images to be viewable at a viewing location and could be an HMD similar to an aforementioned embodiment.
- the gaze axis could be determined to be an axis extending from a center of the HMD wearer's pupil.
- the gaze axis could be similar to the gaze axis 430 depicted in FIG. 4B .
- the pupil location could be determined using eye glint images or by other eye-tracking techniques detailed above.
- a reference axis related to the HMD may also be determined in the method (Step 504 ).
- the reference axis could be determined by the processor 112 based on the orientation of the HMD and may correspond to the apparent center of the HMD field of view, for example. Other reference axes are possible.
- a tracking rate related to the movement of virtual images may be adjusted based on the gaze axis and the reference axis (Step 506 ).
- the tracking rate may be the rate at which virtual images are panning and/or scrolling across the HMD field of view.
- the tracking rate may depend upon the data that is displayed. For instance, words on a document may scroll in a vertical fashion from the bottom to the top of the HMD field of view to simulate reading downwards along a printed page. In this case, the display may move the text upwards at a tracking rate of around 1 second per line.
- a tracking rate could also be related to the motion of a HMD wearer. For instance, when displaying virtual images that may be user interface menus, for instance, the HMD may attempt to base the tracking rate of the virtual images on the rate of HMD movement. More specifically, an HMD wearer may access a user interface by changing the orientation of the HMD (for instance rotating one's head and/or body to select different elements of the user interface). Thus, the user interface could be at least partially anchored to locations and objects in the real world and the virtual images could be adjusted or panned at a rate proportional to the rate of change of the HMD orientation.
- the tracking rate could be adjusted by the angle difference between the gaze axis and the reference axis. For instance, the tracking rate could be increased if the angle difference between the gaze axis and the reference axis is large and the tracking rate could be decreased if the angle difference is small. The tracking rate could also be adjusted based on the direction of the vector between the reference axis and the gaze axis. More detailed examples are given below.
- the virtual images may be adjusted within the field of view based on the gaze axis, the reference axis, and the tracking rate (Step 508 ). If the tracking rate is adjusted lower, the movement rate of the virtual images may slow, for instance. Conversely, with a higher tracking rate, the virtual images may appear to move more quickly within the field of view.
- FIGS. 6A, 6B, 6C, and 6D illustrate an example in which a determined gaze axis controls text scrolling.
- virtual images including text are presented within a field of view 602 .
- the text 604 may be scrolling slowly upwards at a normal tracking rate similar in fashion to credits at the end of a movie (around one line per second, for instance).
- a gaze point 606 may be ascertained related to a gaze axis and thus to the position of an eye pupil of a wearer of an HMD.
- the reference axis may be considered as originating from the wearer's eye and going through the apparent center of the field of view 602 . When a wearer is reading normally and the angle between the reference axis and the gaze axis is relatively small, the tracking rate may stay unchanged and the text 604 may continue to scroll upwards.
- the wearer of the HMD may move his or her gaze point from a central location 610 to a location near the bottom of the field of view 612 , as shown in a particular field of view 608 .
- the processor 112 may adjust the tracking rate of the virtual images to increase the tracking rate, such as illustrated in field of view 614 .
- the processor 112 may determine that the wearer is reading quickly and try to supply more text by increasing the tracking rate.
- the upward movement rate of text 604 may increase.
- the wearer of the HMD may move his or her gaze point from a central location 610 to a location 618 near the top of the field of view 602 .
- the processor 112 may adjust the tracking rate of the virtual images to decrease the tracking rate.
- the processor 112 may determine that the wearer is reading slowly and try to supply text to the reader's eye more slowly by decreasing the tracking rate.
- the upward movement rate of text 604 may decrease.
- FIG. 6D depicts a scenario 620 wherein a tracking rate may be adjusted to zero.
- the wearer of the HMD may move his or her gaze point to a central location 622 of the field of view 602 , as shown in a particular field of view 616 . Further, the wearer of the HMD may fixate his or her eye gaze point upon the central location 622 for some predetermined period of time. When this eye gaze point position and/or the eye gaze point fixation is detected by the processor 112 , the processor 112 may adjust the tracking rate of the virtual images to further decrease or zero the tracking rate.
- the processor 112 may determine that the wearer wants to focus on a particular element of the virtual image and may provide a more stable virtual image by decreasing or zeroing the tracking rate.
- the movement of text 604 may decrease further in rate or stop completely.
- gaze axis movements near the reference axis may provide for smaller tracking rates than gaze axis movements at larger angle differences.
- eye gaze movements around a reference axis may provide finer virtual image tracking control while gazing farther away from the reference axis may provide coarse virtual image panning and scrolling tracking control.
- FIGS. 6A, 6B, 6C, and 6D illustrate an example embodiment in which an HMD may adjust text scrolling while the HMD could be stationary.
- example embodiments in which the HMD adjusts virtual images while simultaneously translating and/or rotating may also be considered.
- a HMD may display a user interface in which the virtual images are substantially anchored to the real-world environment. This ‘world-fixed’ user interface could appear to the HMD user as though the virtual images of the user interface are substantially fixed to an inside surface of an imaginary ring that surrounds the user's head. That is, when a HMD user turns his or her head to the left, the virtual images rotate within his or her field of view to the right, and vice versa.
- the HMD user may access and navigate menus and icons in the user interface by moving the HMD and by using the aforementioned HMD reference axis as a pointing device or cursor.
- accessing this menu may be complicated if the HMD user is in motion. For instance, when travelling on a subway, the HMD user may round a corner, which may affect the HMD position as well as orientation.
- a user interface menu controlled only by the HMD position and/or orientation may produce errors when the HMD changes its relative reference position such as inadvertent menu selection or rotation of icons in the viewable user interface.
- the utilization of an eye-tracking system could reduce inadvertent movement of the virtual images due to changes in HMD position and orientation. For instance, if the HMD position and/or orientation changes but the eye-tracking system detects no corresponding eye movement (anticipatory eye gaze changes, for example), the HMD may be configured to not adjust the virtual images with respect to the changing HMD position.
- the example embodiment 700 is illustrated in FIG. 7A in which the HMD user is riding a subway forward (position 704 to position 706 ) and then the subway curves left (position 706 to position 708 ).
- the HMD user may be accessing a user interface menu that could include icons and/or menus associated with files 718 , photos 720 , e-mail 722 , contacts 724 and a calendar 726 .
- FIG. 7B illustrates a possible view that the HMD user may see when located at position 704 .
- the e-mail menu 722 could be centered within the HMD display 714 and the gaze point 710 could be determined to be near the center of the HMD display 714 .
- the user interface icons and/or menus could be arranged in imaginary ring that surrounds the HMD user's head.
- the user interface may be configured to remain substantially ‘world-fixed’ and rotate in the opposite direction at least due to HMD panning and eye gaze change. For instance, if the HMD user rotates the HMD to the right and gazes towards the right side of the HMD display 714 , the user interface menus may rotate to the left, allowing the contacts 724 and calendar 726 to be displayed in the HMD display 714 .
- displacements such as moving straight ahead may not necessarily create a change in the displayed objects. For instance, if the HMD user moves forward in the subway car from position 704 to position 706 , the menu may stay substantially the same.
- displacements in HMD location could represent inputs that may cause the displayed objects to be adjusted. For instance, physical movements of the HMD could represent ‘walking’ through the user interface in three-dimensional space.
- FIG. 7D illustrates an embodiment 732 where the HMD user may be travelling on the subway car at position 708 .
- the HMD user is traveling forward and also rotating gradually to the left.
- the controller may interpret the HMD movement as an intentional movement to turn left. In general, this motion may lead to rotating the menus to the right. In this case, the photos 720 menu may be moved towards the center of the HMD display 716 .
- FIG. 7E illustrates an embodiment 736 that may use an eye-tracking system to correctly determine the actual desired speed of user interface movement.
- the HMD user is moving through position 708 , while moving forward and rotating to the left.
- an eye gaze point 738 may be determined to remain at the center of the HMD display 716 .
- the eye gaze fixation may cause the menu to not rotate.
- HMDs may present many different types of information in the form of virtual images to a wearer. Accordingly, each of these various virtual images may be assigned a different tracking rate and vector, which may be based on at least one or more context-related factors such as wearer reading speed, HMD orientation, HMD motion, HMD location, gaze axis, etc.
- a further example of how gaze direction may be used to correct for motion of the HMD is illustrated by the pseudo-code set forth below in Table 1.
- the routine may be called every time that there is a new sensor reading relating to the orientation of the HMD, for example, a sensor reading from gyroscope 120 , accelerometer 124 , or other component of HMD-Tracking system 104 .
- the routine is able to calculate two variables, adjustedX and adjustedY, which relate to how far the displayed images are to be moved in the x and y directions, respectively, using the moveScreen (adjustedX, adjustedY) function.
- the variables dx and dy which represent the distances that the wearer's head has travelled in the x and y directions, respectively, are determined from the sensor data.
- the variables vx and vy which represent the x and y displacements between the wearer's gaze location on the screen and the center of the screen, may be calculated based on the angle difference between the gaze axis and a reference axis that goes through the center of the screen.
- the variable adjustedX may then be calculated as a function of dx and vx
- the variable adjustedY may be calculated as a function of dy and vy. In this way, the wearer of the HMD is able to keep the position of the displayed images fixed, notwithstanding motion of the HMD, by gazing at the center of the screen.
- var radiusX adjustable number of units
- var radiusY adjustable number of units
- callback method that is called every time there is a new sensor reading.
- function handleHeadMovement( ) ⁇ // calculate the distance the head has moved. This distance might // be faulty due to sensor drift or unintended movement of the wearer, // as when riding the subway or walking around a corner
- var dx relative distance that the head travelled on the x axis
- var dy relative distance that the head travelled on the y axis
- var vx distance the eye gaze is from the center of the screen on the x axis
- var vy distance the eye gaze is from the center of the screen on the y axis; // adjust the distance travelled based on the distance the eye gaze is from // the center.
- Non-Transitory Computer Readable Medium to Determine Speed of Image Movement Using Eye Gaze Detection 4.
- the non-transitory computer readable medium could be, for example, a random access memory (RAM), a read-only memory (ROM), a flash memory, a cache memory, one or more magnetically encoded discs, one or more optically encoded discs, or any other form of non-transitory data storage.
- the non-transitory computer readable medium could also be distributed among multiple data storage elements, which could be remotely located from each other.
- the computing device that executes the stored instructions could be a wearable computing device, such as a wearable computing device 100 illustrated in FIG. 1 .
- the computing device that executes the stored instructions could be another computing device, such as a server in a server network.
- a non-transitory computer readable medium may store instructions executable by the processor 112 to perform various functions. For instance, instructions that could be used to carry out method 500 may be stored in memory 114 and could be executed by processor 112 .
- the processor 112 upon receiving gaze information from the eye-tracking system 102 , carry out instructions to determine a gaze axis and a reference axis as well as to control the HMD 100 to display virtual images within the HMD field of view and adjust a tracking rate based on the gaze axis and the reference axis.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Optics & Photonics (AREA)
- Computer Graphics (AREA)
- Computer Hardware Design (AREA)
- Software Systems (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
- This application claims priority to and is a continuation of U.S. patent application Ser. No. 14/605,587, filed on Jan. 26, 2015, entitled “Imaging Method.” Application Ser. No. 14/605,587 claims priority to and is a divisional of U.S. patent application Ser. No. 13/287,390, filed on Nov. 2, 2011, entitled “Imaging Method.” Both of these aforementioned applications are hereby incorporated by reference in their entirety.
- Wearable systems can integrate various elements, such as miniaturized computers, input devices, sensors, detectors, image displays, wireless communication devices as well as image and audio processors, into a device that can be worn by a user. Such devices provide a mobile and lightweight solution to communicating, computing and interacting with one's environment. With the advance of technologies associated with wearable systems and miniaturized optical elements, it has become possible to consider wearable compact optical displays that augment the wearer's experience of the real world.
- By placing an image display element close to the wearer's eye(s), an artificial image can be made to overlay the wearer's view of the real world. Such image display elements are incorporated into systems also referred to as “near-eye displays”, “head-mounted displays” (HMDs) or “heads-up displays” (HUDs). Depending upon the size of the display element and the distance to the wearer's eye, the artificial image may fill or nearly fill the wearer's field of view.
- In a first aspect, a head-mounted display (HMD) is provided. The HMD includes a head-mounted support, an optical system, an infrared light source, a camera, and a computer. The optical system is attached to the head-mounted support and includes a display panel configured to generate a virtual image, wherein the virtual image is viewable from a viewing location. The infrared light source is configured to illuminate the viewing location with infrared light such that infrared light is reflected from the viewing location as reflected infrared light and the camera is configured to image the viewing location by collecting the reflected infrared light. The computer is configured to determine a gaze axis based on one or more images of the viewing location obtained by the camera and control the display panel to move the virtual images within a field of view based on the gaze axis, a reference axis related to the HMD, and a tracking rate.
- In a second aspect, a method is provided. The method includes determining a gaze axis within a field of view of a head-mounted display (HMD), wherein the HMD is configured to display virtual images within the field of view. The method further includes determining a reference axis related to the HMD, adjusting a tracking rate based on the gaze axis and the reference axis, and moving the virtual images within the field of view based on the gaze axis, the reference axis and the tracking rate.
- In a third aspect, a non-transitory computer readable medium is provided. The non-transitory computer readable medium includes instructions executable by a computing device to cause the computing device to perform functions including, receiving eye-tracking images from a head-mounted display (HMD), wherein the HMD is configured to display virtual images within a field of view. The non-transitory computer readable medium further includes determining a gaze axis from the eye-tracking images, determining a reference axis related to the HMD, calculating an angle difference between the gaze axis and the reference axis, and adjusting a tracking rate based on the angle difference. The non-transitory computer readable medium further includes controlling the HMD to display the virtual images based upon the gaze axis, the reference axis, and the tracking rate.
-
FIG. 1 is schematic diagram of a wearable computing device, in accordance with an example embodiment. -
FIG. 2 is a top view of an optical system, in accordance with an example embodiment. -
FIG. 3A is a front view of a head-mounted display, in accordance with an example embodiment. -
FIG. 3B is a top view of the head-mounted display ofFIG. 3A , in accordance with an example embodiment. -
FIG. 3C is a side view of the head-mounted display ofFIG. 3A andFIG. 3B , in accordance with an example embodiment. -
FIG. 4A is a side view of a head-mounted display with a forward gaze axis, in accordance with an example embodiment. -
FIG. 4B is a side view of the head-mounted display ofFIG. 4A with an upward gaze axis, in accordance with an example embodiment. -
FIG. 5 is a flowchart of a method, in accordance with an example embodiment. -
FIG. 6A is a field of view of a head-mounted display showing scrolling text, in accordance with an example embodiment. -
FIG. 6B is a field of view of a head-mounted display showing scrolling text, in accordance with an example embodiment. -
FIG. 6C is a field of view of a head-mounted display showing scrolling text, in accordance with an example embodiment. -
FIG. 6D is a field of view of a head-mounted display showing scrolling text, in accordance with an example embodiment. -
FIG. 7A is an overhead view of HMD user traveling on a subway, in accordance with an example embodiment. -
FIG. 7B is a field of view of a HMD user interface, in accordance with an example embodiment. -
FIG. 7C is a field of view of a HMD user interface, in accordance with an example embodiment. -
FIG. 7D is a field of view of a HMD user interface, in accordance with an example embodiment. -
FIG. 7E is a field of view of a HMD user interface, in accordance with an example embodiment. - In the following detailed description, reference is made to the accompanying figures, which form a part thereof. In the figures, similar symbols typically identify similar components, unless context dictates otherwise. The illustrative embodiments described in the detailed description and figures are not meant to be limiting. Other embodiments may be utilized, and other changes may be made, without departing from the spirit or scope of the subject matter presented herein. It will be readily understood that the aspects of the present disclosure, as generally described herein, and illustrated in the figures, can be arranged, substituted, combined, separated, and designed in a wide variety of different configurations, all of which are contemplated herein.
- 1. Overview
- A head-mounted display (HMD) may enable its wearer to observe the wearer's real-world surroundings and also view a displayed image, such as a computer-generated image. In some cases, the displayed image may overlay a portion of the wearer's field of view of the real world. Thus, while the wearer of the HMD is going about his or her daily activities, such as walking, driving, exercising, etc., the wearer may be able to see a displayed image generated by the HMD at the same time that the wearer is looking out at his or her real-world surroundings.
- The displayed image, also known as a virtual image, might include, for example, graphics, text, and/or video. The content of the displayed image could relate to any number of contexts, including but not limited to the wearer's current environment, an activity in which the wearer is currently engaged, the biometric status of the wearer, and any audio, video, or textual communications that have been directed to the wearer. The images displayed by the HMD may also be part of an interactive user interface. For example, the HMD could be part of a wearable computing device. Thus, the images displayed by the HMD could include menus, selection boxes, navigation icons, or other user interface features that enable the wearer to invoke functions of the wearable computing device or otherwise interact with the wearable computing device.
- The images displayed by the HMD could appear anywhere in the wearer's field of view. For example, the displayed image might occur at or near the center of the wearer's field of view, or the displayed image might be confined to the top, bottom, or a corner of the wearer's field of view. Alternatively, the displayed image might be at the periphery of or entirely outside of the wearer's normal field of view. For example, the displayed image might be positioned such that it is not visible when the wearer looks straight ahead but is visible when the wearer looks in a specific direction, such as up, down, or to one side. In addition, the displayed image might overlay only a small portion of the wearer's field of view, or the displayed image might fill most or all of the wearer's field of view. The displayed image could be displayed continuously or only at certain times (e.g., only when the wearer is engaged in certain activities).
- The displayed images may appear fixed relative to the wearer's environment. For instance, the virtual images may appear anchored to a particular object or location within the wearer's environment. Alternatively, displayed images may appear fixed relative to the wearer's field of view. For example, the HMD may include a graphical user interface that may stay substantially anchored to the wearer's field of view regardless of the HMD orientation. In practice, both types of virtual imagery may be implemented together in an HMD.
- To display a virtual image to the wearer, an optical system in the HMD may include a light source, such as a light-emitting diode (LED), that is configured to illuminate a display panel, such as a liquid crystal-on-silicon (LCOS) display. The display panel generates light patterns by spatially modulating the light from the light source, and the light patterns may be viewable as virtual images at a viewing location.
- The HMD may obtain data from the wearer in order to perform certain functions, for instance to provide context-sensitive information to the wearer. In an example embodiment, the HMD may obtain information regarding the wearer and the wearer's environment and respond accordingly. For instance, the HMD may use a pupil position recognition technique, wherein if the HMD recognizes that the wearer's pupil location, and thus a corresponding gaze axis, is inclined with respect to a reference axis, the HMD may display virtual images related to objects located above the wearer. Conversely, the HMD may recognize, by a similar pupil position recognition technique, that the wearer is looking downward. Accordingly, the HMD may display virtual images related to objects located below a reference axis of the wearer.
- In order to determine the actual position of a HMD wearer's pupil and to determine a corresponding gaze axis, the wearer's pupil may be illuminated by an infrared light source or multiple infrared light sources. An infrared camera may image the pupil. The infrared light source(s) could be located in the HMD optical path, or could alternatively be located off-axis. The infrared camera could also be located in the HMD optical path or off-axis. Possible eye tracking modalities that could be used include dark pupil imaging and dual-glint Purkinje image tracking, among other techniques known in the art.
- A processor may implement an image processing algorithm to find the edges or extents of the imaged pupil. The image processing algorithms may include pattern recognition, Canny edge detection, thresholding, contrast detection, or differential edge detection, to name a few. Those skilled in the art will understand that a variety of different image processing techniques could be used individually or in combination with other methods in order to obtain pupil location. After image processing, the processor may determine a gaze axis, which may be defined as an axis extending from a viewing location and through a gaze point located within the wearer's field of view.
- The processor may also determine a reference axis, which may be defined as an axis extending from a viewing location and through a point in space. The point in space may include the apparent center of the display of the HMD or a target object, among other possibilities.
- Once a gaze axis is determined, the processor may act to adjust various components of the displayed virtual image based on an angle difference between the gaze axis and the reference axis and a tracking rate. For example, an upward scrolling list of text information may be presented to a user of an HMD, similar to the traditional display of credits at the end of a movie. In this embodiment, the reference axis may include an axis that extends through the apparent center of the HMD display. If a gaze axis is determined that indicates the user is gazing down near the bottom of the display with respect to the reference axis, the tracking rate may be increased such that the rate of upward scrolling is increased, providing more text to the user.
- Conversely, if a gaze axis is determined that indicates that the user is gazing near the top of the screen with respect to the reference axis, the tracking rate may be decreased such that the rate of upward scrolling is slowed, allowing the user to catch up in reading the text.
- Alternatively or additionally, if a gaze axis is determined to be near the reference axis (i.e. the user is looking at the middle of the screen), the tracking rate may be decreased or set to zero (eliminating scrolling of the virtual image). In this respect, a user may be able to focus his/her attention on specific text.
- It will be evident to those skilled in the art that there are a variety of ways to implement such virtual image adjustment in a HMD system. The details of such implementations may depend on, for example, the type of data provided, the local environmental conditions, the location of the user, and the task to be performed.
- Certain illustrative examples of using eye-tracking data to adjust a virtual image displayed by a HMD are described below. It is to be understood, however, that other embodiments are possible and are implicitly considered within the context of the following example embodiments.
- 2. Adjusting Virtual Images in a Head-Mounted Display Using Eye-Tracking
-
FIG. 1 is schematic diagram of a wearable computing device or a head-mounted display (HMD) 100 that may include several different components and subsystems. In one example,HMD 100 includes an eye-trackingsystem 102, a HMD-trackingsystem 104, anoptical system 106, peripherals, apower supply 110, aprocessor 112, amemory 114, and auser interface 115. The eye-trackingsystem 102 may include hardware such as aninfrared camera 116 and at least one infraredlight source 118. The HMD-trackingsystem 104 may include agyroscope 120, a global positioning system (GPS) 122, and anaccelerometer 124. Theoptical system 106 may include, in one embodiment, adisplay panel 126, a displaylight source 128, andoptics 130.Peripherals 108 may include, for example, awireless communication interface 134, atouchpad 136, amicrophone 138, acamera 140, and aspeaker 142. - In an example embodiment,
HMD 100 includes a see-through display. Thus, the wearer ofHMD 100 may observe a portion of the real-world environment, i.e., in a particular field of view provided by theoptical system 106. In addition,HMD 100 is operable to display virtual images that are superimposed on the field of view, for example, to provide an “augmented reality” experience. Some of the virtual images displayed byHMD 100 may be superimposed over particular objects in the field of view.HMD 100 may also display images that appear to hover within the field of view instead of being associated with particular objects in the field of view. - Components of the
HMD 100 may be configured to work in an interconnected fashion with other components within or outside their respective systems. For instance, in an example embodiment, theinfrared camera 116 may image one or both of the HMD wearer's eyes. Theinfrared camera 116 may deliver image information to theprocessor 112, which may access thememory 112 and make a determination regarding the direction of the HMD wearer's gaze, also termed a gaze axis. Theprocessor 112 may further accept input from theGPS unit 122, thegyroscope 120, and/or theaccelerometer 124 to determine the location and orientation of theHMD 100. Subsequently, theprocessor 112 may control theuser interface 115 and thedisplay panel 126 to display virtual images to the HMD wearer that may include context-specific information based on the HMD location and orientation as well as the HMD wearer's gaze axis. -
HMD 100 could be configured as, for example, eyeglasses, goggles, a helmet, a hat, a visor, a headband, or in some other form that can be supported on or from the wearer's head. Further,HMD 100 may be configured to display images to both of the wearer's eyes, for example, using two see-through displays. Alternatively,HMD 100 may include only a single see-through display and may display images to only one of the wearer's eyes, either the left eye or the right eye. In other embodiments,HMD 100 may include an opaque display configured to display images to a first eye or both eyes of the HMD wearer. In embodiments where an opaque display is presented to the first eye of the HMD wearer, a view of the real-world environment could be available to a second eye of the HMD wearer. - A
power supply 110 may provide power to various HMD components and could represent, for example, a rechargeable lithium-ion battery. Various other power supply materials and types known in the art are possible. - The function of the
HMD 100 may be controlled by aprocessor 112 that executes instructions stored in a non-transitory computer readable medium, such as thememory 114. Thus,processor 112 in combination with instructions stored in thememory 114 may function as a controller ofHMD 100. As such,processor 112 may control theuser interface 115 to adjust what images are displayed byHMD 100. Theprocessor 112 may also control thewireless communication interface 134 and various other components of theHMD 100. Theprocessor 112 may additionally represent a plurality of computing devices that may serve to control individual components or subsystems of theHMD 100. - In addition to instructions that may be executed by the
processor 112, thememory 114 may store data that may include a set of calibrated wearer eye pupil positions and a collection of past eye pupil positions. Thus, thememory 114 may function as a database of information related to gaze direction. Such information may be used byHMD 100 to anticipate where the user will look and determine what images are to be displayed to the wearer. Calibrated wearer eye pupil positions may include, for instance, information regarding the extents or range of the wearer's eye pupil movement (right/left and upwards/downwards) as well as wearer eye pupil positions that may relate to various reference axes. - Reference axes could represent, for example, an axis extending from a viewing location and through a target object or the apparent center of a field of view (i.e. the reference axis may correspond to a center line of the field of view). Other possibilities for reference axes exist. Thus, a reference axis may further represent a basis for determining dynamic gaze direction.
- In addition, information may be stored in the
memory 114 regarding possible control instructions that may be enacted using eye movements. For instance, two consecutive wearer eye blinks may represent a control instruction directing theHMD 100 to capture an image with aperipheral camera 140. Control instructions could also include the dwell-based selection of a target object. For instance, if a wearer fixates visually upon a particular virtual image or real-world object for longer than a predetermined time period, a control instruction may be generated to select the virtual image or real-world object as a target object. Many other control instructions are possible. - In addition to the aforementioned features,
memory 114 could store various recorded data from previous HMD/user interactions. For instance, multiple images of a HMD wearer's eye(s) could be averaged to obtain an averaged eye gaze axis. This could lessen the effect of saccadic eye movements or saccades, in which the eye moves in a rapid and somewhat random manner around an eye gaze axis. These saccades help humans build up a mental image of a field of view with better resolution than if the eye remained static, and by averaging a number of eye images within a particular time period, an average gaze axis could be determined with less saccadic ‘noise’. - Additionally,
memory 114 could store recorded data regarding recent eye gaze axes for various application-based functions. For instance, the recent variance of the eye gaze axis could be coupled to scrolling images generated by theHMD 100. In this embodiment, if recent eye gaze axis variance is high, the images (e.g. text or other images) could scroll faster. If the eye gaze axis variance is low, the images may scroll slower or stop altogether. In this context, a lower variance in eye gaze axis could indicate the HMD wearer is concentrating on one particular gaze location, whereas a higher eye gaze axis variance means the opposite—the HMD wearer may be quickly scanning a document and desire a faster scrolling speed. - Depending on the content that is presented on the HMD display, the variance may differ depending on the axis along which it is measured. For example, the horizontal variance of a HMD wearer's eye gaze may be high while the vertical variance may be relatively low. This could indicate to the
HMD 100 that the wearer is reading text. Accordingly, text scrolling/tracking could be adjusted in a different or more controlled fashion compared to ‘non-reading’ scrolling/panning/pagination situations. - The
HMD 100 may include auser interface 115 for providing information to the wearer or receiving input from the wearer. Theuser interface 115 could be associated with, for example, the displayed virtual images, a touchpad, a keypad, buttons, a microphone, and/or other peripheral input devices. Theprocessor 112 may control the functioning of theHMD 100 based on input received through theuser interface 115. For example, theprocessor 112 may utilize user input from theuser interface 115 to control how theHMD 100 displays images within a field of view or determine what images theHMD 100 displays. - The
infrared camera 116 may be utilized by the eye-trackingsystem 102 to capture images of a viewing location associated with theHMD 100. Thus, theinfrared camera 116 may image the eye of a HMD wearer that may be located at the viewing location. The images could be either video images or still images. The images obtained by theinfrared camera 116 regarding the HMD wearer's eye may help determine where the wearer is looking within the HMD field of view, for instance by ascertaining the location of the HMD wearer's eye pupil. - Analysis of the images obtained by the
infrared camera 116 could be performed by theprocessor 112 in conjunction with thememory 114. - The imaging of the viewing location could occur continuously or at discrete times depending upon, for instance, user interactions with the
user interface 115. Theinfrared camera 116 could be integrated into theoptical system 106. Alternatively, theinfrared camera 116 could be mounted separately from theoptical system 106 and/orHMD 100. Furthermore, theinfrared camera 116 could additionally represent a visible light camera with sensing capabilities in the infrared wavelengths. - The infrared
light source 118 could represent one or more infrared light-emitting diodes (LEDs) or infrared laser diodes that may illuminate a viewing location. Thus, one or both eyes of a wearer of theHMD 100 may be illuminated by the infraredlight source 118. The infraredlight source 118 may be positioned along an optical axis common to the infrared camera, and/or the infraredlight source 118 may be positioned elsewhere. The infraredlight source 118 could be mounted separately from theoptical system 106 and/orHMD 100. The infraredlight source 118 may illuminate the viewing location continuously or may be turned on at discrete times. Additionally, when illuminated, the infraredlight source 118 may be modulated at a particular frequency. - The HMD-tracking
system 104 could be configured to provide a HMD position and HMD orientation to theprocessor 112. This position and orientation data may help determine a reference axis to which a gaze axis is compared. For instance, the reference axis may correspond to the orientation of the HMD. - The
gyroscope 120 could be a microelectromechanical system (MEMS) gyroscope or a fiber optic gyroscope. Thegyroscope 120 may be configured to provide orientation information to theprocessor 112. TheGPS unit 122 could be a receiver that obtains clock and other signals from GPS satellites and may be configured to provide real-time location information to theprocessor 112. The HMD-trackingsystem 104 could further include anaccelerometer 124 configured to provide motion input data to theprocessor 112. - The
optical system 106 could represent components configured to provide virtual images to a viewing location. An example ofoptical system 106 is described in detail below. - Various
peripheral devices 108 may be included in theHMD 100 and may serve to provide information to and from a wearer of theHMD 100. In one example, theHMD 100 may include awireless communication interface 134 for wirelessly communicating with one or more devices directly or via a communication network. For example,wireless communication interface 134 could use 3G cellular communication, such as CDMA, EVDO, GSM/GPRS, or 4G cellular communication, such as WiMAX or LTE. Alternatively,wireless communication interface 134 could communicate with a wireless local area network (WLAN), for example, using WiFi. In some embodiments,wireless communication interface 134 could communicate directly with a device, for example, using an infrared link, Bluetooth, or ZigBee. - Although
FIG. 1 shows various components of the HMD 100 (i.e.,wireless communication interface 134,processor 112,memory 114,infrared camera 116,display panel 126,GPS 122, and user interface 115) as being integrated intoHMD 100, one or more of these components could be physically separate fromHMD 100. For example,infrared camera 116 could be mounted on the wearer separate fromHMD 100. Thus, theHMD 100 could be part of a wearable computing device in the form of separate devices that can be worn on or carried by the wearer. The separate components that make up the wearable computing device could be communicatively coupled together in either a wired or wireless fashion. -
FIG. 2 illustrates a top view of anoptical system 200 that is configured to display a virtual image superimposed upon a real-world scene viewable along aviewing axis 204. For clarity, adistal portion 232 and aproximal portion 234 represent optically-coupled portions of theoptical system 200 that may or may not be physically separated. An example embodiment includes adisplay panel 206 that may be illuminated by alight source 208. Light emitted from thelight source 208 is incident upon thedistal beam splitter 210. Thelight source 208 may include one or more light-emitting diodes (LEDs) and/or laser diodes. Thelight source 208 may further include a linear polarizer that acts to pass one particular polarization to the rest of the optical system. - In an example embodiment, the
distal beam splitter 210 is a polarizing beam splitter that reflects light depending upon the polarization of light incident upon the beam splitter. To illustrate, s-polarized light from thelight source 208 may be preferentially reflected by a distal beam-splittinginterface 212 towards thedisplay panel 206. Thedisplay panel 206 in the example embodiment is a liquid crystal-on-silicon (LCOS) display, but could also be a digital light projector (DLP) micro-mirror display, or other type of reflective display panel. Thedisplay panel 206 acts to spatially-modulate the incident light to generate a light pattern. Alternatively, thedisplay panel 206 may be an emissive-type display such as an organic light-emitting diode (OLED) display. - In the example in which the
display panel 206 is a LCOS display panel, thedisplay panel 206 generates a light pattern with a polarization perpendicular to the polarization of light initially incident upon the panel. In this example embodiment, thedisplay panel 206 converts incident s-polarized light into a light pattern with p-polarization. The generated light pattern from thedisplay panel 206 is directed towards thedistal beam splitter 210. The p-polarized light pattern passes through thedistal beam splitter 210 and is directed along anoptical axis 214 towards the proximal region of theoptical system 200. In an example embodiment, theproximal beam splitter 216 is also a polarizing beam splitter. The light pattern is at least partially transmitted through theproximal beam splitter 216 to the image former 218. In an example embodiment, image former 218 includes aconcave mirror 230 and a proximal quarter-wave plate 228. The light pattern passes through the proximal quarter-wave plate 228 and is reflected by theconcave mirror 230. - The reflected light pattern passes back through proximal quarter-
wave plate 228. Through the interactions with the proximal quarter-wave plate 228 and theconcave mirror 230, the light patterns are converted to the s-polarization and are formed into a viewable image. This viewable image is incident upon theproximal beam splitter 216 and the viewable image is reflected from proximalbeam splitting interface 220 towards aviewing location 222 along aviewing axis 204. A real-world scene is viewable through aviewing window 224. Theviewing window 224 may include a linear polarizer in order to reduce stray light within the optical system. Light from theviewing window 224 is at least partially transmitted through theproximal beam splitter 216. Thus, both a virtual image and a real-world image are viewable to theviewing location 222 through theproximal beam splitter 216. - Although
FIG. 2 depicts thedistal portion 232 of the optical system housing as to the left of theproximal portion 234 of the optical system housing when viewed from above, it is understood that other embodiments are possible to physically realize theoptical system 200, including thedistal portion 232 being configured to be to the right, below and above with respect to theproximal portion 234. Further, although an example embodiment describes an image former 218 as comprising aconcave mirror 230, it is understood by those skilled in the art that the image former 218 may comprise a different optical element, such as an optical lens or a diffractive optic element. - In one embodiment, the
proximal beam splitter 216, thedistal beam splitter 210, and other components ofoptical system 200 are made of glass. Alternatively, some or all of such optical components may be partially or entirely plastic, which can also function to reduce the weight ofoptical system 200. A suitable plastic material is Zeonex® E48R cyclo olefin optical grade polymer which is available from Zeon Chemicals L.P., Louisville, Ky. Another suitable plastic material is polymethyl methacrylate (PMMA). - An example embodiment may include an infrared
light source 226 that is configured to illuminate theviewing location 222. AlthoughFIG. 2 depicts the infraredlight source 226 as adjacent toviewing window 224, those skilled in the art will understand that the infraredlight source 226 could be located elsewhere, such as on the side of theproximal beam splitter 216 that is adjacent to theviewing location 222 or in thedistal portion 232 of theoptical system 200. The infraredlight source 226 may represent, for example, one or more infrared light-emitting diodes (LEDs). Infrared LEDs with a small size may be implemented, such as the Vishay Technology TSML 1000 product. - Further, those skilled in the art will understand that, for best eye-tracking accuracy, it may be advantageous to obtain infrared images of the eye pupil using light sources that illuminate the eye from positions off-axis and/or on-axis with respect to the
viewing axis 204. Therefore, the infraredlight source 226 may include one or more LEDs located at different locations in theoptical system 200. - Infrared light generated from the infrared
light source 226 is configured to be incident upon theviewing location 222. Thus, the wearer's eye pupil may be illuminated with the infrared light. The infrared light may be reflected from the wearer's eye back along theviewing axis 204 towards theproximal beam splitter 216. A portion of the reflected infrared light may be reflected from thebeam splitting interface 220 towards the image former 218. - In order to transmit infrared light to an
infrared camera 202, the image former 218 may include a dichroic thin film configured to selectively reflect or transmit incident light depending upon the wavelength of the incident light. For instance, the dichroic thin film may be configured to pass infrared light while reflecting visible light. In an example embodiment, the visible light pattern generated by thedisplay panel 206 may be reflected by theconcave mirror 230 and the visible light pattern may be formed into a viewable image. The infrared light may thus be preferably transmitted through theconcave mirror 230 toinfrared camera 202. Dichroic thin film coatings are available commercially from companies such as JML Optical Industries and Precision Glass & Optics (PG&O) and comprise multiple layers of dielectric and/or metal films. These dichroic coatings are also called ‘cold mirrors’. - In an example embodiment, a small aperture or apertures may be introduced into the image former 218, which may be realized by one or more pinholes in the
concave mirror 230. In this example embodiment, most of the visible and infrared light is reflected off of and formed by the image former 218 into an image viewable by the HMD wearer. Some of the visible and infrared light passes through the aperture and is incident upon theinfrared camera 202. Theinfrared camera 202 may selectively filter and detect the infrared light from the combination of visible and infrared light to obtain information regarding the wearer's eye pupil location. Alternatively, the infraredlight source 226 may be modulated to provide a frequency reference for a lock-in amplifier or phase-locked loop in order that the infrared light signal is obtained efficiently. Also, the visiblelight source 208 may be modulated and infrared light detection could be performed when the visiblelight source 208 is off, for example. Those with skill in the art will understand that there are other variations of transducing an infrared light signal mixed with a visible light signal with an infrared camera and that those variations are included implicitly in this specification. -
FIG. 3A presents a front view of a head-mounted display (HMD) 300 in an example embodiment that includes a head-mountedsupport 309.FIGS. 3B and 3C present the top and side views, respectively, of the HMD inFIG. 3A . Although this example embodiment is provided in an eyeglasses format, it will be understood that wearable systems and HMDs may take other forms, such as hats, goggles, masks, headbands and helmets. The head-mountedsupport 309 includes lens frames 314 and 316, acenter frame support 318,lens elements arms center frame support 318 and side-arms support 309 to the wearer's head via the wearer's nose and ears, respectively. Each of theframe elements arms support 309. Alternatively or additionally, head-mountedsupport 309 may support external wiring.Lens elements left eye 308 may look throughleft lens 312 and the wearer'sright eye 306 may look throughright lens 310.Optical systems FIG. 2 , may be positioned in front oflenses FIGS. 3A, 3B, and 3C .Optical systems support 309 using support mounts 324 and 326, respectively. Furthermore,optical systems lens elements - Although this example includes an optical system for each of the wearer's eyes, it is to be understood that a HMD might include an optical system for only one of the wearer's eyes (either
left eye 308 or right eye 306). As described inFIG. 2 , the HMD wearer may simultaneously observe fromoptical systems 302 and 304 a real-world image with an overlaid virtual image. TheHMD 300 may include various elements such as aprocessor 340, atouchpad 342, amicrophone 344, and abutton 346. Thecomputer 340 may use data from, among other sources, various sensors and cameras to determine the virtual image that should be displayed to the user. In an example embodiment, as described earlier, an infrared light source or sources may illuminate the viewing position(s) 308 and 306, i.e. the wearer's eye(s), and the reflected infrared light may be preferentially collected with an infrared camera. - Those skilled in the art would understand that other user input devices, user output devices, wireless communication devices, sensors, and cameras may be reasonably included in such a wearable computing system.
-
FIGS. 4A and 4B depict side and front views of an eye as well as schematic drawings of pupil location information under different conditions. One way to determine a gaze axis of a person is to ascertain the position of the person's eye pupil with respect to a reference point, such as a viewing location. To track eye pupil movements, infrared light is generally reflected off of a person's eye. The reflected light may be collected and detected with an infrared detector. Upon imaging of the eye, image processing can be conducted with aprocessor 112 in order to determine, for instance, the extents and centroid location of the person's pupil. The other known means and methods of eye-tracking, including the use of visible light illumination and/or imaging techniques are possible. - For example, in an
embodiment 400, a person may be looking directly forward as depicted inFIG. 4A . Theeye 412 is open and thepupil 418 is located along areference axis 410. After image processing, which may include edge detection, the position of the pupil may be determined to be atpupil location 422. In this example, theprocessor 112 may subsequently determine that the gaze axis based on thepupil location 422 coincides with areference axis 410. Virtual image display position and movement may be adjusted due to the determinedpupil location 422. For instance, theprocessor 112 may adjust a tracking rate to zero when the gaze axis and the reference axis are equivalent or nearly equivalent. This may allow a user to slowly read critical text or closely examine a virtual image, for example. - In an
example embodiment 424, as illustrated inFIG. 4B , a person may be looking upwards with respect to areference axis 428. Theeye 434 is open and the pupil location is generally higher than a reference point 440. In this situation, imaging the person'spupil 438 with infrared light may result in adetermined pupil position 442. Theprocessor 112 may determine that thegaze axis 430 that is above thereference axis 428. Theangle difference 432 may represent the absolute difference in angle between thereference axis 428 and thegaze axis 430. Theprocessor 112 may calculate theangle difference 432 and, based on theangle difference 432, adjust a tracking rate. For instance, alarge angle difference 432 could represent an adjustment in tracking rate such that the tracking rate is higher, for instance to scroll a virtual image across a field of view at a faster rate. - Other embodiments could include the use of different eye gaze determination techniques. For instance, instead of using the eye pupil to determine gaze axis, it is possible to track eye motions using the boundary between the sclera and iris (416 and 436 in
FIGS. 4A and 4B ). For the purposes of determining an eye gaze axis, finding the centroid of the sclera/iris boundary may be equivalent to finding the centroid of a pupil. - 3. A Method for Adjusting Virtual Images within a Field of View Based on a Gaze Axis, a Reference Axis, and a Tracking Rate.
- A
method 500 is provided for adjusting virtual images within a field of view based on a gaze axis, a reference axis and a tracking rate.Method 500 could be performed using an HMD that is configured as shown in any ofFIGS. 1-3C or configured in some other way.FIG. 5 illustrates the steps in an example method, however, it is understood that in other embodiments, the steps may appear in different order and steps may be added or subtracted. - In the method, a gaze axis is determined within a field of view of a head-mounted display (HMD) (Step 502). The HMD is generally configured to display virtual images to be viewable at a viewing location and could be an HMD similar to an aforementioned embodiment. The gaze axis could be determined to be an axis extending from a center of the HMD wearer's pupil. The gaze axis could be similar to the
gaze axis 430 depicted inFIG. 4B . The pupil location could be determined using eye glint images or by other eye-tracking techniques detailed above. - A reference axis related to the HMD may also be determined in the method (Step 504). The reference axis could be determined by the
processor 112 based on the orientation of the HMD and may correspond to the apparent center of the HMD field of view, for example. Other reference axes are possible. - A tracking rate related to the movement of virtual images may be adjusted based on the gaze axis and the reference axis (Step 506). The tracking rate may be the rate at which virtual images are panning and/or scrolling across the HMD field of view. The tracking rate may depend upon the data that is displayed. For instance, words on a document may scroll in a vertical fashion from the bottom to the top of the HMD field of view to simulate reading downwards along a printed page. In this case, the display may move the text upwards at a tracking rate of around 1 second per line.
- A tracking rate could also be related to the motion of a HMD wearer. For instance, when displaying virtual images that may be user interface menus, for instance, the HMD may attempt to base the tracking rate of the virtual images on the rate of HMD movement. More specifically, an HMD wearer may access a user interface by changing the orientation of the HMD (for instance rotating one's head and/or body to select different elements of the user interface). Thus, the user interface could be at least partially anchored to locations and objects in the real world and the virtual images could be adjusted or panned at a rate proportional to the rate of change of the HMD orientation.
- Further, the tracking rate could be adjusted by the angle difference between the gaze axis and the reference axis. For instance, the tracking rate could be increased if the angle difference between the gaze axis and the reference axis is large and the tracking rate could be decreased if the angle difference is small. The tracking rate could also be adjusted based on the direction of the vector between the reference axis and the gaze axis. More detailed examples are given below.
- The virtual images may be adjusted within the field of view based on the gaze axis, the reference axis, and the tracking rate (Step 508). If the tracking rate is adjusted lower, the movement rate of the virtual images may slow, for instance. Conversely, with a higher tracking rate, the virtual images may appear to move more quickly within the field of view.
-
FIGS. 6A, 6B, 6C, and 6D illustrate an example in which a determined gaze axis controls text scrolling. In theexample embodiment 600, virtual images including text are presented within a field ofview 602. Thetext 604 may be scrolling slowly upwards at a normal tracking rate similar in fashion to credits at the end of a movie (around one line per second, for instance). Agaze point 606 may be ascertained related to a gaze axis and thus to the position of an eye pupil of a wearer of an HMD. In this example embodiment, the reference axis may be considered as originating from the wearer's eye and going through the apparent center of the field ofview 602. When a wearer is reading normally and the angle between the reference axis and the gaze axis is relatively small, the tracking rate may stay unchanged and thetext 604 may continue to scroll upwards. - In
FIG. 6B , the wearer of the HMD may move his or her gaze point from acentral location 610 to a location near the bottom of the field ofview 612, as shown in a particular field ofview 608. When this change in eye gaze point is detected by theprocessor 112, theprocessor 112 may adjust the tracking rate of the virtual images to increase the tracking rate, such as illustrated in field ofview 614. In particular, theprocessor 112 may determine that the wearer is reading quickly and try to supply more text by increasing the tracking rate. Thus, the upward movement rate oftext 604 may increase. - In a
related scenario 616 depicted inFIG. 6C , whiletext 604 is scrolling upward within a HMD field ofview 602, the wearer of the HMD may move his or her gaze point from acentral location 610 to alocation 618 near the top of the field ofview 602. When this change in eye gaze point is detected by theprocessor 112, theprocessor 112 may adjust the tracking rate of the virtual images to decrease the tracking rate. In particular, theprocessor 112 may determine that the wearer is reading slowly and try to supply text to the reader's eye more slowly by decreasing the tracking rate. Thus, the upward movement rate oftext 604 may decrease. -
FIG. 6D depicts ascenario 620 wherein a tracking rate may be adjusted to zero. For example, whiletext 604 is scrolling upward within a HMD field ofview 602, the wearer of the HMD may move his or her gaze point to acentral location 622 of the field ofview 602, as shown in a particular field ofview 616. Further, the wearer of the HMD may fixate his or her eye gaze point upon thecentral location 622 for some predetermined period of time. When this eye gaze point position and/or the eye gaze point fixation is detected by theprocessor 112, theprocessor 112 may adjust the tracking rate of the virtual images to further decrease or zero the tracking rate. In particular, theprocessor 112 may determine that the wearer wants to focus on a particular element of the virtual image and may provide a more stable virtual image by decreasing or zeroing the tracking rate. Thus, the movement oftext 604 may decrease further in rate or stop completely. Furthermore, gaze axis movements near the reference axis may provide for smaller tracking rates than gaze axis movements at larger angle differences. Thus, eye gaze movements around a reference axis may provide finer virtual image tracking control while gazing farther away from the reference axis may provide coarse virtual image panning and scrolling tracking control. -
FIGS. 6A, 6B, 6C, and 6D illustrate an example embodiment in which an HMD may adjust text scrolling while the HMD could be stationary. However, example embodiments in which the HMD adjusts virtual images while simultaneously translating and/or rotating may also be considered. For instance, a HMD may display a user interface in which the virtual images are substantially anchored to the real-world environment. This ‘world-fixed’ user interface could appear to the HMD user as though the virtual images of the user interface are substantially fixed to an inside surface of an imaginary ring that surrounds the user's head. That is, when a HMD user turns his or her head to the left, the virtual images rotate within his or her field of view to the right, and vice versa. - In one situation, the HMD user may access and navigate menus and icons in the user interface by moving the HMD and by using the aforementioned HMD reference axis as a pointing device or cursor. However, accessing this menu may be complicated if the HMD user is in motion. For instance, when travelling on a subway, the HMD user may round a corner, which may affect the HMD position as well as orientation. Thus, a user interface menu controlled only by the HMD position and/or orientation may produce errors when the HMD changes its relative reference position such as inadvertent menu selection or rotation of icons in the viewable user interface.
- In an example embodiment, the utilization of an eye-tracking system could reduce inadvertent movement of the virtual images due to changes in HMD position and orientation. For instance, if the HMD position and/or orientation changes but the eye-tracking system detects no corresponding eye movement (anticipatory eye gaze changes, for example), the HMD may be configured to not adjust the virtual images with respect to the changing HMD position.
- The
example embodiment 700 is illustrated inFIG. 7A in which the HMD user is riding a subway forward (position 704 to position 706) and then the subway curves left (position 706 to position 708). In theexample embodiment 712, the HMD user may be accessing a user interface menu that could include icons and/or menus associated withfiles 718,photos 720,e-mail 722,contacts 724 and acalendar 726.FIG. 7B illustrates a possible view that the HMD user may see when located atposition 704. Thee-mail menu 722 could be centered within theHMD display 714 and thegaze point 710 could be determined to be near the center of theHMD display 714. - As described above, the user interface icons and/or menus could be arranged in imaginary ring that surrounds the HMD user's head. In the
example embodiment 712, if at rest, the user interface may be configured to remain substantially ‘world-fixed’ and rotate in the opposite direction at least due to HMD panning and eye gaze change. For instance, if the HMD user rotates the HMD to the right and gazes towards the right side of theHMD display 714, the user interface menus may rotate to the left, allowing thecontacts 724 andcalendar 726 to be displayed in theHMD display 714. - However, as shown in
FIG. 7C , displacements, such as moving straight ahead may not necessarily create a change in the displayed objects. For instance, if the HMD user moves forward in the subway car fromposition 704 toposition 706, the menu may stay substantially the same. Alternatively, displacements in HMD location could represent inputs that may cause the displayed objects to be adjusted. For instance, physical movements of the HMD could represent ‘walking’ through the user interface in three-dimensional space. -
FIG. 7D illustrates anembodiment 732 where the HMD user may be travelling on the subway car atposition 708. At thatposition 708, the HMD user is traveling forward and also rotating gradually to the left. If the user interface is not corrected by an eye-tracking system, the controller may interpret the HMD movement as an intentional movement to turn left. In general, this motion may lead to rotating the menus to the right. In this case, thephotos 720 menu may be moved towards the center of theHMD display 716. - However,
FIG. 7E illustrates anembodiment 736 that may use an eye-tracking system to correctly determine the actual desired speed of user interface movement. In this case, the HMD user is moving throughposition 708, while moving forward and rotating to the left. At the same time, aneye gaze point 738 may be determined to remain at the center of theHMD display 716. The eye gaze fixation may cause the menu to not rotate. - Other embodiments of adjusting the tracking rate of virtual images in an HMD system are possible and are not meant to be limited by the above discussion. Those skilled in the art will understand that HMDs may present many different types of information in the form of virtual images to a wearer. Accordingly, each of these various virtual images may be assigned a different tracking rate and vector, which may be based on at least one or more context-related factors such as wearer reading speed, HMD orientation, HMD motion, HMD location, gaze axis, etc.
- A further example of how gaze direction may be used to correct for motion of the HMD is illustrated by the pseudo-code set forth below in Table 1. The routine may be called every time that there is a new sensor reading relating to the orientation of the HMD, for example, a sensor reading from
gyroscope 120,accelerometer 124, or other component of HMD-Tracking system 104. The routine is able to calculate two variables, adjustedX and adjustedY, which relate to how far the displayed images are to be moved in the x and y directions, respectively, using the moveScreen (adjustedX, adjustedY) function. Specifically, the variables dx and dy, which represent the distances that the wearer's head has travelled in the x and y directions, respectively, are determined from the sensor data. The variables vx and vy, which represent the x and y displacements between the wearer's gaze location on the screen and the center of the screen, may be calculated based on the angle difference between the gaze axis and a reference axis that goes through the center of the screen. The variable adjustedX may then be calculated as a function of dx and vx, and the variable adjustedY may be calculated as a function of dy and vy. In this way, the wearer of the HMD is able to keep the position of the displayed images fixed, notwithstanding motion of the HMD, by gazing at the center of the screen. -
TABLE 1 var radiusX = adjustable number of units; var radiusY = adjustable number of units; // callback method that is called every time there is a new sensor reading. function handleHeadMovement( ) { // calculate the distance the head has moved. This distance might // be faulty due to sensor drift or unintended movement of the wearer, // as when riding the subway or walking around a corner var dx = relative distance that the head travelled on the x axis; var dy = relative distance that the head travelled on the y axis; // calculate the variance of the eye gaze from the center of the screen var vx = distance the eye gaze is from the center of the screen on the x axis; var vy = distance the eye gaze is from the center of the screen on the y axis; // adjust the distance travelled based on the distance the eye gaze is from // the center. Make sure that the distance travelled is only decremented, // never augmented var adjustedX = dx * Math.min( 1, vx / radiusX ); var adjustedY = dy * Math.min( 1, vy / radiusY ); // move the screen to reflect the adjusted movement. moveScreen( adjustedX, adjustedY ); } - 4. Non-Transitory Computer Readable Medium to Determine Speed of Image Movement Using Eye Gaze Detection.
- Some or all of the functions described above in
method 500 and illustrated inFIGS. 5, 6A, 6B, 6C, 6D, 7A, 7B, 7C, 7D, and 7E may be performed by a computing device in response to the execution of instructions stored in a non-transitory computer readable medium. The non-transitory computer readable medium could be, for example, a random access memory (RAM), a read-only memory (ROM), a flash memory, a cache memory, one or more magnetically encoded discs, one or more optically encoded discs, or any other form of non-transitory data storage. The non-transitory computer readable medium could also be distributed among multiple data storage elements, which could be remotely located from each other. The computing device that executes the stored instructions could be a wearable computing device, such as awearable computing device 100 illustrated inFIG. 1 . Alternatively, the computing device that executes the stored instructions could be another computing device, such as a server in a server network. A non-transitory computer readable medium may store instructions executable by theprocessor 112 to perform various functions. For instance, instructions that could be used to carry outmethod 500 may be stored inmemory 114 and could be executed byprocessor 112. In such an embodiment, upon receiving gaze information from the eye-trackingsystem 102, theprocessor 112 carry out instructions to determine a gaze axis and a reference axis as well as to control theHMD 100 to display virtual images within the HMD field of view and adjust a tracking rate based on the gaze axis and the reference axis. - The above detailed description describes various features and functions of the disclosed systems, devices, and methods with reference to the accompanying figures. While various aspects and embodiments have been disclosed herein, other aspects and embodiments will be apparent to those skilled in the art. The various aspects and embodiments disclosed herein are for purposes of illustration and are not intended to be limiting, with the true scope and spirit being indicated by the following claims.
Claims (21)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/096,181 US20160252956A1 (en) | 2011-11-02 | 2016-04-11 | Imaging Method |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/287,390 US8970452B2 (en) | 2011-11-02 | 2011-11-02 | Imaging method |
US14/605,587 US20150169054A1 (en) | 2011-11-02 | 2015-01-26 | Imaging Method |
US15/096,181 US20160252956A1 (en) | 2011-11-02 | 2016-04-11 | Imaging Method |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/605,587 Continuation US20150169054A1 (en) | 2011-11-02 | 2015-01-26 | Imaging Method |
Publications (1)
Publication Number | Publication Date |
---|---|
US20160252956A1 true US20160252956A1 (en) | 2016-09-01 |
Family
ID=48171863
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/287,390 Active 2033-02-04 US8970452B2 (en) | 2011-11-02 | 2011-11-02 | Imaging method |
US14/605,587 Abandoned US20150169054A1 (en) | 2011-11-02 | 2015-01-26 | Imaging Method |
US15/096,181 Abandoned US20160252956A1 (en) | 2011-11-02 | 2016-04-11 | Imaging Method |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/287,390 Active 2033-02-04 US8970452B2 (en) | 2011-11-02 | 2011-11-02 | Imaging method |
US14/605,587 Abandoned US20150169054A1 (en) | 2011-11-02 | 2015-01-26 | Imaging Method |
Country Status (2)
Country | Link |
---|---|
US (3) | US8970452B2 (en) |
WO (1) | WO2013066634A1 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190349575A1 (en) * | 2018-05-14 | 2019-11-14 | Dell Products, L.P. | SYSTEMS AND METHODS FOR USING PERIPHERAL VISION IN VIRTUAL, AUGMENTED, AND MIXED REALITY (xR) APPLICATIONS |
US11126342B2 (en) | 2018-01-04 | 2021-09-21 | Samsung Electronics Co., Ltd. | Electronic device for controlling image display based on scroll input and method thereof |
US11165971B1 (en) | 2020-12-15 | 2021-11-02 | International Business Machines Corporation | Smart contact lens based collaborative video capturing |
Families Citing this family (221)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9952664B2 (en) * | 2014-01-21 | 2018-04-24 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9400390B2 (en) | 2014-01-24 | 2016-07-26 | Osterhout Group, Inc. | Peripheral lighting for head worn computing |
US9229233B2 (en) | 2014-02-11 | 2016-01-05 | Osterhout Group, Inc. | Micro Doppler presentations in head worn computing |
US9298007B2 (en) | 2014-01-21 | 2016-03-29 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9965681B2 (en) | 2008-12-16 | 2018-05-08 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US20150277120A1 (en) | 2014-01-21 | 2015-10-01 | Osterhout Group, Inc. | Optical configurations for head worn computing |
US9715112B2 (en) | 2014-01-21 | 2017-07-25 | Osterhout Group, Inc. | Suppression of stray light in head worn computing |
US9366867B2 (en) | 2014-07-08 | 2016-06-14 | Osterhout Group, Inc. | Optical systems for see-through displays |
US20150205111A1 (en) | 2014-01-21 | 2015-07-23 | Osterhout Group, Inc. | Optical configurations for head worn computing |
US20100306825A1 (en) | 2009-05-27 | 2010-12-02 | Lucid Ventures, Inc. | System and method for facilitating user interaction with a simulated object associated with a physical location |
FI20100380L (en) * | 2010-11-17 | 2012-05-18 | Icergo Oy | The method presents readable information on a digital display board |
US9812091B2 (en) | 2011-02-18 | 2017-11-07 | Kyocera Corporation | Automatic scrolling speed control by tracking user's eye |
JP5606953B2 (en) * | 2011-02-18 | 2014-10-15 | 京セラ株式会社 | Information display device |
US8879155B1 (en) | 2011-11-09 | 2014-11-04 | Google Inc. | Measurement method and system |
US10598929B2 (en) | 2011-11-09 | 2020-03-24 | Google Llc | Measurement method and system |
US10354291B1 (en) | 2011-11-09 | 2019-07-16 | Google Llc | Distributing media to displays |
KR101891786B1 (en) * | 2011-11-29 | 2018-08-27 | 삼성전자주식회사 | Operation Method For User Function based on a Eye-Tracking and Portable Device supporting the same |
US9497501B2 (en) | 2011-12-06 | 2016-11-15 | Microsoft Technology Licensing, Llc | Augmented reality virtual monitor |
US9223138B2 (en) | 2011-12-23 | 2015-12-29 | Microsoft Technology Licensing, Llc | Pixel opacity for augmented reality |
US8941722B2 (en) * | 2012-01-03 | 2015-01-27 | Sony Corporation | Automatic intelligent focus control of video |
US9829715B2 (en) | 2012-01-23 | 2017-11-28 | Nvidia Corporation | Eyewear device for transmitting signal and communication method thereof |
US9606586B2 (en) | 2012-01-23 | 2017-03-28 | Microsoft Technology Licensing, Llc | Heat transfer device |
US20130201305A1 (en) * | 2012-02-06 | 2013-08-08 | Research In Motion Corporation | Division of a graphical display into regions |
US9726887B2 (en) | 2012-02-15 | 2017-08-08 | Microsoft Technology Licensing, Llc | Imaging structure color conversion |
US9368546B2 (en) | 2012-02-15 | 2016-06-14 | Microsoft Technology Licensing, Llc | Imaging structure with embedded light sources |
US9297996B2 (en) | 2012-02-15 | 2016-03-29 | Microsoft Technology Licensing, Llc | Laser illumination scanning |
US9779643B2 (en) | 2012-02-15 | 2017-10-03 | Microsoft Technology Licensing, Llc | Imaging structure emitter configurations |
US9578318B2 (en) | 2012-03-14 | 2017-02-21 | Microsoft Technology Licensing, Llc | Imaging structure emitter calibration |
US10469916B1 (en) | 2012-03-23 | 2019-11-05 | Google Llc | Providing media content to a wearable device |
US11068049B2 (en) | 2012-03-23 | 2021-07-20 | Microsoft Technology Licensing, Llc | Light guide display and field of view |
US9558590B2 (en) * | 2012-03-28 | 2017-01-31 | Microsoft Technology Licensing, Llc | Augmented reality light guide display |
US10191515B2 (en) * | 2012-03-28 | 2019-01-29 | Microsoft Technology Licensing, Llc | Mobile device light guide display |
US9471833B1 (en) * | 2012-04-03 | 2016-10-18 | Intuit Inc. | Character recognition using images at different angles |
US9717981B2 (en) | 2012-04-05 | 2017-08-01 | Microsoft Technology Licensing, Llc | Augmented reality and physical games |
US20130293580A1 (en) * | 2012-05-01 | 2013-11-07 | Zambala Lllp | System and method for selecting targets in an augmented reality environment |
US8893164B1 (en) | 2012-05-16 | 2014-11-18 | Google Inc. | Audio system |
US10502876B2 (en) | 2012-05-22 | 2019-12-10 | Microsoft Technology Licensing, Llc | Waveguide optics focus elements |
US8989535B2 (en) | 2012-06-04 | 2015-03-24 | Microsoft Technology Licensing, Llc | Multiple waveguide imaging structure |
US9557565B2 (en) * | 2012-07-02 | 2017-01-31 | Nvidia Corporation | Near-eye optical deconvolution displays |
US9494797B2 (en) * | 2012-07-02 | 2016-11-15 | Nvidia Corporation | Near-eye parallax barrier displays |
US9841537B2 (en) | 2012-07-02 | 2017-12-12 | Nvidia Corporation | Near-eye microlens array displays |
USRE47984E1 (en) * | 2012-07-02 | 2020-05-12 | Nvidia Corporation | Near-eye optical deconvolution displays |
US9142185B2 (en) * | 2012-08-30 | 2015-09-22 | Atheer, Inc. | Method and apparatus for selectively presenting content |
US9268136B1 (en) * | 2012-09-28 | 2016-02-23 | Google Inc. | Use of comparative sensor data to determine orientation of head relative to body |
JP6040715B2 (en) * | 2012-11-06 | 2016-12-07 | ソニー株式会社 | Image display apparatus, image display method, and computer program |
US20140152558A1 (en) * | 2012-11-30 | 2014-06-05 | Tom Salter | Direct hologram manipulation using imu |
US10192358B2 (en) | 2012-12-20 | 2019-01-29 | Microsoft Technology Licensing, Llc | Auto-stereoscopic augmented reality display |
WO2014115387A1 (en) * | 2013-01-28 | 2014-07-31 | ソニー株式会社 | Information processor, information processing method and program |
US20150193061A1 (en) * | 2013-01-29 | 2015-07-09 | Google Inc. | User's computing experience based on the user's computing activity |
US9791921B2 (en) | 2013-02-19 | 2017-10-17 | Microsoft Technology Licensing, Llc | Context-aware augmented reality object commands |
US9864498B2 (en) * | 2013-03-13 | 2018-01-09 | Tobii Ab | Automatic scrolling based on gaze detection |
US9619020B2 (en) | 2013-03-01 | 2017-04-11 | Tobii Ab | Delay warp gaze interaction |
KR20230173231A (en) * | 2013-03-11 | 2023-12-26 | 매직 립, 인코포레이티드 | System and method for augmented and virtual reality |
US20140267581A1 (en) | 2013-03-15 | 2014-09-18 | John Cronin | Real time virtual reality leveraging web cams and ip cams and web cam and ip cam networks |
US20140280505A1 (en) | 2013-03-15 | 2014-09-18 | John Cronin | Virtual reality interaction with 3d printing |
US20140280503A1 (en) | 2013-03-15 | 2014-09-18 | John Cronin | System and methods for effective virtual reality visitor interface |
US20140280644A1 (en) | 2013-03-15 | 2014-09-18 | John Cronin | Real time unified communications interaction of a predefined location in a virtual reality location |
US20140280506A1 (en) | 2013-03-15 | 2014-09-18 | John Cronin | Virtual reality enhanced through browser connections |
KR102560629B1 (en) | 2013-03-15 | 2023-07-26 | 매직 립, 인코포레이티드 | Display system and method |
US9838506B1 (en) | 2013-03-15 | 2017-12-05 | Sony Interactive Entertainment America Llc | Virtual reality universe representation changes viewing based upon client side parameters |
US20140280502A1 (en) | 2013-03-15 | 2014-09-18 | John Cronin | Crowd and cloud enabled virtual reality distributed location network |
TWI649675B (en) * | 2013-03-28 | 2019-02-01 | 新力股份有限公司 | Display device |
DE102013210354A1 (en) | 2013-06-04 | 2014-12-04 | Bayerische Motoren Werke Aktiengesellschaft | Eye-controlled interaction for data glasses |
DE102013210588A1 (en) | 2013-06-07 | 2014-12-11 | Bayerische Motoren Werke Aktiengesellschaft | Display system with data glasses |
DE102013210587A1 (en) | 2013-06-07 | 2014-12-24 | Bayerische Motoren Werke Aktiengesellschaft | Display system with data glasses |
GB201310368D0 (en) * | 2013-06-11 | 2013-07-24 | Sony Comp Entertainment Europe | Head-mountable apparatus and systems |
GB201310367D0 (en) * | 2013-06-11 | 2013-07-24 | Sony Comp Entertainment Europe | Head-mountable apparatus and systems |
US10175483B2 (en) * | 2013-06-18 | 2019-01-08 | Microsoft Technology Licensing, Llc | Hybrid world/body locked HUD on an HMD |
US9625723B2 (en) * | 2013-06-25 | 2017-04-18 | Microsoft Technology Licensing, Llc | Eye-tracking system using a freeform prism |
US10228561B2 (en) | 2013-06-25 | 2019-03-12 | Microsoft Technology Licensing, Llc | Eye-tracking system using a freeform prism and gaze-detection light |
US9582075B2 (en) | 2013-07-19 | 2017-02-28 | Nvidia Corporation | Gaze-tracking eye illumination from display |
US9880325B2 (en) | 2013-08-14 | 2018-01-30 | Nvidia Corporation | Hybrid optics for near-eye displays |
DE102013013698B4 (en) * | 2013-08-16 | 2024-10-02 | Audi Ag | Method for operating electronic data glasses |
US9665172B2 (en) * | 2013-09-03 | 2017-05-30 | Tobii Ab | Portable eye tracking device |
WO2015034801A2 (en) * | 2013-09-06 | 2015-03-12 | 3M Innovative Properties Company | Head mounted display with eye tracking |
KR20150037254A (en) * | 2013-09-30 | 2015-04-08 | 엘지전자 주식회사 | Wearable display device and method of controlling layer |
US10405786B2 (en) | 2013-10-09 | 2019-09-10 | Nedim T. SAHIN | Systems, environment and methods for evaluation and management of autism spectrum disorder using a wearable data collection device |
US9936916B2 (en) | 2013-10-09 | 2018-04-10 | Nedim T. SAHIN | Systems, environment and methods for identification and analysis of recurring transitory physiological states and events using a portable data collection device |
TW201516467A (en) * | 2013-10-25 | 2015-05-01 | Quanta Comp Inc | Head mounted display and imaging method thereof |
US10558262B2 (en) | 2013-11-18 | 2020-02-11 | Tobii Ab | Component determination and gaze provoked interaction |
US9949637B1 (en) | 2013-11-25 | 2018-04-24 | Verily Life Sciences Llc | Fluorescent imaging on a head-mountable device |
US9760898B2 (en) * | 2014-01-06 | 2017-09-12 | The Nielsen Company (Us), Llc | Methods and apparatus to detect engagement with media presented on wearable media devices |
US9594246B2 (en) | 2014-01-21 | 2017-03-14 | Osterhout Group, Inc. | See-through computer display systems |
US9529195B2 (en) | 2014-01-21 | 2016-12-27 | Osterhout Group, Inc. | See-through computer display systems |
US10254856B2 (en) | 2014-01-17 | 2019-04-09 | Osterhout Group, Inc. | External user interface for head worn computing |
US9671613B2 (en) | 2014-09-26 | 2017-06-06 | Osterhout Group, Inc. | See-through computer display systems |
US9299194B2 (en) | 2014-02-14 | 2016-03-29 | Osterhout Group, Inc. | Secure sharing in head worn computing |
US9939934B2 (en) | 2014-01-17 | 2018-04-10 | Osterhout Group, Inc. | External user interface for head worn computing |
US9575321B2 (en) | 2014-06-09 | 2017-02-21 | Osterhout Group, Inc. | Content presentation in head worn computing |
US10649220B2 (en) | 2014-06-09 | 2020-05-12 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11103122B2 (en) | 2014-07-15 | 2021-08-31 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US10684687B2 (en) | 2014-12-03 | 2020-06-16 | Mentor Acquisition One, Llc | See-through computer display systems |
US11227294B2 (en) | 2014-04-03 | 2022-01-18 | Mentor Acquisition One, Llc | Sight information collection in head worn computing |
US9810906B2 (en) | 2014-06-17 | 2017-11-07 | Osterhout Group, Inc. | External user interface for head worn computing |
US10191279B2 (en) | 2014-03-17 | 2019-01-29 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US20160019715A1 (en) | 2014-07-15 | 2016-01-21 | Osterhout Group, Inc. | Content presentation in head worn computing |
US9829707B2 (en) | 2014-08-12 | 2017-11-28 | Osterhout Group, Inc. | Measuring content brightness in head worn computing |
US9746686B2 (en) | 2014-05-19 | 2017-08-29 | Osterhout Group, Inc. | Content position calibration in head worn computing |
US9841599B2 (en) | 2014-06-05 | 2017-12-12 | Osterhout Group, Inc. | Optical configurations for head-worn see-through displays |
US9366868B2 (en) | 2014-09-26 | 2016-06-14 | Osterhout Group, Inc. | See-through computer display systems |
US9448409B2 (en) | 2014-11-26 | 2016-09-20 | Osterhout Group, Inc. | See-through computer display systems |
US20150277118A1 (en) | 2014-03-28 | 2015-10-01 | Osterhout Group, Inc. | Sensor dependent content position in head worn computing |
US11892644B2 (en) | 2014-01-21 | 2024-02-06 | Mentor Acquisition One, Llc | See-through computer display systems |
US9753288B2 (en) | 2014-01-21 | 2017-09-05 | Osterhout Group, Inc. | See-through computer display systems |
US9651784B2 (en) | 2014-01-21 | 2017-05-16 | Osterhout Group, Inc. | See-through computer display systems |
US9532714B2 (en) | 2014-01-21 | 2017-01-03 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US11737666B2 (en) | 2014-01-21 | 2023-08-29 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US11669163B2 (en) | 2014-01-21 | 2023-06-06 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US9811153B2 (en) | 2014-01-21 | 2017-11-07 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9766463B2 (en) | 2014-01-21 | 2017-09-19 | Osterhout Group, Inc. | See-through computer display systems |
US12105281B2 (en) | 2014-01-21 | 2024-10-01 | Mentor Acquisition One, Llc | See-through computer display systems |
US20150205135A1 (en) | 2014-01-21 | 2015-07-23 | Osterhout Group, Inc. | See-through computer display systems |
US9651788B2 (en) | 2014-01-21 | 2017-05-16 | Osterhout Group, Inc. | See-through computer display systems |
US9811159B2 (en) | 2014-01-21 | 2017-11-07 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9310610B2 (en) | 2014-01-21 | 2016-04-12 | Osterhout Group, Inc. | See-through computer display systems |
US9494800B2 (en) | 2014-01-21 | 2016-11-15 | Osterhout Group, Inc. | See-through computer display systems |
US11487110B2 (en) | 2014-01-21 | 2022-11-01 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US12093453B2 (en) | 2014-01-21 | 2024-09-17 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US9836122B2 (en) | 2014-01-21 | 2017-12-05 | Osterhout Group, Inc. | Eye glint imaging in see-through computer display systems |
US9846308B2 (en) | 2014-01-24 | 2017-12-19 | Osterhout Group, Inc. | Haptic systems for head-worn computers |
US9437159B2 (en) | 2014-01-25 | 2016-09-06 | Sony Interactive Entertainment America Llc | Environmental interrupt in a head-mounted display and utilization of non field of view real estate |
US9588343B2 (en) | 2014-01-25 | 2017-03-07 | Sony Interactive Entertainment America Llc | Menu navigation in a head-mounted display |
US10353460B2 (en) * | 2014-01-29 | 2019-07-16 | Tarek A Shazly | Eye and head tracking device |
JP5956479B2 (en) * | 2014-01-29 | 2016-07-27 | 株式会社東芝 | Display device and gaze estimation device |
US10067341B1 (en) | 2014-02-04 | 2018-09-04 | Intelligent Technologies International, Inc. | Enhanced heads-up display system |
US9401540B2 (en) | 2014-02-11 | 2016-07-26 | Osterhout Group, Inc. | Spatial location presentation in head worn computing |
US9773349B2 (en) * | 2014-02-19 | 2017-09-26 | Daqri, Llc | Active parallax correction |
TWI489140B (en) * | 2014-02-24 | 2015-06-21 | Quanta Comp Inc | Head mounted display apparatus |
CA2939922A1 (en) | 2014-02-24 | 2015-08-27 | Brain Power, Llc | Systems, environment and methods for evaluation and management of autism spectrum disorder using a wearable data collection device |
US20170123491A1 (en) * | 2014-03-17 | 2017-05-04 | Itu Business Development A/S | Computer-implemented gaze interaction method and apparatus |
JP6326901B2 (en) * | 2014-03-26 | 2018-05-23 | 富士ゼロックス株式会社 | Image processing apparatus and program |
US20160187651A1 (en) | 2014-03-28 | 2016-06-30 | Osterhout Group, Inc. | Safety for a vehicle operator with an hmd |
KR20160143749A (en) | 2014-04-09 | 2016-12-14 | 쓰리엠 이노베이티브 프로퍼티즈 컴파니 | Head mounted display and low conspicuity pupil illuminator |
EP2933707B1 (en) * | 2014-04-14 | 2017-12-06 | iOnRoad Technologies Ltd. | Head mounted display presentation adjustment |
US9423842B2 (en) | 2014-09-18 | 2016-08-23 | Osterhout Group, Inc. | Thermal management for head-worn computer |
US9672210B2 (en) | 2014-04-25 | 2017-06-06 | Osterhout Group, Inc. | Language translation with head-worn computing |
US9651787B2 (en) | 2014-04-25 | 2017-05-16 | Osterhout Group, Inc. | Speaker assembly for headworn computer |
US10853589B2 (en) | 2014-04-25 | 2020-12-01 | Mentor Acquisition One, Llc | Language translation with head-worn computing |
EP3144775B1 (en) | 2014-05-09 | 2020-07-08 | Sony Corporation | Information processing system and information processing method |
US10663740B2 (en) | 2014-06-09 | 2020-05-26 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US9304235B2 (en) | 2014-07-30 | 2016-04-05 | Microsoft Technology Licensing, Llc | Microfabrication |
US10254942B2 (en) | 2014-07-31 | 2019-04-09 | Microsoft Technology Licensing, Llc | Adaptive sizing and positioning of application windows |
US10678412B2 (en) | 2014-07-31 | 2020-06-09 | Microsoft Technology Licensing, Llc | Dynamic joint dividers for application windows |
US10592080B2 (en) | 2014-07-31 | 2020-03-17 | Microsoft Technology Licensing, Llc | Assisted presentation of application windows |
FR3025711B1 (en) * | 2014-09-15 | 2020-12-25 | Acep France | DEVICE FOR MEASURING THE PHYSIOLOGICAL CHARACTERISTICS OF THE EYE, FOR A PATIENT WEARING TINTED GLASSES |
KR102194787B1 (en) * | 2014-09-24 | 2020-12-24 | 삼성전자주식회사 | Apparatus and method for user based sensor data acquiring |
JP2016090689A (en) * | 2014-10-31 | 2016-05-23 | セイコーエプソン株式会社 | Image display device and drawing method |
DE102014222355A1 (en) | 2014-11-03 | 2016-05-04 | Bayerische Motoren Werke Aktiengesellschaft | Fatigue detection with sensors of data glasses |
EP3220374A4 (en) * | 2014-11-12 | 2018-07-18 | Fujitsu Limited | Wearable device, display control method, and display control program |
US9684172B2 (en) | 2014-12-03 | 2017-06-20 | Osterhout Group, Inc. | Head worn computer display systems |
GB2533366A (en) | 2014-12-18 | 2016-06-22 | Ibm | Methods of controlling document display devices and document display devices |
USD743963S1 (en) | 2014-12-22 | 2015-11-24 | Osterhout Group, Inc. | Air mouse |
USD751552S1 (en) | 2014-12-31 | 2016-03-15 | Osterhout Group, Inc. | Computer glasses |
USD753114S1 (en) | 2015-01-05 | 2016-04-05 | Osterhout Group, Inc. | Air mouse |
US10740971B2 (en) * | 2015-01-20 | 2020-08-11 | Microsoft Technology Licensing, Llc | Augmented reality field of view object follower |
DE102015200750A1 (en) * | 2015-01-20 | 2016-07-21 | Bayerische Motoren Werke Aktiengesellschaft | Display of representations on an HMD |
US9372347B1 (en) | 2015-02-09 | 2016-06-21 | Microsoft Technology Licensing, Llc | Display system |
US9827209B2 (en) | 2015-02-09 | 2017-11-28 | Microsoft Technology Licensing, Llc | Display system |
US9513480B2 (en) | 2015-02-09 | 2016-12-06 | Microsoft Technology Licensing, Llc | Waveguide |
US11086216B2 (en) | 2015-02-09 | 2021-08-10 | Microsoft Technology Licensing, Llc | Generating electronic components |
US9429692B1 (en) | 2015-02-09 | 2016-08-30 | Microsoft Technology Licensing, Llc | Optical components |
US10317677B2 (en) | 2015-02-09 | 2019-06-11 | Microsoft Technology Licensing, Llc | Display system |
US10018844B2 (en) | 2015-02-09 | 2018-07-10 | Microsoft Technology Licensing, Llc | Wearable image display system |
US9535253B2 (en) | 2015-02-09 | 2017-01-03 | Microsoft Technology Licensing, Llc | Display system |
US9423360B1 (en) | 2015-02-09 | 2016-08-23 | Microsoft Technology Licensing, Llc | Optical components |
US20160239985A1 (en) | 2015-02-17 | 2016-08-18 | Osterhout Group, Inc. | See-through computer display systems |
WO2016168785A1 (en) | 2015-04-17 | 2016-10-20 | Tulip Interfaces, Inc. | Augmented manufacturing system |
EP3308539A1 (en) * | 2015-06-12 | 2018-04-18 | Microsoft Technology Licensing, LLC | Display for stereoscopic augmented reality |
EP3109734A1 (en) * | 2015-06-22 | 2016-12-28 | Samsung Electronics Co., Ltd | Three-dimensional user interface for head-mountable display |
US10416835B2 (en) | 2015-06-22 | 2019-09-17 | Samsung Electronics Co., Ltd. | Three-dimensional user interface for head-mountable display |
US9939644B2 (en) | 2015-06-25 | 2018-04-10 | Intel Corporation | Technologies for controlling vision correction of a wearable computing device |
CN104991345A (en) * | 2015-07-31 | 2015-10-21 | 北京亮亮视野科技有限公司 | Mobile display system and smart glasses with application of mobile display system |
JP6334484B2 (en) * | 2015-09-01 | 2018-05-30 | 株式会社東芝 | Glasses-type wearable device, control method thereof, and information management server |
US9726891B2 (en) * | 2015-09-03 | 2017-08-08 | Microsoft Technology Licensing, Llc | Left and right eye optical paths with shared optical element for head-mounted display device |
US20170092002A1 (en) * | 2015-09-30 | 2017-03-30 | Daqri, Llc | User interface for augmented reality system |
EP3367213A4 (en) * | 2015-10-22 | 2019-04-17 | LG Electronics Inc. | Mobile terminal and control method therefor |
US10466780B1 (en) * | 2015-10-26 | 2019-11-05 | Pillantas | Systems and methods for eye tracking calibration, eye vergence gestures for interface control, and visual aids therefor |
US10338677B2 (en) * | 2015-10-28 | 2019-07-02 | Microsoft Technology Licensing, Llc | Adjusting image frames based on tracking motion of eyes |
JP6560974B2 (en) | 2015-12-17 | 2019-08-14 | 株式会社ソニー・インタラクティブエンタテインメント | Information processing apparatus and operation reception method |
WO2017147801A1 (en) * | 2016-03-02 | 2017-09-08 | 陈台国 | System for adjusting depth of field of multi-depth display and method thereof |
US10466491B2 (en) | 2016-06-01 | 2019-11-05 | Mentor Acquisition One, Llc | Modular systems for head-worn computers |
US10824253B2 (en) | 2016-05-09 | 2020-11-03 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US9910284B1 (en) | 2016-09-08 | 2018-03-06 | Osterhout Group, Inc. | Optical systems for head-worn computers |
US10684478B2 (en) | 2016-05-09 | 2020-06-16 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US10134174B2 (en) * | 2016-06-13 | 2018-11-20 | Microsoft Technology Licensing, Llc | Texture mapping with render-baked animation |
US10955987B2 (en) * | 2016-10-04 | 2021-03-23 | Facebook, Inc. | Three-dimensional user interface |
CN106531073B (en) * | 2017-01-03 | 2018-11-20 | 京东方科技集团股份有限公司 | Processing circuit, display methods and the display device of display screen |
US10341606B2 (en) * | 2017-05-24 | 2019-07-02 | SA Photonics, Inc. | Systems and method of transmitting information from monochrome sensors |
CN107272904B (en) * | 2017-06-28 | 2021-05-18 | 联想(北京)有限公司 | Image display method and electronic equipment |
US10108261B1 (en) * | 2017-07-05 | 2018-10-23 | Oculus Vr, Llc | Eye tracking based on light polarization |
US10578869B2 (en) | 2017-07-24 | 2020-03-03 | Mentor Acquisition One, Llc | See-through computer display systems with adjustable zoom cameras |
US10422995B2 (en) | 2017-07-24 | 2019-09-24 | Mentor Acquisition One, Llc | See-through computer display systems with stray light management |
US11409105B2 (en) | 2017-07-24 | 2022-08-09 | Mentor Acquisition One, Llc | See-through computer display systems |
US10969584B2 (en) | 2017-08-04 | 2021-04-06 | Mentor Acquisition One, Llc | Image expansion optic for head-worn computer |
US10394034B2 (en) * | 2017-08-15 | 2019-08-27 | Microsoft Technology Licensing, Llc | Eye-tracking with MEMS scanning and optical relay |
WO2019084325A1 (en) * | 2017-10-27 | 2019-05-02 | Magic Leap, Inc. | Virtual reticle for augmented reality systems |
US10311584B1 (en) | 2017-11-09 | 2019-06-04 | Facebook Technologies, Llc | Estimation of absolute depth from polarization measurements |
TWI644126B (en) * | 2017-11-28 | 2018-12-11 | 沅聖科技股份有限公司 | Head mounted display and image device |
WO2019108211A1 (en) * | 2017-11-30 | 2019-06-06 | Hewlett-Packard Development Company, L.P. | Augmented reality based virtual dashboard implementations |
WO2019143117A1 (en) | 2018-01-18 | 2019-07-25 | Samsung Electronics Co., Ltd. | Method and apparatus for adjusting augmented reality content |
JP6582205B2 (en) * | 2018-02-28 | 2019-10-02 | 株式会社コナミデジタルエンタテインメント | Information processing apparatus, information processing apparatus program, head mounted display, and information processing system |
US10878781B2 (en) * | 2018-08-09 | 2020-12-29 | Chun-Ding HUANG | Image processing method and head-mounted display system |
US20200125169A1 (en) * | 2018-10-18 | 2020-04-23 | Eyetech Digital Systems, Inc. | Systems and Methods for Correcting Lens Distortion in Head Mounted Displays |
CN109188700B (en) * | 2018-10-30 | 2021-05-11 | 京东方科技集团股份有限公司 | Optical display system and AR/VR display device |
US10630925B1 (en) | 2018-12-03 | 2020-04-21 | Facebook Technologies, Llc | Depth determination using polarization of light and camera assembly with augmented pixels |
US10791282B2 (en) | 2018-12-13 | 2020-09-29 | Fenwick & West LLP | High dynamic range camera assembly with augmented pixels |
US10855896B1 (en) | 2018-12-13 | 2020-12-01 | Facebook Technologies, Llc | Depth determination using time-of-flight and camera assembly with augmented pixels |
US10791286B2 (en) | 2018-12-13 | 2020-09-29 | Facebook Technologies, Llc | Differentiated imaging using camera assembly with augmented pixels |
CN109656373B (en) * | 2019-01-02 | 2020-11-10 | 京东方科技集团股份有限公司 | Fixation point positioning method and positioning device, display equipment and storage medium |
CN109542240B (en) * | 2019-02-01 | 2020-07-10 | 京东方科技集团股份有限公司 | Eyeball tracking device and method |
CN110263657B (en) * | 2019-05-24 | 2023-04-18 | 亿信科技发展有限公司 | Human eye tracking method, device, system, equipment and storage medium |
CN110267025B (en) * | 2019-07-03 | 2021-04-13 | 京东方科技集团股份有限公司 | Rendering method and device for virtual 3D display and display method and system thereof |
US10902623B1 (en) | 2019-11-19 | 2021-01-26 | Facebook Technologies, Llc | Three-dimensional imaging with spatial and temporal coding for depth camera assembly |
US11194160B1 (en) | 2020-01-21 | 2021-12-07 | Facebook Technologies, Llc | High frame rate reconstruction with N-tap camera sensor |
US11733530B1 (en) | 2020-09-24 | 2023-08-22 | Apple Inc. | Head-mountable device having light seal element with adjustable opacity |
CN112416125A (en) * | 2020-11-17 | 2021-02-26 | 青岛小鸟看看科技有限公司 | VR head-mounted all-in-one machine |
CN114527864B (en) * | 2020-11-19 | 2024-03-15 | 京东方科技集团股份有限公司 | Augmented reality text display system, method, equipment and medium |
KR20240093875A (en) * | 2020-12-17 | 2024-06-24 | 진 랩스 인코포레이티드 | Event camera system for pupil detection and eye tracking |
US11681370B2 (en) * | 2021-09-16 | 2023-06-20 | Htc Corporation | Handheld controller and control method |
US12093450B2 (en) * | 2022-10-25 | 2024-09-17 | Meta Platforms Technologies, Llc | Scanning display with eye-tracking |
US11988828B1 (en) * | 2022-11-16 | 2024-05-21 | Meta Platforms Technologies, Llc | Multi-pupil display and eye-tracking with interferometric sensing |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6637883B1 (en) * | 2003-01-23 | 2003-10-28 | Vishwas V. Tengshe | Gaze tracking system and method |
Family Cites Families (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004280127A (en) | 1994-06-23 | 2004-10-07 | Seiko Epson Corp | Head mounted type display device |
US6351273B1 (en) | 1997-04-30 | 2002-02-26 | Jerome H. Lemelson | System and methods for controlling automatic scrolling of information on a display or screen |
CA2310114A1 (en) | 1998-02-02 | 1999-08-02 | Steve Mann | Wearable camera system with viewfinder means |
US6603491B2 (en) * | 2000-05-26 | 2003-08-05 | Jerome H. Lemelson | System and methods for controlling automatic scrolling of information on a display or screen |
US6578962B1 (en) | 2001-04-27 | 2003-06-17 | International Business Machines Corporation | Calibration-free eye gaze tracking |
US6886137B2 (en) | 2001-05-29 | 2005-04-26 | International Business Machines Corporation | Eye gaze control of dynamic information presentation |
US7306337B2 (en) | 2003-03-06 | 2007-12-11 | Rensselaer Polytechnic Institute | Calibration-free gaze tracking under natural head movement |
US7401920B1 (en) | 2003-05-20 | 2008-07-22 | Elbit Systems Ltd. | Head mounted eye tracking and display system |
US7963652B2 (en) | 2003-11-14 | 2011-06-21 | Queen's University At Kingston | Method and apparatus for calibration-free eye tracking |
KR100749033B1 (en) | 2004-03-03 | 2007-08-13 | 학교법인 한국정보통신학원 | A method for manipulating a terminal using user's glint, and an apparatus |
US20110077548A1 (en) | 2004-04-01 | 2011-03-31 | Torch William C | Biosensors, communicators, and controllers monitoring eye movement and methods for using them |
CN102068237A (en) | 2004-04-01 | 2011-05-25 | 威廉·C·托奇 | Controllers and Methods for Monitoring Eye Movement, System and Method for Controlling Calculation Device |
US7331929B2 (en) | 2004-10-01 | 2008-02-19 | General Electric Company | Method and apparatus for surgical operating room information display gaze detection and user prioritization for control |
WO2007085682A1 (en) | 2006-01-26 | 2007-08-02 | Nokia Corporation | Eye tracker device |
US9323055B2 (en) | 2006-05-26 | 2016-04-26 | Exelis, Inc. | System and method to display maintenance and operational instructions of an apparatus using augmented reality |
US7542210B2 (en) | 2006-06-29 | 2009-06-02 | Chirieleison Sr Anthony | Eye tracking head mounted display |
JP2009157634A (en) | 2007-12-26 | 2009-07-16 | Fuji Xerox Co Ltd | Irradiation control device, irradiation control program, and visual line analysis system |
US8436788B2 (en) | 2008-01-08 | 2013-05-07 | Lockheed Martin Corporation | Method and apparatus for displaying |
WO2009101238A1 (en) | 2008-02-14 | 2009-08-20 | Nokia Corporation | Device and method for determining gaze direction |
KR101004930B1 (en) | 2008-07-10 | 2010-12-28 | 성균관대학교산학협력단 | Full browsing method using gaze detection and handheld terminal performing the method |
JP5565258B2 (en) * | 2010-10-12 | 2014-08-06 | ソニー株式会社 | Image processing apparatus, image processing method, and program |
US9632315B2 (en) * | 2010-10-21 | 2017-04-25 | Lockheed Martin Corporation | Head-mounted display apparatus employing one or more fresnel lenses |
US8611015B2 (en) | 2011-11-22 | 2013-12-17 | Google Inc. | User interface |
-
2011
- 2011-11-02 US US13/287,390 patent/US8970452B2/en active Active
-
2012
- 2012-10-18 WO PCT/US2012/060846 patent/WO2013066634A1/en active Application Filing
-
2015
- 2015-01-26 US US14/605,587 patent/US20150169054A1/en not_active Abandoned
-
2016
- 2016-04-11 US US15/096,181 patent/US20160252956A1/en not_active Abandoned
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6637883B1 (en) * | 2003-01-23 | 2003-10-28 | Vishwas V. Tengshe | Gaze tracking system and method |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11126342B2 (en) | 2018-01-04 | 2021-09-21 | Samsung Electronics Co., Ltd. | Electronic device for controlling image display based on scroll input and method thereof |
US20190349575A1 (en) * | 2018-05-14 | 2019-11-14 | Dell Products, L.P. | SYSTEMS AND METHODS FOR USING PERIPHERAL VISION IN VIRTUAL, AUGMENTED, AND MIXED REALITY (xR) APPLICATIONS |
US11595637B2 (en) * | 2018-05-14 | 2023-02-28 | Dell Products, L.P. | Systems and methods for using peripheral vision in virtual, augmented, and mixed reality (xR) applications |
US11165971B1 (en) | 2020-12-15 | 2021-11-02 | International Business Machines Corporation | Smart contact lens based collaborative video capturing |
Also Published As
Publication number | Publication date |
---|---|
US20150169054A1 (en) | 2015-06-18 |
US20130106674A1 (en) | 2013-05-02 |
WO2013066634A1 (en) | 2013-05-10 |
US8970452B2 (en) | 2015-03-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8970452B2 (en) | Imaging method | |
US8786953B2 (en) | User interface | |
US8767306B1 (en) | Display system | |
US9213185B1 (en) | Display scaling based on movement of a head-mounted display | |
US10055642B2 (en) | Staredown to produce changes in information density and type | |
US8971570B1 (en) | Dual LED usage for glint detection | |
US8955973B2 (en) | Method and system for input detection using structured light projection | |
US8939584B2 (en) | Unlocking method for a computing system | |
US9285872B1 (en) | Using head gesture and eye position to wake a head mounted device | |
US20130241805A1 (en) | Using Convergence Angle to Select Among Different UI Elements | |
US9007301B1 (en) | User interface | |
US20130088413A1 (en) | Method to Autofocus on Near-Eye Display | |
US9261959B1 (en) | Input detection | |
US20140247286A1 (en) | Active Stabilization for Heads-Up Displays | |
US20150153572A1 (en) | Adjustment of Location of Superimposed Image | |
US20150003819A1 (en) | Camera auto-focus based on eye gaze | |
US20150084864A1 (en) | Input Method | |
US20130222638A1 (en) | Image Capture Based on Gaze Detection | |
US9335919B2 (en) | Virtual shade | |
US20210248766A1 (en) | Electronic device and operation method thereof | |
KR20240030881A (en) | Method for outputting a virtual content and an electronic device supporting the same | |
CN118394205A (en) | Mixed reality interactions using eye tracking techniques |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: GOOGLE INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WHEELER, AARON JOSEPH;RAFFLE, HAYES SOLOS;SIGNING DATES FROM 20160412 TO 20160413;REEL/FRAME:038268/0856 |
|
AS | Assignment |
Owner name: GOOGLE LLC, CALIFORNIA Free format text: CHANGE OF NAME;ASSIGNOR:GOOGLE INC.;REEL/FRAME:044129/0001 Effective date: 20170929 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |