US20160274365A1 - Systems, devices, and methods for wearable heads-up displays with heterogeneous display quality - Google Patents
Systems, devices, and methods for wearable heads-up displays with heterogeneous display quality Download PDFInfo
- Publication number
- US20160274365A1 US20160274365A1 US15/070,887 US201615070887A US2016274365A1 US 20160274365 A1 US20160274365 A1 US 20160274365A1 US 201615070887 A US201615070887 A US 201615070887A US 2016274365 A1 US2016274365 A1 US 2016274365A1
- Authority
- US
- United States
- Prior art keywords
- user
- region
- interest
- virtual content
- display
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title claims abstract description 47
- 238000012545 processing Methods 0.000 claims abstract description 22
- 230000002093 peripheral effect Effects 0.000 abstract description 2
- 230000004304 visual acuity Effects 0.000 abstract description 2
- 230000008878 coupling Effects 0.000 description 11
- 238000010168 coupling process Methods 0.000 description 11
- 238000005859 coupling reaction Methods 0.000 description 11
- 230000003287 optical effect Effects 0.000 description 10
- 238000010586 diagram Methods 0.000 description 7
- 210000003128 head Anatomy 0.000 description 7
- 230000003122 modulative effect Effects 0.000 description 7
- 210000001747 pupil Anatomy 0.000 description 6
- 238000013461 design Methods 0.000 description 5
- 230000037361 pathway Effects 0.000 description 5
- 210000001525 retina Anatomy 0.000 description 5
- 230000003247 decreasing effect Effects 0.000 description 4
- 239000000463 material Substances 0.000 description 4
- 238000001816 cooling Methods 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 238000005286 illumination Methods 0.000 description 2
- 210000001210 retinal vessel Anatomy 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 239000003795 chemical substances by application Substances 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 210000004087 cornea Anatomy 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 230000004438 eyesight Effects 0.000 description 1
- 210000000873 fovea centralis Anatomy 0.000 description 1
- 230000007274 generation of a signal involved in cell-cell signaling Effects 0.000 description 1
- 239000011521 glass Substances 0.000 description 1
- 210000000554 iris Anatomy 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000001404 mediated effect Effects 0.000 description 1
- 230000000116 mitigating effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 230000002207 retinal effect Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 229910052709 silver Inorganic materials 0.000 description 1
- 239000004332 silver Substances 0.000 description 1
- -1 silver halide compound Chemical class 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0179—Display position adjusting means not related to the information to be displayed
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G3/00—Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes
- G09G3/02—Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes by tracing or scanning a light beam on a screen
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/0093—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0172—Head mounted characterised by optical features
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0118—Head-up displays characterised by optical features comprising devices for improving the contrast of the display / brillance control visibility
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0123—Head-up displays characterised by optical features comprising devices increasing the field of view
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/014—Head-up displays characterised by optical features comprising information/image processing systems
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0147—Head-up displays characterised by optical features comprising a device modifying the resolution of the displayed image
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0172—Head mounted characterised by optical features
- G02B2027/0174—Head mounted characterised by optical features holographic
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B2027/0178—Eyeglass type
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0179—Display position adjusting means not related to the information to be displayed
- G02B2027/0187—Display position adjusting means not related to the information to be displayed slaved to motion of at least a part of the body of the user, e.g. head, eye
Definitions
- the present systems, devices, and methods generally relate to wearable heads-up displays and particularly relate to projector-based wearable heads-up displays.
- a head-mounted display is an electronic device that is worn on a user's head and, when so worn, secures at least one electronic display within a viewable field of at least one of the user's eyes, regardless of the position or orientation of the user's head.
- a wearable heads-up display is a head-mounted display that enables the user to see displayed content but also does not prevent the user from being able to see their external environment.
- the “display” component of a wearable heads-up display is either transparent or at a periphery of the user's field of view so that it does not completely block the user from being able to see their external environment. Examples of wearable heads-up displays include: the Google Glass®, the Optinvent Ora®, the Epson Moverio®, and the Sony Glasstron®, just to name a few.
- a challenge in the design of wearable heads-up displays is to minimize the bulk of the face-worn apparatus will still providing displayed content with sufficient visual quality.
- a wearable heads-up display may be summarized as including: a modulative light source; a dynamic scanner; and a virtual content control system communicatively coupled to both the modulative light source and the dynamic scanner, the virtual content control system including a processor and a non-transitory processor-readable storage medium communicatively coupled to the processor, wherein the non-transitory processor-readable storage medium stores processor-executable resolution control instructions that, when executed by the processor, cause the wearable heads-up display to: identify a region of interest in the user's field of view; and project virtual content with high resolution in the region of interest and with relatively lower resolution outside of the region of interest.
- the wearable heads-up display may further comprise an eye-tracker communicatively coupled to the virtual content control system, wherein the processor-executable resolution control instructions, when executed by the processor, cause the wearable heads-up display to identify a region of interest in the user's field of view based on a position of the user's foveal region as determined by the eye-tracker.
- a method of operating a wearable heads-up display to display virtual content with non-uniform resolution may be summarized as including: identifying a region of interest in a field of view of a user of the wearable heads-up display; and projecting, by the wearable heads-up display, virtual content with high resolution in the region of interest in the field of view of the user and with relatively lower resolution in regions of the field of view of the user that are outside of the region of interest.
- Identifying a region of interest in a field of view of a user of the wearable heads-up display may include identifying a foveal region in the field of view of the user of the wearable heads-up display.
- the wearable heads-up display may include an eye-tracker and identifying a foveal region in the field of view of the user of the wearable heads-up display may include identifying the foveal region based on a position of an eye of the user as determined by the eye-tracker.
- projecting, by the wearable heads-up display, virtual content with high resolution in the region of interest in the field of view of the user and with relatively lower resolution in regions of the field of view of the user that are outside of the region of interest may include scanning, by the dynamic scanner, virtual content with a first scanning step size in the region of interest and with a second scanning step size in regions of the field of view of the user that are outside of the region of interest, wherein the scanning step size is smaller than the second scanning step size.
- a wearable heads-up display may be summarized as including: a support structure that in use is worn on a head of a user; a projector carried by the support structure; a processor communicatively coupled to the projector; and a non-transitory processor-readable storage medium communicatively coupled to the processor, wherein the non-transitory processor-readable storage medium stores processor-executable virtual content control instructions that, when executed by the processor, cause the wearable heads-up display to: determine a region of interest in a field of view of the user; project virtual content with a first quality level with respect to a first display parameter in the region of interest; and project virtual content with a second quality level with respect to the first display parameter outside of the region of interest, wherein the first quality level is higher than the second quality level.
- the first quality level corresponds to a “high quality” with respect to the first display parameter
- the second quality level corresponds to a “relatively lower quality” with respect to the first display parameter.
- the region of interest in the field of view of the user may include a foveal region of the field of view of the user.
- the wearable heads-up display may further include a fovea tracker carried by the support structure, positioned and oriented to determine a position of a fovea of an eye of the user, wherein the fovea tracker is communicatively coupled to the processor, and wherein the processor-executable virtual content control instructions that, when executed by the processor, cause the wearable heads-up display to determine a region of interest in a field of view of the user, cause the wearable heads-up display to determine the foveal region of the field of view of the user based on the position of the fovea of the eye of the user determined by the fovea tracker.
- the wearable heads-up display may include an eye tracker carried by the support structure, positioned and oriented to determine a gaze direction of an eye of the user, wherein the eye tracker is communicatively coupled to the processor, and wherein the processor-executable virtual content control instructions that, when executed by the processor, cause the wearable heads-up display to determine a region of interest in a field of view of the user, cause the wearable heads-up display to determine a region of interest in the field of view of the user based on the gaze direction of the eye of the user determined by the eye tracker.
- the region of interest in the field of view of the user may include a foveal region of the field of view of the user, and the foveal region of the field of view of the user may be determined by the wearable heads-up display based on the gaze direction of the eye of the user determined by the eye tracker.
- the first display parameter may be selected from a group consisting of: a resolution of virtual content projected by the projector and a brightness of virtual content projected by the projector.
- the projector may include at least one projector selected from a group consisting of: a scanning laser projector and a digital light processing-based projector.
- the wearable heads-up display may further include a holographic combiner carried by the support structure, wherein the holographic combiner is positioned within a field of view of an eye of the user when the support structure is worn on the head of the user.
- the wearable heads-up display may further include a prescription eyeglass lens, wherein the holographic combiner is carried by the prescription eyeglass lens.
- the support structure may have a general shape and appearance of an eyeglasses frame.
- the wearable heads-up display may further include a virtual content control system, wherein both the processor and the non-transitory processor-readable storage medium are included in the virtual content control system.
- a method of operating a wearable heads-up display to display virtual content with non-uniform quality may be summarized as including: determining a region of interest in a field of view of a user of the wearable heads-up display; projecting, by the projector, virtual content with a first quality level with respect to a first display parameter in the region of interest of the field of view of the user; and projecting, by the projector, virtual content with a second quality level with respect to the first display parameter in regions of the field of view of the user that are outside of the region of interest, wherein the first quality level is higher than the second quality level.
- the first quality level corresponds to a “high quality” with respect to the first display parameter
- the second quality level corresponds to a “relatively lower quality” with respect to the first display parameter.
- Determining a region of interest in a field of view of a user of the wearable heads-up display may include determining a foveal region in the field of view of the user.
- the wearable heads-up display may include a fovea tracker and the method may further include determining a position of a fovea of an eye of the user by the fovea tracker.
- Determining a foveal region in the field of view of the user may include determining the foveal region of the field of view of the user based on the position of the fovea of the eye of the user determined by the fovea tracker.
- Projecting, by the projector, virtual content with a first quality level with respect to a first display parameter in the region of interest of the field of view of the user may include projecting, by the projector, virtual content with a first brightness level in the region of interest of the field of view of the user.
- Projecting, by the projector, virtual content with a second quality level with respect to the first display parameter in regions of the field of view of the user that are outside of the region of interest may include projecting, by the projector, virtual content with a second brightness level in regions of the field of view of the user that are outside of the region of interest, wherein the first brightness level is brighter than the second brightness level.
- Projecting, by the projector, virtual content with a first quality level with respect to a first display parameter in the region of interest of the field of view of the user may include projecting, by the projector, virtual content with a first resolution in the region of interest of the field of view of the user.
- Projecting, by the projector, virtual content with a second quality level with respect to the first display parameter in regions of the field of view of the user that are outside of the region of interest may include projecting, by the projector, virtual content with a second resolution in regions of the field of view of the user that are outside of the region of interest, wherein the first resolution is a higher resolution than the second resolution.
- the wearable heads-up display may include a processor and a non-transitory processor-readable storage medium communicatively coupled to the processor and which stores processor-executable virtual content control instructions.
- FIG. 1 is a partial-cutaway perspective view of a wearable heads-up display that provides heterogeneous display quality with respect to at least one display parameter in accordance with the present systems, devices, and methods.
- FIG. 2 is an illustrative diagram showing a plan view of exemplary projected virtual content from a wearable heads-up display that employs heterogeneous (non-uniform) display quality in accordance with the present systems, devices, and methods.
- FIG. 3 is a flow-diagram showing a method of operating a wearable heads-up display to display virtual content with heterogeneous (non-uniform) quality in accordance with the present systems, devices, and methods.
- WHUDs wearable heads-up displays
- Such heterogeneous or non-uniform display quality can advantageously reduce the graphical processing power and/or overall power consumption of a WHUD without compromising the perceived quality of the displayed content (i.e., “virtual content”).
- reducing the graphical processing power and/or overall power consumption of the WHUD in accordance with the present systems, devices, and methods enables the WHUD to employ smaller components (e.g., a smaller processor, a smaller memory, a smaller battery or batteries, and/or a smaller cooling system), which in turn enables the WHUD to adopt a smaller form factor and an overall more pleasing aesthetic design.
- smaller components e.g., a smaller processor, a smaller memory, a smaller battery or batteries, and/or a smaller cooling system
- the perceived quality of virtual content displayed by a WHUD may depend on a number of display parameters, including without limitation: resolution, number of pixels, pixel density, pixel size, brightness, color saturation, sharpness, focus, noise, and so on. All other things being equal, a WHUD that displays virtual content with high quality may generally demand higher graphical processing power and/or generally consume more overall power than a WHUD that displays virtual content with relatively lower image quality. As described above, higher graphical processing power and/or higher overall power consumption can add significant and unwanted bulk to a WHUD by necessitating, for example, larger battery(ies), a larger processor, a larger memory coupled to the processor, a larger display engine, and/or a larger cooling system for the processor and/or for the display engine.
- the present systems, devices, and methods describe WHUDs that strategically display virtual content with heterogeneous or non-uniform display quality (with respect to at least one display parameter) in order to provide virtual content that still appears in high quality to the user without necessitating all, as many, or any larger or more powerful components. In this way, the added bulk of the WHUD is limited and a more aesthetically-pleasing design is realized.
- a “projector-based” WHUD is generally used as an example of a WHUD architecture; however, a person of skill in the art will appreciate that the various teachings described herein may be applied in other, non-projector-based WHUD architectures (e.g., WHUD architectures that employ one or more microdisplay(s) and/or waveguide structures).
- a projector-based WHUD may be a form of virtual retina display in which a projector draws a raster scan onto the eye of the user.
- the projector may include a scanning laser projector, a digital light processing-based projector, or generally any combination of a modulative light source (such as a laser or one or more LED(s)) and a dynamic reflector mechanism (such as one or more dynamic scanner(s) or digital light processor(s)). In the absence of any further measure the projector may project light over a fixed area called the field of view (“FOV”) of the display.
- a modulative light source such as a laser or one or more LED(s)
- a dynamic reflector mechanism such as one or more dynamic scanner(s) or digital light processor(s)
- FOV field of view
- FOV generally refers to the extent that a scene is visible to an observer and is usually characterized by the angle formed at the eye between respective light beams originating from two points at opposite edges of a scene that are both visible from the same eye position.
- the human eye typically has a FOV of almost 180° across the horizontal direction and about 135° across the vertical direction.
- a WHUD typically has a FOV that is less than the FOV of the eye, although it is desirable for a WHUD to be capable of providing virtual content with a FOV as close as possible to the FOV of the eye. Unfortunately, this is typically a great challenge given the close proximity of the WHUD to the eye.
- a larger FOV demands more graphical processing power because a larger FOV generally means there is more virtual content to display.
- a larger FOV has overall higher power consumption because, at least in part, of the higher levels of graphical processing and the overall increased light signal generation necessary to fill the larger FOV.
- the various embodiments described herein include techniques for projecting virtual content with a high FOV while easing the demands (e.g., graphical processing demands and/or power consumption) on the display architecture. This is achieved, at least in part, by projecting virtual content with heterogeneous or non-uniform display quality with respect to at least one display parameter.
- virtual content may be projected with heterogeneous or non-uniform resolution and/or with heterogeneous or non-uniform brightness.
- the virtual content may be projected with relatively high quality with respect to a first display parameter (e.g., resolution or brightness) at and over a particular region of interest/focus and with relatively lower quality with respect to the same first display parameter elsewhere.
- a first display parameter e.g., resolution or brightness
- a large FOV may be displayed to the user while mitigating demands on graphical processing and power.
- This scheme advantageously accounts for the fact that a user's ability to focus is typically not uniform over the eye's entire FOV. In practice, when a user is focusing on a high quality region of interest in a complete FOV, the user may not be able to detect that regions of the FOV that are outside of this region of interest are being projected at lower quality.
- high quality and low quality are often used with respect to one or more display parameter(s). Unless the specific context requires otherwise, such terms are generally used in a relative sense with respect to the same display parameter. “High quality” (and its variants) generally refers to a first quality level with respect to a first display parameter, the first quality level generally equal to the perceived quality of the WHUD with respect to the first display parameter. “Low quality” (and its variants) generally refers to a second quality level with respect to the same first display parameter, the second quality level generally lower than and/or less than the first quality level.
- high quality and low quality are used to denote that the first quality level is higher than the second quality level with respect to the same display parameter.
- the exact amount by which a “high quality” is higher than a “low quality” may depend on a variety of factors, including the specific display parameter and/or other display parameters in the WHUD.
- a first quality level or “high quality” with respect to a display parameter may be 1% higher than a second quality level or “low quality” with respect to the same display parameter
- a first quality level or “high quality” with respect to a display parameter may be 10% higher than a second quality level or “low quality” with respect to the same display parameter
- a first quality level or “high quality” with respect to a display parameter may be 25% higher than a second quality level or “low quality” with respect to the same display parameter
- a first quality level or “high quality” with respect to a display parameter may be 50% higher than a second quality level or “low quality” with respect to the same display parameter
- a first quality level or “high quality” with respect to a display parameter may be 100% higher than a second quality level or “low quality” with respect to the same display parameter
- a first quality level or “high quality” with respect to a display parameter may be more than 100% higher than a second quality level or “low quality” (e.g., 150%, 200%, and so
- a first quality level being higher than a second quality level may correspond to the actual value of the first quality level being greater than or less than the second quality level depending on the specific display parameter. For example, if the display parameter in question is pixel density, then in order for the first quality level to be higher than the second quality level the pixel density associated with the first quality level may be greater than the pixel density associated with the second quality level; however, if the display parameter in question is the spacing in between pixels, then in order for the first quality level to be higher than the second quality level the spacing in between pixels associated with the first quality level may be less than the spacing in between pixels associated with the second quality level.
- FIG. 1 is a partial-cutaway perspective view of a WHUD 100 that provides heterogeneous display quality with respect to at least one display parameter in accordance with the present systems, devices, and methods.
- WHUD 100 includes a support structure 110 that in use is worn on the head of a user and has a general shape and appearance of an eyeglasses (e.g., sunglasses) frame.
- Support structure 110 carries multiple components, including: a projector 120 (a scanning laser projector in the illustrated example), a holographic combiner 130 , and an exit pupil expansion optic 150 . Portions of projector 120 and exit pupil expansion optic 150 may be contained within an inner volume of support structure 110 ; however, FIG.
- support structure 110 also carries a virtual content control system 160 communicatively coupled to projector 120 .
- Virtual content control system 160 comprises a processor 161 and a non-transitory processor-readable storage medium or memory 162 communicatively coupled to the processor 162 .
- Memory 162 stores processor-executable virtual content control data and/or instructions 163 that, when executed by processor 161 , cause WHUD 100 to provide heterogeneous display quality with respect to at least one display parameter as discussed in more detail later on.
- the term “carries” and variants such as “carried by” are generally used to refer to a physical coupling between two objects.
- the physical coupling may be direct physical coupling (i.e., with direct physical contact between the two objects) or indirect physical coupling that may be mediated by one or more additional objects.
- the term carries and variants such as “carried by” are meant to generally encompass all manner of direct and indirect physical coupling, including without limitation: carried on, carried within, physically coupled to, and/or supported by, with or without any number of intermediary physical objects therebetween.
- Projector 120 is a scanning laser projector, though as previously described other forms of projectors may similarly be used, such as a digital light processing-based projector.
- Projector 120 includes multiple laser diodes (e.g., a red laser diode, a green laser diode, and/or a blue laser diode) and at least one scan mirror (e.g., a single two-dimensional scan mirror or two one-dimensional scan mirrors, which may be, e.g., MEMS-based or piezo-based).
- a person of skill in the art will appreciate that the teachings herein may be applied in WHUDs that employ non-projector-based display architectures, such as WHUDs that employ microdisplays and/or waveguide structures.
- Holographic combiner 130 is positioned within a field of view of at least one eye of the user when support structure 110 is worn on the head of the user. Holographic combiner 130 is sufficiently optically transparent to permit light from the user's environment (i.e., “environmental light”) to pass through to the user's eye.
- support structure 110 further carries a transparent eyeglass lens 140 (e.g., a prescription eyeglass lens) and holographic combiner 130 comprises at least one layer of holographic material that is adhered to, affixed to, laminated with, carried in or upon, or otherwise integrated with eyeglass lens 140 .
- the at least one layer of holographic material may include a photopolymer film such as Bayfol®HX available from Bayer MaterialScience AG or a silver halide compound and may, for example, be integrated with transparent lens 140 using any of the techniques described in U.S. Provisional Patent Application Ser. No. 62/214,600.
- Holographic combiner 130 includes at least one hologram in or on the at least one layer of holographic material. With holographic combiner 130 positioned in a field of view of an eye of the user when support structure 110 is worn on the head of the user, the at least one hologram of holographic combiner 130 is positioned and oriented to redirect light originating from projector 120 towards the eye of the user. In particular, the at least one hologram is positioned and oriented to receive light signals that originate from projector 120 and converge those light signals to at least one exit pupil at or proximate the eye of the user.
- Exit pupil expansion optic 150 is positioned in an optical path between projector 120 and holographic combiner 130 and may take on any of a variety of different forms, including without limitation those described in U.S. patent application Ser. No. 15/046,234, U.S. patent application Ser. No. 15/046,254, and/or U.S. patent application Ser. No. 15/046,269.
- the processor-executable virtual content control instructions (and/or data) 163 when executed by processor 161 of virtual content control system, cause WHUD 100 to provide heterogeneous display quality with respect to at least one display parameter.
- processor-executable virtual content control instruction (and/or data) 163 when executed by processor 161 , cause WHUD 100 to determine a region of interest in a FOV of the user, project virtual content with a high quality with respect to a first display parameter in the region of interest, and project virtual content with a relatively lower quality with respect to the first display parameter outside of the region of interest.
- the first display parameter may include any of a variety of different display parameters depending on the specific implementation, including without limitation: resolution, number of pixels, pixel density, pixel size, brightness, color saturation, sharpness, focus, and/or noise.
- the WHUD ( 100 ) may provide heterogeneous (non-uniform) display quality with respect to multiple different display parameters, such as at least two different display parameters.
- resolution is used, with reference to display quality and/or virtual content projected by a projector ( 1200 , to generally refer to a distribution of pixels or lines that make up a display and/or that make up virtual content of a display.
- the “quality” of resolution may depend on a number of resolution parameters and, accordingly, the quality of resolution may be adjusted (i.e., made higher or lower) by tuning any one or combination of multiple ones of the resolution parameters.
- Exemplary resolution parameters that may be tuned in order to make the display quality of virtual content higher or lower with respect to display resolution include, without limitation: number of pixels, size of pixels, spacing in between pixels, and/or pixel density.
- the display quality of WHUD 100 may be made higher with respect to resolution by increasing the number of pixels, decreasing the size of pixels, and/or increasing the pixel density. Conversely, the display quality of WHUD 100 may be made lower with respect to resolution by decreasing the number of pixels, increasing the size of pixels, and/or decreasing the pixel density.
- Resolution is just one example of a display parameter that may be varied of the FOV of a WHUD to provide heterogeneous (non-uniform) display quality in accordance with the present systems, devices, and methods.
- Brightness is another example of such a display parameter.
- the display quality of WHUD 100 may be made higher with respect to brightness by increasing the brightness and the display quality of WHUD 100 may be made lower with respect to brightness by decreasing the brightness.
- providing heterogeneous (non-uniform) display quality with respect to a first display parameter may include heterogeneously varying at least a second display parameter over the FOV of a WHUD in order to compensate for one or more effect(s) of providing heterogeneous display quality with respect to the first display parameter.
- regions of a WHUD's FOV that are displayed with relatively low quality resolution may be displayed with relatively higher brightness to compensate and reduce the likelihood that the user will perceive the non-uniformity in resolution.
- the region of interest in the FOV of the user in which virtual content is displayed (e.g., projected) with high quality with respect to a first display parameter may be determined (e.g., identified, deduced, or defined) in a variety of different ways.
- the region of interest may be an attribute of the virtual content itself and correspond to a region of the virtual content where the user is expected to attract their attention based on the nature of the virtual content. For example, if the virtual content comprises a block of text overlaid on a textured background, virtual content control system 160 may determine (e.g., define or deduce) that the region of interest corresponds to the block of text as this is likely where the user will direct their attention.
- virtual content control system 160 may define a region of interest in order to strategically direct the user's attention (e.g., guide the user's gaze) to that region of the virtual content, for example, to highlight a new alert or notification and draw the user's attention thereto or to highlight a particular position on a map.
- the user's attention e.g., guide the user's gaze
- the region of interest may be identified by the WHUD ( 100 ) based on one or more property(ies) of the user's eye.
- WHUD 100 includes a sensor 170 carried by support structure 110 , where sensor 170 is operative to sense, measure, detect, monitor, and/or track one or more property(ies) of the user's eye.
- Sensor 170 is communicatively coupled to virtual content control system 160 and data from sensor 170 that is indicative or representative of one or more property(ies) of the user's eye may be used by virtual content control system 160 to determine a region of interest in a FOV of the user.
- Two exemplary eye properties that may be sensed, measured, detected, monitored, and/or tracked by sensor 170 and used by virtual content control system 160 to determine a region of interest in the user's FOV are now described.
- the region of interest in the FOV of the user may include a foveal region of the FOV of the user.
- the foveal region in the FOV of the user may generally correspond to light rays that impinge on the fovea (i.e., the “fovea centralis”) on the retina of the user's eye.
- the fovea is a depression in the inner surface of the retina (usually about 1.5 mm wide) that includes a relatively higher density of cone cells compared to the rest of the retinal surface. Due to this high density of cone cells, the fovea is generally the region of the retina that provides the sharpest (e.g., most detailed) vision and/or the highest visual acuity.
- humans have generally evolved to direct their gaze (e.g., adjust their eye position) so that light coming from the detailed object impinges on the fovea.
- WHUD 100 projects virtual content with high quality (with respect to a first display parameter) in the foveal region of the user's FOV by aligning the virtual content with the user's eye so that the high quality region of the virtual content aligns with (e.g., impinges on) the fovea of the retina of the user's eye.
- sensor 170 may include a fovea tracker that is communicatively coupled to virtual content control system 160 (e.g., communicatively coupled to processor 161 of virtual content control system 160 ).
- Fovea tracker 170 is positioned and oriented to determine a position of the fovea of the user's eye and processor-executable virtual content control instructions 163 may, when executed by processor 161 , cause WHUD 100 to determine (e.g., identify) the foveal region of the user's FOV based on the position of the fovea of the user's eye determined by fovea tracker 170 .
- fovea tracker 170 may employ a variety of different techniques.
- fovea tracker 170 may comprise an illumination source (e.g., a light source, such as an infrared light source) and/or an optical sensor such as a camera, a video camera, or a photodetector.
- an illumination source e.g., a light source, such as an infrared light source
- an optical sensor such as a camera, a video camera, or a photodetector.
- the optical sensor component of fovea tracker 170 may sense, detect, measure, monitor, and/or track retinal blood vessels and/or other features on the inside of the user's eye from which the position of the fovea may be determined (e.g., identified).
- the optical sensor component of fovea tracker 170 may capture images of the user's eye and a processor communicatively coupled to the optical sensor (e.g., processor 161 ) may process the images to determine (e.g., identify) the position of the fovea based on, for example, discernible features of the retina (e.g., retinal blood vessels) in the images. Processing the mages by the processor may include executing, by the processor, processor-readable image processing data and/or instructions stored in a non-transitory processor-readable storage medium or memory (e.g., memory 162 ) of WHUD 100 .
- a processor communicatively coupled to the optical sensor
- processing the mages by the processor may include executing, by the processor, processor-readable image processing data and/or instructions stored in a non-transitory processor-readable storage medium or memory (e.g., memory 162 ) of WHUD 100 .
- the region of interest in the FOV of the user may be determined by WHUD 100 based on the gaze direction of the user.
- sensor 170 may include an eye tracker carried by support structure 110 and positioned and oriented to determine a gaze direction of the eye of the user.
- Eye tracker 170 may be communicatively coupled to virtual content control system 160 (e.g., communicatively coupled to processor 161 of virtual content control system 160 ) and processor-executable virtual content control instructions 163 may, when executed by processor 161 , cause WHUD 100 to determine (e.g., identify) a region of interest in the FOV of the user based on the gaze direction of the user's eye determined by eye tracker 170 .
- eye tracker 170 itself may determine the gaze direction of the user's eye and relay this information to processor 161 , or processor 161 may determine the gaze direction of the user's eye based on data and/or information provided by eye tracker 170 .
- Eye tracker 170 may employ any of a variety of different eye tracking technologies depending on the specific implementation.
- eye tracker 170 may employ any or all of the systems, devices, and methods described in U.S. Provisional Patent Application Ser. No. 62/167,767; U.S. Provisional Patent Application Ser. No. 62/271,135; U.S. Provisional Patent Application Ser. No. 62/245,792; and/or U.S. Provisional Patent Application Ser. No. 62/281,041.
- virtual content control system 160 may position the region of interest to align with the gaze direction of the user's eye so that the region of interest appears substantially centrally in the user's FOV and remains in this position for all eye positions over a wide range of eye positions. This approach may cause the region of interest to at least partially align with the foveal region in the user's FOV without direct determination of the position of the user's fovea.
- the position of the fovea in the user's eye may be determined (e.g., deduced) by virtual content control system 160 based on the gaze direction of the user's eye because the position of the fovea in the user's eye is generally fixed relative to the positions of the pupil, iris, cornea, and/or other features of the user's eye that may be sensed, measured, detected, monitored, and/or tracked by eye tracker 170 in determining the gaze direction of the user's eye.
- virtual content is dynamically projected with highest quality (with respect to at least one display parameter) in the region of the user's FOV that corresponds to the user's fovea (e.g., the foveal region of the user's FOV) and with relatively lower quality (with respect to the same at least one display parameter) elsewhere in the user's FOV (i.e., in regions of the user's FOV outside of the foveal region).
- the virtual content is “dynamic” in the sense that the high quality region “follows” the user's fovea (i.e., follows the foveal region in the user's FOV) based on the user's fovea position, eye position, and/or gaze direction as determined by sensor 170 . Since the user's ability to focus over the entire FOV is non-uniform, it is unnecessary to project (and to provide sufficient infrastructure, e.g., graphical processing power and overall system power to render the system capable of projecting) virtual content with high quality over the entire FOV. Rather, in accordance with the present systems, devices, and methods, only the foveal region (or another region of interest) of the virtual content may be projected at high quality while the peripheral region(s) of the virtual content may be projected at comparatively lower quality.
- FIG. 2 is an illustrative diagram showing a plan view of exemplary projected virtual content from a WHUD 200 that employs heterogeneous (non-uniform) display quality in accordance with the present systems, devices, and methods.
- virtual content corresponding to the “foveal region” of the user's FOV as determined by an on-board fovea-tracking system and/or an on-board eye-tracking system, not illustrated in FIG.
- FIG. 3 is a flow-diagram showing a method 300 of operating a WHUD to display virtual content with heterogeneous (non-uniform) quality in accordance with the present systems, devices, and methods.
- the WHUD includes a projector and may be substantially similar to WHUD 100 from FIG. 1 .
- Method 300 includes three acts 301 , 302 , and 303 , though those of skill in the art will appreciate that in alternative embodiments certain acts may be omitted and/or additional acts may be added. Those of skill in the art will also appreciate that the illustrated order of the acts is shown for exemplary purposes only and may change in alternative embodiments.
- the term “user” refers to a person that is wearing the WHUD.
- a region of interest in the user's FOV is determined.
- This region of interest may be determined by the WHUD itself, for example, by the processor of a virtual content control system carried by the WHUD.
- this region of interest may be determined (e.g., defined) by (i.e., within) a software application executed by a processor on-board the WHUD based on an intention to motivate the user to focus on this particular region.
- this region of interest may be determined (e.g., identified or deduced) by the WHUD based on data and/or information provided by one or more sensor(s) (such as a fovea tracker and/or and eye tracker) based on the position of the fovea of the user's eye, the position of the user's eye, and/or the gaze direction of the user's eye.
- the region of interest may or may not include a foveal region of the user's FOV.
- the projector of the WHUD projects virtual content with a high quality with respect to a first display parameter in the region of interest in the FOV of the user.
- the projector of the WHUD projects virtual content with a relatively lower quality with respect to the first display parameter in regions of the field of view of the user that are outside of the region of interest.
- the first display parameter may include any of a variety of different display parameters depending on the specific implementation, including without limitation: resolution, number of pixels, pixel density, pixel size, brightness, color saturation, sharpness, focus, and/or noise.
- the projector of the WHUD may project virtual content with a high brightness in the region of interest of the FOV of the user and at 303 the projector may project virtual content with a relatively lower brightness in regions of the FOV of the user that are outside of the region of interest.
- the projector of the WHUD may project virtual content with a high resolution in the region of interest of the FOV of the user and at 303 the projector may project virtual content with a relatively lower resolution in regions of the FOV of the user that are outside of the region of interest.
- the quality of the resolution of the virtual content may be varied in a number of different ways, especially when a projector is used in the display system of the WHUD.
- the quality of the resolution may be varied by adjusting the number pixels in the virtual content, the size of the pixels in the virtual content, the size of the gaps between pixels in the virtual content, and/or the density of pixels in the virtual content.
- the quality of the resolution of virtual content may be varied by adjusting either or both of the light modulation of the modulative light source (e.g., laser diodes, LEDs, or similar) and/or the operation of the one or more dynamically-variable reflector(s) (e.g., scan mirror(s)).
- the projector may project virtual content with a high resolution in the region of interest of the FOV of the user by projecting virtual content with a first light modulation frequency in the region of interest in the FOV of the user and at 303 the projector may project virtual content with a relatively lower resolution in regions of the FOV of the user that are outside of the region of interest by projecting virtual content with a second light modulation frequency in regions of the FOV of the user that are outside of the region of interest.
- the first light modulation frequency is greater than the second light modulation frequency.
- the projector e.g., a scanning laser projector
- the projector may project virtual content with a high resolution in the region of interest of the FOV of the user by scanning virtual content with a first scanning step size in the region of interest in the FOV of the user and at 303 the projector may project virtual content with a relatively lower resolution in regions of the FOV of the user that are outside of the region of interest by scanning virtual content with a second scanning step size in regions of the FOV of the user that are outside of the region of interest.
- the first scanning step size is smaller than the second scanning step size.
- using a scanning laser projector heterogeneous (non-uniform) display resolution may be achieved by operating either or both of the modulative light source and/or the dynamic scanner to project relatively fewer/larger pixels outside of the user's foveal region and relatively more/smaller pixels within the user's foveal region.
- the result may be more concentrated image detail (i.e., a higher number/concentration of distinct pixels) in the user's foveal region (as dynamically determined by the eye-tracker(s)) and reduced image detail (i.e., a lower number/concentration of distinct pixels) outside of the user's foveal region.
- the discrepancy is pixel concentration may significantly save on graphical processing and power consumption while nevertheless remaining substantially undetectable to the user, because the user typically cannot focus to high degrees of resolution on those regions of their field of view that are outside of the foveal region.
- the WHUDs described herein may include one or more on-board power sources (e.g., one or more battery(ies)), a wireless transceiver for sending/receiving wireless communications, and/or a tethered connector port for coupling to a computer and/or charging the one or more on-board power source(s).
- on-board power sources e.g., one or more battery(ies)
- wireless transceiver for sending/receiving wireless communications
- a tethered connector port for coupling to a computer and/or charging the one or more on-board power source(s).
- the WHUDs described herein may receive and respond to commands from the user in one or more of a variety of ways, including without limitation: voice commands through a microphone; touch commands through buttons, switches, or a touch sensitive surface; and/or gesture-based commands through gesture detection systems as described in, for example, U.S. Non-Provisional patent application Ser. No. 14/155,087, U.S. Non-Provisional patent application Ser. No. 14/155,107, PCT Patent Application PCT/US2014/057029, and/or U.S. Provisional Patent Application Ser. No. 62/236,060, all of which are incorporated by reference herein in their entirety.
- WHUDs described herein may include any or all of the technologies described in U.S. Provisional Patent Application Ser. No. 62/156,736, and/or U.S. Provisional Patent Application Ser. No. 62/242,844.
- communicative as in “communicative pathway,” “communicative coupling,” and in variants such as “communicatively coupled,” is generally used to refer to any engineered arrangement for transferring and/or exchanging information.
- exemplary communicative pathways include, but are not limited to, electrically conductive pathways (e.g., electrically conductive wires, electrically conductive traces), magnetic pathways (e.g., magnetic media), and/or optical pathways (e.g., optical fiber), and exemplary communicative couplings include, but are not limited to, electrical couplings, magnetic couplings, and/or optical couplings.
- infinitive verb forms are often used. Examples include, without limitation: “to detect,” “to provide,” “to transmit,” “to communicate,” “to process,” “to route,” and the like. Unless the specific context requires otherwise, such infinitive verb forms are used in an open, inclusive sense, that is as “to, at least, detect,” “to, at least, provide,” “to, at least, transmit,” and so on.
- logic or information can be stored on any processor-readable medium for use by or in connection with any processor-related system or method.
- a memory is a processor-readable medium that is an electronic, magnetic, optical, or other physical device or means that contains or stores a computer and/or processor program.
- Logic and/or the information can be embodied in any processor-readable medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions associated with logic and/or information.
- a “non-transitory processor-readable medium” can be any element that can store the program associated with logic and/or information for use by or in connection with the instruction execution system, apparatus, and/or device.
- the processor-readable medium can be, for example, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus or device.
- the computer readable medium would include the following: a portable computer diskette (magnetic, compact flash card, secure digital, or the like), a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM, EEPROM, or Flash memory), a portable compact disc read-only memory (CDROM), digital tape, and other non-transitory media.
- a portable computer diskette magnetic, compact flash card, secure digital, or the like
- RAM random access memory
- ROM read-only memory
- EPROM erasable programmable read-only memory
- CDROM compact disc read-only memory
- digital tape digital tape
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Optics & Photonics (AREA)
- Engineering & Computer Science (AREA)
- Computer Hardware Design (AREA)
- Theoretical Computer Science (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
- 1. Technical Field
- The present systems, devices, and methods generally relate to wearable heads-up displays and particularly relate to projector-based wearable heads-up displays.
- 2. Description of the Related Art
- A head-mounted display is an electronic device that is worn on a user's head and, when so worn, secures at least one electronic display within a viewable field of at least one of the user's eyes, regardless of the position or orientation of the user's head. A wearable heads-up display is a head-mounted display that enables the user to see displayed content but also does not prevent the user from being able to see their external environment. The “display” component of a wearable heads-up display is either transparent or at a periphery of the user's field of view so that it does not completely block the user from being able to see their external environment. Examples of wearable heads-up displays include: the Google Glass®, the Optinvent Ora®, the Epson Moverio®, and the Sony Glasstron®, just to name a few.
- The optical performance of a wearable heads-up display is an important factor in its design. When it comes to face-worn devices, however, users also care a lot about aesthetics. This is clearly highlighted by the immensity of the eyeglass (including sunglass) frame industry. Independent of their performance limitations, many of the aforementioned examples of wearable heads-up displays have struggled to find traction in consumer markets because, at least in part, they lack fashion appeal. Most wearable heads-up displays presented to date employ large display components and, as a result, most wearable heads-up displays presented to date are considerably bulkier and less stylish than conventional eyeglass frames.
- A challenge in the design of wearable heads-up displays is to minimize the bulk of the face-worn apparatus will still providing displayed content with sufficient visual quality. There is a need in the art for wearable heads-up displays of more aesthetically-appealing design that are capable of providing high-quality images to the user without limiting the user's ability to see their external environment.
- A wearable heads-up display may be summarized as including: a modulative light source; a dynamic scanner; and a virtual content control system communicatively coupled to both the modulative light source and the dynamic scanner, the virtual content control system including a processor and a non-transitory processor-readable storage medium communicatively coupled to the processor, wherein the non-transitory processor-readable storage medium stores processor-executable resolution control instructions that, when executed by the processor, cause the wearable heads-up display to: identify a region of interest in the user's field of view; and project virtual content with high resolution in the region of interest and with relatively lower resolution outside of the region of interest. The wearable heads-up display may further comprise an eye-tracker communicatively coupled to the virtual content control system, wherein the processor-executable resolution control instructions, when executed by the processor, cause the wearable heads-up display to identify a region of interest in the user's field of view based on a position of the user's foveal region as determined by the eye-tracker.
- A method of operating a wearable heads-up display to display virtual content with non-uniform resolution may be summarized as including: identifying a region of interest in a field of view of a user of the wearable heads-up display; and projecting, by the wearable heads-up display, virtual content with high resolution in the region of interest in the field of view of the user and with relatively lower resolution in regions of the field of view of the user that are outside of the region of interest. Identifying a region of interest in a field of view of a user of the wearable heads-up display may include identifying a foveal region in the field of view of the user of the wearable heads-up display. The wearable heads-up display may include an eye-tracker and identifying a foveal region in the field of view of the user of the wearable heads-up display may include identifying the foveal region based on a position of an eye of the user as determined by the eye-tracker.
- The wearable heads-up display may comprise a modulative light source and a dynamic scanner. Projecting, by the wearable heads-up display, virtual content with high resolution in the region of interest in the field of view of the user and with relatively lower resolution in regions of the field of view of the user that are outside of the region of interest may include projecting, by the modulative light source, virtual content with a first light modulation frequency in the region of interest and with a second light modulation frequency in regions of the field of view of the user that are outside of the region of interest, wherein the first light modulation frequency is greater than the second light modulation frequency. Either in addition to or instead of such adjustments to the light modulation frequency, projecting, by the wearable heads-up display, virtual content with high resolution in the region of interest in the field of view of the user and with relatively lower resolution in regions of the field of view of the user that are outside of the region of interest may include scanning, by the dynamic scanner, virtual content with a first scanning step size in the region of interest and with a second scanning step size in regions of the field of view of the user that are outside of the region of interest, wherein the scanning step size is smaller than the second scanning step size.
- A wearable heads-up display may be summarized as including: a support structure that in use is worn on a head of a user; a projector carried by the support structure; a processor communicatively coupled to the projector; and a non-transitory processor-readable storage medium communicatively coupled to the processor, wherein the non-transitory processor-readable storage medium stores processor-executable virtual content control instructions that, when executed by the processor, cause the wearable heads-up display to: determine a region of interest in a field of view of the user; project virtual content with a first quality level with respect to a first display parameter in the region of interest; and project virtual content with a second quality level with respect to the first display parameter outside of the region of interest, wherein the first quality level is higher than the second quality level. In other words, the first quality level corresponds to a “high quality” with respect to the first display parameter and the second quality level corresponds to a “relatively lower quality” with respect to the first display parameter.
- The region of interest in the field of view of the user may include a foveal region of the field of view of the user. The wearable heads-up display may further include a fovea tracker carried by the support structure, positioned and oriented to determine a position of a fovea of an eye of the user, wherein the fovea tracker is communicatively coupled to the processor, and wherein the processor-executable virtual content control instructions that, when executed by the processor, cause the wearable heads-up display to determine a region of interest in a field of view of the user, cause the wearable heads-up display to determine the foveal region of the field of view of the user based on the position of the fovea of the eye of the user determined by the fovea tracker.
- The wearable heads-up display may include an eye tracker carried by the support structure, positioned and oriented to determine a gaze direction of an eye of the user, wherein the eye tracker is communicatively coupled to the processor, and wherein the processor-executable virtual content control instructions that, when executed by the processor, cause the wearable heads-up display to determine a region of interest in a field of view of the user, cause the wearable heads-up display to determine a region of interest in the field of view of the user based on the gaze direction of the eye of the user determined by the eye tracker. The region of interest in the field of view of the user may include a foveal region of the field of view of the user, and the foveal region of the field of view of the user may be determined by the wearable heads-up display based on the gaze direction of the eye of the user determined by the eye tracker.
- The first display parameter may be selected from a group consisting of: a resolution of virtual content projected by the projector and a brightness of virtual content projected by the projector. The projector may include at least one projector selected from a group consisting of: a scanning laser projector and a digital light processing-based projector.
- The wearable heads-up display may further include a holographic combiner carried by the support structure, wherein the holographic combiner is positioned within a field of view of an eye of the user when the support structure is worn on the head of the user. The wearable heads-up display may further include a prescription eyeglass lens, wherein the holographic combiner is carried by the prescription eyeglass lens.
- The support structure may have a general shape and appearance of an eyeglasses frame.
- The wearable heads-up display may further include a virtual content control system, wherein both the processor and the non-transitory processor-readable storage medium are included in the virtual content control system.
- A method of operating a wearable heads-up display to display virtual content with non-uniform quality, the wearable heads-up display including a projector, may be summarized as including: determining a region of interest in a field of view of a user of the wearable heads-up display; projecting, by the projector, virtual content with a first quality level with respect to a first display parameter in the region of interest of the field of view of the user; and projecting, by the projector, virtual content with a second quality level with respect to the first display parameter in regions of the field of view of the user that are outside of the region of interest, wherein the first quality level is higher than the second quality level. In other words, the first quality level corresponds to a “high quality” with respect to the first display parameter and the second quality level corresponds to a “relatively lower quality” with respect to the first display parameter.
- Determining a region of interest in a field of view of a user of the wearable heads-up display may include determining a foveal region in the field of view of the user. The wearable heads-up display may include a fovea tracker and the method may further include determining a position of a fovea of an eye of the user by the fovea tracker. Determining a foveal region in the field of view of the user may include determining the foveal region of the field of view of the user based on the position of the fovea of the eye of the user determined by the fovea tracker.
- The wearable heads-up display may include an eye tracker and the method may further include determining a gaze direction of an eye of the user by the eye tracker. Determining a region of interest in a field of view of a user of the wearable heads-up display may include determining the region of interest in the field of view of the user based on the gaze direction of the eye of the user determined by the eye tracker.
- Projecting, by the projector, virtual content with a first quality level with respect to a first display parameter in the region of interest of the field of view of the user may include projecting, by the projector, virtual content with a first brightness level in the region of interest of the field of view of the user. Projecting, by the projector, virtual content with a second quality level with respect to the first display parameter in regions of the field of view of the user that are outside of the region of interest may include projecting, by the projector, virtual content with a second brightness level in regions of the field of view of the user that are outside of the region of interest, wherein the first brightness level is brighter than the second brightness level.
- Projecting, by the projector, virtual content with a first quality level with respect to a first display parameter in the region of interest of the field of view of the user may include projecting, by the projector, virtual content with a first resolution in the region of interest of the field of view of the user. Projecting, by the projector, virtual content with a second quality level with respect to the first display parameter in regions of the field of view of the user that are outside of the region of interest may include projecting, by the projector, virtual content with a second resolution in regions of the field of view of the user that are outside of the region of interest, wherein the first resolution is a higher resolution than the second resolution. Projecting, by the projector, virtual content with a first resolution in the region of interest of the field of view of the user may include projecting, by the projector, virtual content with a first light modulation frequency in the region of interest of the field of view of the user; and projecting, by the projector, virtual content with a second resolution in regions of the field of view of the user that are outside of the region of interest may include projecting, by the projector, virtual content with a second light modulation frequency in regions of the field of view of the user that are outside of the region of interest, wherein the first light modulation frequency is greater than the second light modulation frequency. Either alternatively or in addition, projecting, by the projector, virtual content with a first resolution in the region of interest of the field of view of the user may include scanning, by the projector, virtual content with a first scanning step size in the region of interest of the field of view of the user; and projecting, by the projector, virtual content with a second resolution in regions of the field of view of the user that are outside of the region of interest may include projecting, by the projector, virtual content with a second scanning step size in regions of the field of view of the user that are outside of the region of interest, wherein the first scanning step size is smaller than the second scanning step size.
- The wearable heads-up display may include a processor and a non-transitory processor-readable storage medium communicatively coupled to the processor and which stores processor-executable virtual content control instructions. In this case: determining a region of interest in a field of view of a user of the wearable heads-up display may include executing the processor-executable virtual content control instructions by the processor to cause the wearable heads-up display to determine the region of interest in the field of view of the user; projecting, by the projector, virtual content with a first quality level with respect to a first display parameter in the region of interest of the field of view of the user may include executing the processor-executable virtual content control instructions by the processor to cause the projector to project virtual content with the first quality level with respect to the first display parameter in the region of interest of the field of view of the user; and projecting, by the projector, virtual content with a second quality level with respect to the first display parameter in regions of the field of view of the user that are outside of the region of interest may include executing the processor-executable virtual content control instructions by the processor to cause the projector to project virtual content with the second quality level with respect to the first display parameter in regions of the field of view of the user that are outside of the region of interest.
- In the drawings, identical reference numbers identify similar elements or acts. The sizes and relative positions of elements in the drawings are not necessarily drawn to scale. For example, the shapes of various elements and angles are not necessarily drawn to scale, and some of these elements are arbitrarily enlarged and positioned to improve drawing legibility. Further, the particular shapes of the elements as drawn are not necessarily intended to convey any information regarding the actual shape of the particular elements, and have been solely selected for ease of recognition in the drawings.
-
FIG. 1 is a partial-cutaway perspective view of a wearable heads-up display that provides heterogeneous display quality with respect to at least one display parameter in accordance with the present systems, devices, and methods. -
FIG. 2 is an illustrative diagram showing a plan view of exemplary projected virtual content from a wearable heads-up display that employs heterogeneous (non-uniform) display quality in accordance with the present systems, devices, and methods. -
FIG. 3 is a flow-diagram showing a method of operating a wearable heads-up display to display virtual content with heterogeneous (non-uniform) quality in accordance with the present systems, devices, and methods. - In the following description, certain specific details are set forth in order to provide a thorough understanding of various disclosed embodiments. However, one skilled in the relevant art will recognize that embodiments may be practiced without one or more of these specific details, or with other methods, components, materials, etc. In other instances, well-known structures associated with portable electronic devices and head-worn devices, have not been shown or described in detail to avoid unnecessarily obscuring descriptions of the embodiments.
- Unless the context requires otherwise, throughout the specification and claims which follow, the word “comprise” and variations thereof, such as, “comprises” and “comprising” are to be construed in an open, inclusive sense, that is as “including, but not limited to.”
- Reference throughout this specification to “one embodiment” or “an embodiment” means that a particular feature, structures, or characteristics may be combined in any suitable manner in one or more embodiments.
- As used in this specification and the appended claims, the singular forms “a,” “an,” and “the” include plural referents unless the content clearly dictates otherwise. It should also be noted that the term “or” is generally employed in its broadest sense, that is as meaning “and/or” unless the content clearly dictates otherwise.
- The headings and Abstract of the Disclosure provided herein are for convenience only and do not interpret the scope or meaning of the embodiments.
- The various embodiments described herein provide systems, devices, and methods for wearable heads-up displays (“WHUDs”) with heterogeneous or non-uniform display quality. Such heterogeneous or non-uniform display quality can advantageously reduce the graphical processing power and/or overall power consumption of a WHUD without compromising the perceived quality of the displayed content (i.e., “virtual content”). Relative to implementations having higher graphical processing power and/or higher overall power consumption, reducing the graphical processing power and/or overall power consumption of the WHUD in accordance with the present systems, devices, and methods enables the WHUD to employ smaller components (e.g., a smaller processor, a smaller memory, a smaller battery or batteries, and/or a smaller cooling system), which in turn enables the WHUD to adopt a smaller form factor and an overall more pleasing aesthetic design.
- The perceived quality of virtual content displayed by a WHUD may depend on a number of display parameters, including without limitation: resolution, number of pixels, pixel density, pixel size, brightness, color saturation, sharpness, focus, noise, and so on. All other things being equal, a WHUD that displays virtual content with high quality may generally demand higher graphical processing power and/or generally consume more overall power than a WHUD that displays virtual content with relatively lower image quality. As described above, higher graphical processing power and/or higher overall power consumption can add significant and unwanted bulk to a WHUD by necessitating, for example, larger battery(ies), a larger processor, a larger memory coupled to the processor, a larger display engine, and/or a larger cooling system for the processor and/or for the display engine. The present systems, devices, and methods describe WHUDs that strategically display virtual content with heterogeneous or non-uniform display quality (with respect to at least one display parameter) in order to provide virtual content that still appears in high quality to the user without necessitating all, as many, or any larger or more powerful components. In this way, the added bulk of the WHUD is limited and a more aesthetically-pleasing design is realized.
- Throughout this specification and the appended claims, a “projector-based” WHUD is generally used as an example of a WHUD architecture; however, a person of skill in the art will appreciate that the various teachings described herein may be applied in other, non-projector-based WHUD architectures (e.g., WHUD architectures that employ one or more microdisplay(s) and/or waveguide structures). Generally, a projector-based WHUD may be a form of virtual retina display in which a projector draws a raster scan onto the eye of the user. The projector may include a scanning laser projector, a digital light processing-based projector, or generally any combination of a modulative light source (such as a laser or one or more LED(s)) and a dynamic reflector mechanism (such as one or more dynamic scanner(s) or digital light processor(s)). In the absence of any further measure the projector may project light over a fixed area called the field of view (“FOV”) of the display.
- FOV generally refers to the extent that a scene is visible to an observer and is usually characterized by the angle formed at the eye between respective light beams originating from two points at opposite edges of a scene that are both visible from the same eye position. The human eye typically has a FOV of almost 180° across the horizontal direction and about 135° across the vertical direction. A WHUD typically has a FOV that is less than the FOV of the eye, although it is desirable for a WHUD to be capable of providing virtual content with a FOV as close as possible to the FOV of the eye. Unfortunately, this is typically a great challenge given the close proximity of the WHUD to the eye. Furthermore, providing images with the full 180°×135° FOV can be very demanding of the display architecture, at least in terms of graphical processing power and overall power consumption. In conventional WHUD implementations and with all other things being equal, a larger FOV demands more graphical processing power because a larger FOV generally means there is more virtual content to display. Likewise, in conventional WHUD implementations and with all other things being equal, a larger FOV has overall higher power consumption because, at least in part, of the higher levels of graphical processing and the overall increased light signal generation necessary to fill the larger FOV. Even if a WHUD architecture is capable of accommodating such graphical processing and power consumption, doing so can add significant and unwanted bulk in the form of, for example, larger battery(ies), a larger processor, a larger memory coupled to the processor, a larger projector and/or larger projector components, and/or a larger cooling system for the processor and/or for the projector. Accordingly, the various embodiments described herein include techniques for projecting virtual content with a high FOV while easing the demands (e.g., graphical processing demands and/or power consumption) on the display architecture. This is achieved, at least in part, by projecting virtual content with heterogeneous or non-uniform display quality with respect to at least one display parameter. For example, virtual content may be projected with heterogeneous or non-uniform resolution and/or with heterogeneous or non-uniform brightness. In particular, the virtual content may be projected with relatively high quality with respect to a first display parameter (e.g., resolution or brightness) at and over a particular region of interest/focus and with relatively lower quality with respect to the same first display parameter elsewhere. By concentrating the display quality in a specific region (or in specific regions) of the full displayed FOV, a large FOV may be displayed to the user while mitigating demands on graphical processing and power. This scheme advantageously accounts for the fact that a user's ability to focus is typically not uniform over the eye's entire FOV. In practice, when a user is focusing on a high quality region of interest in a complete FOV, the user may not be able to detect that regions of the FOV that are outside of this region of interest are being projected at lower quality.
- Throughout this specification, the terms “high quality” and “low quality”, as well as variants such as “higher quality” and “lower quality” are often used with respect to one or more display parameter(s). Unless the specific context requires otherwise, such terms are generally used in a relative sense with respect to the same display parameter. “High quality” (and its variants) generally refers to a first quality level with respect to a first display parameter, the first quality level generally equal to the perceived quality of the WHUD with respect to the first display parameter. “Low quality” (and its variants) generally refers to a second quality level with respect to the same first display parameter, the second quality level generally lower than and/or less than the first quality level. The terms “high quality” and “low quality” are used to denote that the first quality level is higher than the second quality level with respect to the same display parameter. The exact amount by which a “high quality” is higher than a “low quality” may depend on a variety of factors, including the specific display parameter and/or other display parameters in the WHUD. For example, a first quality level or “high quality” with respect to a display parameter may be 1% higher than a second quality level or “low quality” with respect to the same display parameter, a first quality level or “high quality” with respect to a display parameter may be 10% higher than a second quality level or “low quality” with respect to the same display parameter, a first quality level or “high quality” with respect to a display parameter may be 25% higher than a second quality level or “low quality” with respect to the same display parameter, a first quality level or “high quality” with respect to a display parameter may be 50% higher than a second quality level or “low quality” with respect to the same display parameter, a first quality level or “high quality” with respect to a display parameter may be 100% higher than a second quality level or “low quality” with respect to the same display parameter, or a first quality level or “high quality” with respect to a display parameter may be more than 100% higher than a second quality level or “low quality” (e.g., 150%, 200%, and so on) with respect to the same display parameter. A first quality level being higher than a second quality level may correspond to the actual value of the first quality level being greater than or less than the second quality level depending on the specific display parameter. For example, if the display parameter in question is pixel density, then in order for the first quality level to be higher than the second quality level the pixel density associated with the first quality level may be greater than the pixel density associated with the second quality level; however, if the display parameter in question is the spacing in between pixels, then in order for the first quality level to be higher than the second quality level the spacing in between pixels associated with the first quality level may be less than the spacing in between pixels associated with the second quality level.
-
FIG. 1 is a partial-cutaway perspective view of a WHUD 100 that provides heterogeneous display quality with respect to at least one display parameter in accordance with the present systems, devices, and methods.WHUD 100 includes asupport structure 110 that in use is worn on the head of a user and has a general shape and appearance of an eyeglasses (e.g., sunglasses) frame.Support structure 110 carries multiple components, including: a projector 120 (a scanning laser projector in the illustrated example), aholographic combiner 130, and an exitpupil expansion optic 150. Portions ofprojector 120 and exitpupil expansion optic 150 may be contained within an inner volume ofsupport structure 110; however,FIG. 1 provides a partial-cutaway view in which regions ofsupport structure 110 have been removed in order to render visible portions ofprojector 120 and exitpupil expansion optic 150 that may otherwise be concealed. In accordance with the present systems, devices, and methods,support structure 110 also carries a virtualcontent control system 160 communicatively coupled toprojector 120. Virtualcontent control system 160 comprises aprocessor 161 and a non-transitory processor-readable storage medium ormemory 162 communicatively coupled to theprocessor 162.Memory 162 stores processor-executable virtual content control data and/orinstructions 163 that, when executed byprocessor 161, cause WHUD 100 to provide heterogeneous display quality with respect to at least one display parameter as discussed in more detail later on. - Throughout this specification and the appended claims, the term “carries” and variants such as “carried by” are generally used to refer to a physical coupling between two objects. The physical coupling may be direct physical coupling (i.e., with direct physical contact between the two objects) or indirect physical coupling that may be mediated by one or more additional objects. Thus, the term carries and variants such as “carried by” are meant to generally encompass all manner of direct and indirect physical coupling, including without limitation: carried on, carried within, physically coupled to, and/or supported by, with or without any number of intermediary physical objects therebetween.
-
Projector 120 is a scanning laser projector, though as previously described other forms of projectors may similarly be used, such as a digital light processing-based projector.Projector 120 includes multiple laser diodes (e.g., a red laser diode, a green laser diode, and/or a blue laser diode) and at least one scan mirror (e.g., a single two-dimensional scan mirror or two one-dimensional scan mirrors, which may be, e.g., MEMS-based or piezo-based). As previously described, a person of skill in the art will appreciate that the teachings herein may be applied in WHUDs that employ non-projector-based display architectures, such as WHUDs that employ microdisplays and/or waveguide structures. -
Holographic combiner 130 is positioned within a field of view of at least one eye of the user whensupport structure 110 is worn on the head of the user.Holographic combiner 130 is sufficiently optically transparent to permit light from the user's environment (i.e., “environmental light”) to pass through to the user's eye. In the illustrated example ofFIG. 1 ,support structure 110 further carries a transparent eyeglass lens 140 (e.g., a prescription eyeglass lens) andholographic combiner 130 comprises at least one layer of holographic material that is adhered to, affixed to, laminated with, carried in or upon, or otherwise integrated witheyeglass lens 140. The at least one layer of holographic material may include a photopolymer film such as Bayfol®HX available from Bayer MaterialScience AG or a silver halide compound and may, for example, be integrated withtransparent lens 140 using any of the techniques described in U.S. Provisional Patent Application Ser. No. 62/214,600.Holographic combiner 130 includes at least one hologram in or on the at least one layer of holographic material. Withholographic combiner 130 positioned in a field of view of an eye of the user whensupport structure 110 is worn on the head of the user, the at least one hologram ofholographic combiner 130 is positioned and oriented to redirect light originating fromprojector 120 towards the eye of the user. In particular, the at least one hologram is positioned and oriented to receive light signals that originate fromprojector 120 and converge those light signals to at least one exit pupil at or proximate the eye of the user. - Exit
pupil expansion optic 150 is positioned in an optical path betweenprojector 120 andholographic combiner 130 and may take on any of a variety of different forms, including without limitation those described in U.S. patent application Ser. No. 15/046,234, U.S. patent application Ser. No. 15/046,254, and/or U.S. patent application Ser. No. 15/046,269. - In accordance with the present systems, devices, and methods, the processor-executable virtual content control instructions (and/or data) 163, when executed by
processor 161 of virtual content control system, cause WHUD 100 to provide heterogeneous display quality with respect to at least one display parameter. Specifically, when executed byprocessor 161, processor-executable virtual content control instruction (and/or data) 163cause WHUD 100 to determine a region of interest in a FOV of the user, project virtual content with a high quality with respect to a first display parameter in the region of interest, and project virtual content with a relatively lower quality with respect to the first display parameter outside of the region of interest. As previously described, the first display parameter may include any of a variety of different display parameters depending on the specific implementation, including without limitation: resolution, number of pixels, pixel density, pixel size, brightness, color saturation, sharpness, focus, and/or noise. In some implementations, the WHUD (100) may provide heterogeneous (non-uniform) display quality with respect to multiple different display parameters, such as at least two different display parameters. - Throughout this specification and the appended claims the term “resolution” is used, with reference to display quality and/or virtual content projected by a projector (1200, to generally refer to a distribution of pixels or lines that make up a display and/or that make up virtual content of a display. In accordance with the present systems, devices, and methods, the “quality” of resolution may depend on a number of resolution parameters and, accordingly, the quality of resolution may be adjusted (i.e., made higher or lower) by tuning any one or combination of multiple ones of the resolution parameters. Exemplary resolution parameters that may be tuned in order to make the display quality of virtual content higher or lower with respect to display resolution include, without limitation: number of pixels, size of pixels, spacing in between pixels, and/or pixel density. For example, the display quality of WHUD 100 may be made higher with respect to resolution by increasing the number of pixels, decreasing the size of pixels, and/or increasing the pixel density. Conversely, the display quality of WHUD 100 may be made lower with respect to resolution by decreasing the number of pixels, increasing the size of pixels, and/or decreasing the pixel density.
- Resolution is just one example of a display parameter that may be varied of the FOV of a WHUD to provide heterogeneous (non-uniform) display quality in accordance with the present systems, devices, and methods. Brightness is another example of such a display parameter. For example, the display quality of WHUD 100 may be made higher with respect to brightness by increasing the brightness and the display quality of WHUD 100 may be made lower with respect to brightness by decreasing the brightness.
- In some implementations, providing heterogeneous (non-uniform) display quality with respect to a first display parameter may include heterogeneously varying at least a second display parameter over the FOV of a WHUD in order to compensate for one or more effect(s) of providing heterogeneous display quality with respect to the first display parameter. For example, regions of a WHUD's FOV that are displayed with relatively low quality resolution may be displayed with relatively higher brightness to compensate and reduce the likelihood that the user will perceive the non-uniformity in resolution.
- The region of interest in the FOV of the user in which virtual content is displayed (e.g., projected) with high quality with respect to a first display parameter may be determined (e.g., identified, deduced, or defined) in a variety of different ways. In some implementations, the region of interest may be an attribute of the virtual content itself and correspond to a region of the virtual content where the user is expected to attract their attention based on the nature of the virtual content. For example, if the virtual content comprises a block of text overlaid on a textured background, virtual
content control system 160 may determine (e.g., define or deduce) that the region of interest corresponds to the block of text as this is likely where the user will direct their attention. In some implementations, virtualcontent control system 160 may define a region of interest in order to strategically direct the user's attention (e.g., guide the user's gaze) to that region of the virtual content, for example, to highlight a new alert or notification and draw the user's attention thereto or to highlight a particular position on a map. - In other implementations, the region of interest may be identified by the WHUD (100) based on one or more property(ies) of the user's eye. For example,
WHUD 100 includes asensor 170 carried bysupport structure 110, wheresensor 170 is operative to sense, measure, detect, monitor, and/or track one or more property(ies) of the user's eye.Sensor 170 is communicatively coupled to virtualcontent control system 160 and data fromsensor 170 that is indicative or representative of one or more property(ies) of the user's eye may be used by virtualcontent control system 160 to determine a region of interest in a FOV of the user. Two exemplary eye properties that may be sensed, measured, detected, monitored, and/or tracked bysensor 170 and used by virtualcontent control system 160 to determine a region of interest in the user's FOV are now described. - In a first example, the region of interest in the FOV of the user may include a foveal region of the FOV of the user. A person of skill in the art will appreciate that the foveal region in the FOV of the user may generally correspond to light rays that impinge on the fovea (i.e., the “fovea centralis”) on the retina of the user's eye. The fovea is a depression in the inner surface of the retina (usually about 1.5 mm wide) that includes a relatively higher density of cone cells compared to the rest of the retinal surface. Due to this high density of cone cells, the fovea is generally the region of the retina that provides the sharpest (e.g., most detailed) vision and/or the highest visual acuity. When viewing an object or particularly fine detail, such as when reading text, humans have generally evolved to direct their gaze (e.g., adjust their eye position) so that light coming from the detailed object impinges on the fovea.
- In this first example,
WHUD 100 projects virtual content with high quality (with respect to a first display parameter) in the foveal region of the user's FOV by aligning the virtual content with the user's eye so that the high quality region of the virtual content aligns with (e.g., impinges on) the fovea of the retina of the user's eye. In order to determine the position of the fovea of the user's eye,sensor 170 may include a fovea tracker that is communicatively coupled to virtual content control system 160 (e.g., communicatively coupled toprocessor 161 of virtual content control system 160).Fovea tracker 170 is positioned and oriented to determine a position of the fovea of the user's eye and processor-executable virtualcontent control instructions 163 may, when executed byprocessor 161, cause WHUD 100 to determine (e.g., identify) the foveal region of the user's FOV based on the position of the fovea of the user's eye determined byfovea tracker 170. - Depending on the specific implementation,
fovea tracker 170 may employ a variety of different techniques. As an example,fovea tracker 170 may comprise an illumination source (e.g., a light source, such as an infrared light source) and/or an optical sensor such as a camera, a video camera, or a photodetector. With the eye sufficiently illuminated (e.g., by an illumination source component of fovea tracker 170), the optical sensor component offovea tracker 170 may sense, detect, measure, monitor, and/or track retinal blood vessels and/or other features on the inside of the user's eye from which the position of the fovea may be determined (e.g., identified). More specifically, the optical sensor component offovea tracker 170 may capture images of the user's eye and a processor communicatively coupled to the optical sensor (e.g., processor 161) may process the images to determine (e.g., identify) the position of the fovea based on, for example, discernible features of the retina (e.g., retinal blood vessels) in the images. Processing the mages by the processor may include executing, by the processor, processor-readable image processing data and/or instructions stored in a non-transitory processor-readable storage medium or memory (e.g., memory 162) of WHUD 100. - In a second example, the region of interest in the FOV of the user may be determined by
WHUD 100 based on the gaze direction of the user. To this end,sensor 170 may include an eye tracker carried bysupport structure 110 and positioned and oriented to determine a gaze direction of the eye of the user.Eye tracker 170 may be communicatively coupled to virtual content control system 160 (e.g., communicatively coupled toprocessor 161 of virtual content control system 160) and processor-executable virtualcontent control instructions 163 may, when executed byprocessor 161, cause WHUD 100 to determine (e.g., identify) a region of interest in the FOV of the user based on the gaze direction of the user's eye determined byeye tracker 170. - A person of skill in the art will appreciate that in different implementations,
eye tracker 170 itself may determine the gaze direction of the user's eye and relay this information toprocessor 161, orprocessor 161 may determine the gaze direction of the user's eye based on data and/or information provided byeye tracker 170. -
Eye tracker 170 may employ any of a variety of different eye tracking technologies depending on the specific implementation. For example,eye tracker 170 may employ any or all of the systems, devices, and methods described in U.S. Provisional Patent Application Ser. No. 62/167,767; U.S. Provisional Patent Application Ser. No. 62/271,135; U.S. Provisional Patent Application Ser. No. 62/245,792; and/or U.S. Provisional Patent Application Ser. No. 62/281,041. - Based on data and/or information about the gaze direction of the user's eye, virtual
content control system 160 may position the region of interest to align with the gaze direction of the user's eye so that the region of interest appears substantially centrally in the user's FOV and remains in this position for all eye positions over a wide range of eye positions. This approach may cause the region of interest to at least partially align with the foveal region in the user's FOV without direct determination of the position of the user's fovea. However, in some implementations the position of the fovea in the user's eye (and the corresponding position of the foveal region in the user's FOV) may be determined (e.g., deduced) by virtualcontent control system 160 based on the gaze direction of the user's eye because the position of the fovea in the user's eye is generally fixed relative to the positions of the pupil, iris, cornea, and/or other features of the user's eye that may be sensed, measured, detected, monitored, and/or tracked byeye tracker 170 in determining the gaze direction of the user's eye. - In some implementations of the present systems, devices, and methods, virtual content is dynamically projected with highest quality (with respect to at least one display parameter) in the region of the user's FOV that corresponds to the user's fovea (e.g., the foveal region of the user's FOV) and with relatively lower quality (with respect to the same at least one display parameter) elsewhere in the user's FOV (i.e., in regions of the user's FOV outside of the foveal region). The virtual content is “dynamic” in the sense that the high quality region “follows” the user's fovea (i.e., follows the foveal region in the user's FOV) based on the user's fovea position, eye position, and/or gaze direction as determined by
sensor 170. Since the user's ability to focus over the entire FOV is non-uniform, it is unnecessary to project (and to provide sufficient infrastructure, e.g., graphical processing power and overall system power to render the system capable of projecting) virtual content with high quality over the entire FOV. Rather, in accordance with the present systems, devices, and methods, only the foveal region (or another region of interest) of the virtual content may be projected at high quality while the peripheral region(s) of the virtual content may be projected at comparatively lower quality. -
FIG. 2 is an illustrative diagram showing a plan view of exemplary projected virtual content from a WHUD 200 that employs heterogeneous (non-uniform) display quality in accordance with the present systems, devices, and methods. In the illustrated example, virtual content corresponding to the “foveal region” of the user's FOV (as determined by an on-board fovea-tracking system and/or an on-board eye-tracking system, not illustrated inFIG. 2 to reduce clutter) is depicted with greater clarity (i.e., sharper focus, higher resolution, and/or higher brightness) compared to regions of the user's FOV that are outside of the foveal region (i.e., non-foveal regions) in order to illustrate that the foveal region has higher display quality than the non-foveal regions. -
FIG. 3 is a flow-diagram showing amethod 300 of operating a WHUD to display virtual content with heterogeneous (non-uniform) quality in accordance with the present systems, devices, and methods. The WHUD includes a projector and may be substantially similar to WHUD 100 fromFIG. 1 .Method 300 includes threeacts - At 301, a region of interest in the user's FOV is determined. This region of interest may be determined by the WHUD itself, for example, by the processor of a virtual content control system carried by the WHUD. In some implementations, this region of interest may be determined (e.g., defined) by (i.e., within) a software application executed by a processor on-board the WHUD based on an intention to motivate the user to focus on this particular region. In alternative implementations, this region of interest may be determined (e.g., identified or deduced) by the WHUD based on data and/or information provided by one or more sensor(s) (such as a fovea tracker and/or and eye tracker) based on the position of the fovea of the user's eye, the position of the user's eye, and/or the gaze direction of the user's eye. The region of interest may or may not include a foveal region of the user's FOV.
- At 302, the projector of the WHUD projects virtual content with a high quality with respect to a first display parameter in the region of interest in the FOV of the user.
- At 303, the projector of the WHUD projects virtual content with a relatively lower quality with respect to the first display parameter in regions of the field of view of the user that are outside of the region of interest.
- As previously described, the first display parameter may include any of a variety of different display parameters depending on the specific implementation, including without limitation: resolution, number of pixels, pixel density, pixel size, brightness, color saturation, sharpness, focus, and/or noise. As an example, at 302 the projector of the WHUD may project virtual content with a high brightness in the region of interest of the FOV of the user and at 303 the projector may project virtual content with a relatively lower brightness in regions of the FOV of the user that are outside of the region of interest. As another example, at 302 the projector of the WHUD may project virtual content with a high resolution in the region of interest of the FOV of the user and at 303 the projector may project virtual content with a relatively lower resolution in regions of the FOV of the user that are outside of the region of interest.
- As also previously described, the quality of the resolution of the virtual content may be varied in a number of different ways, especially when a projector is used in the display system of the WHUD. In some cases, the quality of the resolution may be varied by adjusting the number pixels in the virtual content, the size of the pixels in the virtual content, the size of the gaps between pixels in the virtual content, and/or the density of pixels in the virtual content. In a projector-based WHUD, the quality of the resolution of virtual content may be varied by adjusting either or both of the light modulation of the modulative light source (e.g., laser diodes, LEDs, or similar) and/or the operation of the one or more dynamically-variable reflector(s) (e.g., scan mirror(s)).
- As a first example, at 302 the projector may project virtual content with a high resolution in the region of interest of the FOV of the user by projecting virtual content with a first light modulation frequency in the region of interest in the FOV of the user and at 303 the projector may project virtual content with a relatively lower resolution in regions of the FOV of the user that are outside of the region of interest by projecting virtual content with a second light modulation frequency in regions of the FOV of the user that are outside of the region of interest. In this case, the first light modulation frequency is greater than the second light modulation frequency.
- As a second example, at 302 the projector (e.g., a scanning laser projector) may project virtual content with a high resolution in the region of interest of the FOV of the user by scanning virtual content with a first scanning step size in the region of interest in the FOV of the user and at 303 the projector may project virtual content with a relatively lower resolution in regions of the FOV of the user that are outside of the region of interest by scanning virtual content with a second scanning step size in regions of the FOV of the user that are outside of the region of interest. In this case, the first scanning step size is smaller than the second scanning step size.
- In general, using a scanning laser projector heterogeneous (non-uniform) display resolution may be achieved by operating either or both of the modulative light source and/or the dynamic scanner to project relatively fewer/larger pixels outside of the user's foveal region and relatively more/smaller pixels within the user's foveal region. The result may be more concentrated image detail (i.e., a higher number/concentration of distinct pixels) in the user's foveal region (as dynamically determined by the eye-tracker(s)) and reduced image detail (i.e., a lower number/concentration of distinct pixels) outside of the user's foveal region. The discrepancy is pixel concentration may significantly save on graphical processing and power consumption while nevertheless remaining substantially undetectable to the user, because the user typically cannot focus to high degrees of resolution on those regions of their field of view that are outside of the foveal region.
- The WHUDs described herein may include one or more on-board power sources (e.g., one or more battery(ies)), a wireless transceiver for sending/receiving wireless communications, and/or a tethered connector port for coupling to a computer and/or charging the one or more on-board power source(s).
- The WHUDs described herein may receive and respond to commands from the user in one or more of a variety of ways, including without limitation: voice commands through a microphone; touch commands through buttons, switches, or a touch sensitive surface; and/or gesture-based commands through gesture detection systems as described in, for example, U.S. Non-Provisional patent application Ser. No. 14/155,087, U.S. Non-Provisional patent application Ser. No. 14/155,107, PCT Patent Application PCT/US2014/057029, and/or U.S. Provisional Patent Application Ser. No. 62/236,060, all of which are incorporated by reference herein in their entirety.
- The various implementations of WHUDs described herein may include any or all of the technologies described in U.S. Provisional Patent Application Ser. No. 62/156,736, and/or U.S. Provisional Patent Application Ser. No. 62/242,844.
- Throughout this specification and the appended claims the term “communicative” as in “communicative pathway,” “communicative coupling,” and in variants such as “communicatively coupled,” is generally used to refer to any engineered arrangement for transferring and/or exchanging information. Exemplary communicative pathways include, but are not limited to, electrically conductive pathways (e.g., electrically conductive wires, electrically conductive traces), magnetic pathways (e.g., magnetic media), and/or optical pathways (e.g., optical fiber), and exemplary communicative couplings include, but are not limited to, electrical couplings, magnetic couplings, and/or optical couplings.
- Throughout this specification and the appended claims, infinitive verb forms are often used. Examples include, without limitation: “to detect,” “to provide,” “to transmit,” “to communicate,” “to process,” “to route,” and the like. Unless the specific context requires otherwise, such infinitive verb forms are used in an open, inclusive sense, that is as “to, at least, detect,” “to, at least, provide,” “to, at least, transmit,” and so on.
- The above description of illustrated embodiments, including what is described in the Abstract, is not intended to be exhaustive or to limit the embodiments to the precise forms disclosed. Although specific embodiments of and examples are described herein for illustrative purposes, various equivalent modifications can be made without departing from the spirit and scope of the disclosure, as will be recognized by those skilled in the relevant art. The teachings provided herein of the various embodiments can be applied to other portable and/or wearable electronic devices, not necessarily the exemplary wearable electronic devices generally described above.
- For instance, the foregoing detailed description has set forth various embodiments of the devices and/or processes via the use of block diagrams, schematics, and examples. Insofar as such block diagrams, schematics, and examples contain one or more functions and/or operations, it will be understood by those skilled in the art that each function and/or operation within such block diagrams, flowcharts, or examples can be implemented, individually and/or collectively, by a wide range of hardware, software, firmware, or virtually any combination thereof. In one embodiment, the present subject matter may be implemented via Application Specific Integrated Circuits (ASICs). However, those skilled in the art will recognize that the embodiments disclosed herein, in whole or in part, can be equivalently implemented in standard integrated circuits, as one or more computer programs executed by one or more computers (e.g., as one or more programs running on one or more computer systems), as one or more programs executed by on one or more controllers (e.g., microcontrollers) as one or more programs executed by one or more processors (e.g., microprocessors, central processing units, graphical processing units), as firmware, or as virtually any combination thereof, and that designing the circuitry and/or writing the code for the software and or firmware would be well within the skill of one of ordinary skill in the art in light of the teachings of this disclosure.
- When logic is implemented as software and stored in memory, logic or information can be stored on any processor-readable medium for use by or in connection with any processor-related system or method. In the context of this disclosure, a memory is a processor-readable medium that is an electronic, magnetic, optical, or other physical device or means that contains or stores a computer and/or processor program. Logic and/or the information can be embodied in any processor-readable medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions associated with logic and/or information.
- In the context of this specification, a “non-transitory processor-readable medium” can be any element that can store the program associated with logic and/or information for use by or in connection with the instruction execution system, apparatus, and/or device. The processor-readable medium can be, for example, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus or device. More specific examples (a non-exhaustive list) of the computer readable medium would include the following: a portable computer diskette (magnetic, compact flash card, secure digital, or the like), a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM, EEPROM, or Flash memory), a portable compact disc read-only memory (CDROM), digital tape, and other non-transitory media.
- The various embodiments described above can be combined to provide further embodiments. To the extent that they are not inconsistent with the specific teachings and definitions herein, all of the U.S. patents, U.S. patent application publications, U.S. patent applications, foreign patents, foreign patent applications and non-patent publications referred to in this specification and/or listed in the Application Data Sheet which are owned by Thalmic Labs Inc., including but not limited to: U.S. Provisional Patent Application Ser. No. 62/134,347; U.S. Provisional Patent Application Ser. No. 62/214,600, U.S. Provisional Patent Application Ser. No. 62/268,892, U.S. Provisional Patent Application Ser. No. 62/167,767, U.S. Provisional Patent Application Ser. No. 62/271,135, U.S. Provisional Patent Application Ser. No. 62/245,792, U.S. Provisional Patent Application Ser. No. 62/281,041, U.S. Provisional Patent Application Ser. No. 62/288,947, U.S. Non-Provisional patent application Ser. No. 14/155,087, U.S. Non-Provisional patent application Ser. No. 14/155,107, PCT Patent Application PCT/US2014/057029, U.S. Provisional Patent Application Ser. No. 62/236,060, U.S. Provisional Patent Application Ser. No. 62/156,736, U.S. Non-Provisional patent application Ser. No. 15/046,254, U.S. Non-Provisional patent application Ser. No. 15/046,234, U.S. Non-Provisional patent application Ser. No. 15/046,269, and U.S. Provisional Patent Application Ser. No. 62/242,844, are incorporated herein by reference, in their entirety. Aspects of the embodiments can be modified, if necessary, to employ systems, circuits and concepts of the various patents, applications and publications to provide yet further embodiments.
- These and other changes can be made to the embodiments in light of the above-detailed description. In general, in the following claims, the terms used should not be construed to limit the claims to the specific embodiments disclosed in the specification and the claims, but should be construed to include all possible embodiments along with the full scope of equivalents to which such claims are entitled. Accordingly, the claims are not limited by the disclosure.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/070,887 US20160274365A1 (en) | 2015-03-17 | 2016-03-15 | Systems, devices, and methods for wearable heads-up displays with heterogeneous display quality |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201562134347P | 2015-03-17 | 2015-03-17 | |
US15/070,887 US20160274365A1 (en) | 2015-03-17 | 2016-03-15 | Systems, devices, and methods for wearable heads-up displays with heterogeneous display quality |
Publications (1)
Publication Number | Publication Date |
---|---|
US20160274365A1 true US20160274365A1 (en) | 2016-09-22 |
Family
ID=56924859
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/070,887 Abandoned US20160274365A1 (en) | 2015-03-17 | 2016-03-15 | Systems, devices, and methods for wearable heads-up displays with heterogeneous display quality |
Country Status (1)
Country | Link |
---|---|
US (1) | US20160274365A1 (en) |
Cited By (125)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160018653A1 (en) * | 2014-01-24 | 2016-01-21 | Osterhout Group, Inc. | See-through computer display systems |
US20170098330A1 (en) * | 2015-07-14 | 2017-04-06 | Colopl, Inc. | Method for controlling head mounted display, and program for controlling head mounted display |
US20170124760A1 (en) * | 2015-10-29 | 2017-05-04 | Sony Computer Entertainment Inc. | Foveated geometry tessellation |
US9651783B2 (en) | 2014-01-21 | 2017-05-16 | Osterhout Group, Inc. | See-through computer display systems |
US9651787B2 (en) | 2014-04-25 | 2017-05-16 | Osterhout Group, Inc. | Speaker assembly for headworn computer |
US9684172B2 (en) | 2014-12-03 | 2017-06-20 | Osterhout Group, Inc. | Head worn computer display systems |
USD792400S1 (en) | 2014-12-31 | 2017-07-18 | Osterhout Group, Inc. | Computer glasses |
US9720234B2 (en) | 2014-01-21 | 2017-08-01 | Osterhout Group, Inc. | See-through computer display systems |
US9720241B2 (en) | 2014-06-09 | 2017-08-01 | Osterhout Group, Inc. | Content presentation in head worn computing |
US9740280B2 (en) | 2014-01-21 | 2017-08-22 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9746686B2 (en) | 2014-05-19 | 2017-08-29 | Osterhout Group, Inc. | Content position calibration in head worn computing |
US9753288B2 (en) | 2014-01-21 | 2017-09-05 | Osterhout Group, Inc. | See-through computer display systems |
US9766463B2 (en) | 2014-01-21 | 2017-09-19 | Osterhout Group, Inc. | See-through computer display systems |
US9766449B2 (en) | 2014-06-25 | 2017-09-19 | Thalmic Labs Inc. | Systems, devices, and methods for wearable heads-up displays |
US9772492B2 (en) | 2014-01-21 | 2017-09-26 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9798148B2 (en) | 2014-07-08 | 2017-10-24 | Osterhout Group, Inc. | Optical configurations for head-worn see-through displays |
US9811152B2 (en) | 2014-01-21 | 2017-11-07 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9829707B2 (en) | 2014-08-12 | 2017-11-28 | Osterhout Group, Inc. | Measuring content brightness in head worn computing |
US9836122B2 (en) | 2014-01-21 | 2017-12-05 | Osterhout Group, Inc. | Eye glint imaging in see-through computer display systems |
US9841599B2 (en) | 2014-06-05 | 2017-12-12 | Osterhout Group, Inc. | Optical configurations for head-worn see-through displays |
US9841602B2 (en) | 2014-02-11 | 2017-12-12 | Osterhout Group, Inc. | Location indicating avatar in head worn computing |
US9843093B2 (en) | 2014-02-11 | 2017-12-12 | Osterhout Group, Inc. | Spatial location presentation in head worn computing |
US9904051B2 (en) | 2015-10-23 | 2018-02-27 | Thalmic Labs Inc. | Systems, devices, and methods for laser eye tracking |
US9910284B1 (en) | 2016-09-08 | 2018-03-06 | Osterhout Group, Inc. | Optical systems for head-worn computers |
US9928019B2 (en) | 2014-02-14 | 2018-03-27 | Osterhout Group, Inc. | Object shadowing in head worn computing |
US9939934B2 (en) | 2014-01-17 | 2018-04-10 | Osterhout Group, Inc. | External user interface for head worn computing |
US9958682B1 (en) | 2015-02-17 | 2018-05-01 | Thalmic Labs Inc. | Systems, devices, and methods for splitter optics in wearable heads-up displays |
US9965681B2 (en) | 2008-12-16 | 2018-05-08 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9989764B2 (en) | 2015-02-17 | 2018-06-05 | Thalmic Labs Inc. | Systems, devices, and methods for eyebox expansion in wearable heads-up displays |
US10001644B2 (en) | 2014-01-21 | 2018-06-19 | Osterhout Group, Inc. | See-through computer display systems |
US20180196512A1 (en) * | 2017-01-10 | 2018-07-12 | Samsung Electronics Co., Ltd. | Method for outputting image and electronic device supporting the same |
US10062182B2 (en) | 2015-02-17 | 2018-08-28 | Osterhout Group, Inc. | See-through computer display systems |
WO2018160507A1 (en) * | 2017-03-03 | 2018-09-07 | Microsoft Technology Licensing, Llc | Mems scanning display device |
US10073268B2 (en) | 2015-05-28 | 2018-09-11 | Thalmic Labs Inc. | Display with integrated visible light eye tracking |
US10078224B2 (en) | 2014-09-26 | 2018-09-18 | Osterhout Group, Inc. | See-through computer display systems |
WO2018183405A1 (en) * | 2017-03-27 | 2018-10-04 | Avegant Corp. | Steerable foveal display |
WO2018187096A1 (en) * | 2017-04-03 | 2018-10-11 | Microsoft Technology Licensing, Llc | Wide field of view scanning display |
CN108737724A (en) * | 2017-04-17 | 2018-11-02 | 英特尔公司 | The system and method for capturing and showing for 360 videos |
US10126815B2 (en) | 2016-01-20 | 2018-11-13 | Thalmic Labs Inc. | Systems, devices, and methods for proximity-based eye tracking |
US10133075B2 (en) | 2015-05-04 | 2018-11-20 | Thalmic Labs Inc. | Systems, devices, and methods for angle- and wavelength-multiplexed holographic optical elements |
US10151926B2 (en) | 2016-01-29 | 2018-12-11 | North Inc. | Systems, devices, and methods for preventing eyebox degradation in a wearable heads-up display |
WO2018208361A3 (en) * | 2017-03-03 | 2018-12-20 | Microsoft Technology Licensing, Llc | Mems scanning display device |
US10162182B2 (en) * | 2015-08-03 | 2018-12-25 | Facebook Technologies, Llc | Enhanced pixel resolution through non-uniform ocular projection |
CN109189357A (en) * | 2018-08-30 | 2019-01-11 | Oppo广东移动通信有限公司 | Information display method, device, intelligent glasses and storage medium |
JP2019012258A (en) * | 2017-06-30 | 2019-01-24 | エルジー ディスプレイ カンパニー リミテッド | Display device and gate driving circuit thereof |
US10191279B2 (en) | 2014-03-17 | 2019-01-29 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US20190033595A1 (en) * | 2017-07-26 | 2019-01-31 | Thalmic Labs Inc. | Systems, devices, and methods for narrow waveband laser diodes |
US10215987B2 (en) | 2016-11-10 | 2019-02-26 | North Inc. | Systems, devices, and methods for astigmatism compensation in a wearable heads-up display |
US10230929B2 (en) | 2016-07-27 | 2019-03-12 | North Inc. | Systems, devices, and methods for laser projectors |
US10247858B2 (en) | 2015-10-25 | 2019-04-02 | Facebook Technologies, Llc | Liquid crystal half-wave plate lens |
US10254856B2 (en) | 2014-01-17 | 2019-04-09 | Osterhout Group, Inc. | External user interface for head worn computing |
US10297180B2 (en) | 2015-08-03 | 2019-05-21 | Facebook Technologies, Llc | Compensation of chromatic dispersion in a tunable beam steering device for improved display |
US20190196206A1 (en) * | 2017-12-22 | 2019-06-27 | North Inc. | Grating waveguide combiner for optical engine |
US10338451B2 (en) | 2015-08-03 | 2019-07-02 | Facebook Technologies, Llc | Devices and methods for removing zeroth order leakage in beam steering devices |
US10365549B2 (en) | 2016-04-13 | 2019-07-30 | North Inc. | Systems, devices, and methods for focusing laser projectors |
US10365492B2 (en) | 2016-12-23 | 2019-07-30 | North Inc. | Systems, devices, and methods for beam combining in wearable heads-up displays |
US20190272801A1 (en) * | 2018-03-01 | 2019-09-05 | Beijing Boe Optoelectronics Technology Co., Ltd. | Processing method and processing device for display data, and display device |
US10409057B2 (en) | 2016-11-30 | 2019-09-10 | North Inc. | Systems, devices, and methods for laser eye tracking in wearable heads-up displays |
WO2019170959A1 (en) * | 2018-03-06 | 2019-09-12 | Varjo Technologies Oy | Display apparatus and method of displaying using controllable scanning mirror |
US10416454B2 (en) | 2015-10-25 | 2019-09-17 | Facebook Technologies, Llc | Combination prism array for focusing light |
US10422995B2 (en) | 2017-07-24 | 2019-09-24 | Mentor Acquisition One, Llc | See-through computer display systems with stray light management |
US10437074B2 (en) | 2017-01-25 | 2019-10-08 | North Inc. | Systems, devices, and methods for beam combining in laser projectors |
US10459221B2 (en) | 2016-08-12 | 2019-10-29 | North Inc. | Systems, devices, and methods for variable luminance in wearable heads-up displays |
US10459305B2 (en) | 2015-08-03 | 2019-10-29 | Facebook Technologies, Llc | Time-domain adjustment of phase retardation in a liquid crystal grating for a color display |
USD864959S1 (en) * | 2017-01-04 | 2019-10-29 | Mentor Acquisition One, Llc | Computer glasses |
US10466491B2 (en) | 2016-06-01 | 2019-11-05 | Mentor Acquisition One, Llc | Modular systems for head-worn computers |
US10474296B1 (en) * | 2018-07-12 | 2019-11-12 | Microvision, Inc. | Laser scanning devices and methods with touch detection |
US10488662B2 (en) | 2015-09-04 | 2019-11-26 | North Inc. | Systems, articles, and methods for integrating holographic optical elements with eyeglass lenses |
US10499021B2 (en) * | 2017-04-11 | 2019-12-03 | Microsoft Technology Licensing, Llc | Foveated MEMS scanning display |
DE102018209886A1 (en) * | 2018-06-19 | 2019-12-19 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Device for projecting a laser beam to generate an image on the retina of an eye and glasses device with two such devices |
US10528135B2 (en) | 2013-01-14 | 2020-01-07 | Ctrl-Labs Corporation | Wearable muscle interface systems, devices and methods that interact with content displayed on an electronic display |
US10552676B2 (en) | 2015-08-03 | 2020-02-04 | Facebook Technologies, Llc | Methods and devices for eye tracking based on depth sensing |
USD877237S1 (en) * | 2017-12-07 | 2020-03-03 | Amazon Technologies, Inc. | Wearable device |
US10578869B2 (en) * | 2017-07-24 | 2020-03-03 | Mentor Acquisition One, Llc | See-through computer display systems with adjustable zoom cameras |
US10613323B1 (en) * | 2017-12-13 | 2020-04-07 | Facebook Technologies, Llc | Transition feature for framing multizone optics |
US10649220B2 (en) | 2014-06-09 | 2020-05-12 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US10656822B2 (en) | 2015-10-01 | 2020-05-19 | North Inc. | Systems, devices, and methods for interacting with content displayed on head-mounted displays |
US10663740B2 (en) | 2014-06-09 | 2020-05-26 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US10670928B2 (en) | 2015-12-21 | 2020-06-02 | Facebook Technologies, Llc | Wide angle beam steering for virtual reality and augmented reality |
US10684692B2 (en) | 2014-06-19 | 2020-06-16 | Facebook Technologies, Llc | Systems, devices, and methods for gesture identification |
US10684478B2 (en) | 2016-05-09 | 2020-06-16 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US10684687B2 (en) | 2014-12-03 | 2020-06-16 | Mentor Acquisition One, Llc | See-through computer display systems |
US10719127B1 (en) * | 2018-08-29 | 2020-07-21 | Rockwell Collins, Inc. | Extended life display by utilizing eye tracking |
US10777087B2 (en) * | 2018-12-07 | 2020-09-15 | International Business Machines Corporation | Augmented reality for removing external stimuli |
WO2020190519A1 (en) * | 2019-03-15 | 2020-09-24 | Microsoft Technology Licensing, Llc | Holographic image generated based on eye position |
US10802190B2 (en) | 2015-12-17 | 2020-10-13 | Covestro Llc | Systems, devices, and methods for curved holographic optical elements |
US10824253B2 (en) | 2016-05-09 | 2020-11-03 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US10853589B2 (en) | 2014-04-25 | 2020-12-01 | Mentor Acquisition One, Llc | Language translation with head-worn computing |
CN112105982A (en) * | 2018-05-04 | 2020-12-18 | 哈曼国际工业有限公司 | Head-up display without mirror |
EP3631558A4 (en) * | 2017-05-29 | 2020-12-30 | Eyeway Vision Ltd. | Image projection system |
US10901216B2 (en) | 2017-10-23 | 2021-01-26 | Google Llc | Free space multiple laser diode modules |
US10969584B2 (en) | 2017-08-04 | 2021-04-06 | Mentor Acquisition One, Llc | Image expansion optic for head-worn computer |
US11024007B2 (en) | 2015-04-20 | 2021-06-01 | Intel Corporation | Apparatus and method for non-uniform frame buffer rasterization |
US11032471B2 (en) * | 2016-06-30 | 2021-06-08 | Nokia Technologies Oy | Method and apparatus for providing a visual indication of a point of interest outside of a user's view |
EP3687465A4 (en) * | 2017-09-27 | 2021-06-16 | University of Miami | Digital therapeutic corrective spectacles |
US11079846B2 (en) | 2013-11-12 | 2021-08-03 | Facebook Technologies, Llc | Systems, articles, and methods for capacitive electromyography sensors |
US11104272B2 (en) | 2014-03-28 | 2021-08-31 | Mentor Acquisition One, Llc | System for assisted operator safety using an HMD |
US11103122B2 (en) | 2014-07-15 | 2021-08-31 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11122256B1 (en) * | 2017-08-07 | 2021-09-14 | Apple Inc. | Mixed reality system |
US11126261B2 (en) | 2019-01-07 | 2021-09-21 | Avegant Corp. | Display control system and rendering pipeline |
US11164352B2 (en) * | 2017-04-21 | 2021-11-02 | Intel Corporation | Low power foveated rendering to save power on GPU and/or display |
US11169383B2 (en) | 2018-12-07 | 2021-11-09 | Avegant Corp. | Steerable positioning element |
DE102020206821A1 (en) | 2020-06-02 | 2021-12-02 | Robert Bosch Gesellschaft mit beschränkter Haftung | Method for determining a line of sight of an eye |
DE102020206822A1 (en) | 2020-06-02 | 2021-12-02 | Robert Bosch Gesellschaft mit beschränkter Haftung | Procedure for operating data glasses |
US11269182B2 (en) | 2014-07-15 | 2022-03-08 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
USD946572S1 (en) * | 2020-12-02 | 2022-03-22 | Eissa Nick Asemani | Visor mounted eyewear |
US11398069B2 (en) * | 2017-04-01 | 2022-07-26 | Intel Corporation | Temporal data structures in a ray tracing architecture |
US11409105B2 (en) * | 2017-07-24 | 2022-08-09 | Mentor Acquisition One, Llc | See-through computer display systems |
US11487110B2 (en) | 2014-01-21 | 2022-11-01 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US11586049B2 (en) | 2019-03-29 | 2023-02-21 | Avegant Corp. | Steerable hybrid display using a waveguide |
US11624921B2 (en) | 2020-01-06 | 2023-04-11 | Avegant Corp. | Head mounted system with color specific modulation |
US11635736B2 (en) | 2017-10-19 | 2023-04-25 | Meta Platforms Technologies, Llc | Systems and methods for identifying biological structures associated with neuromuscular source signals |
USD985563S1 (en) * | 2021-01-06 | 2023-05-09 | Samsung Electronics Co., Ltd. | Head-mounted display |
US11644799B2 (en) | 2013-10-04 | 2023-05-09 | Meta Platforms Technologies, Llc | Systems, articles and methods for wearable electronic devices employing contact sensors |
USD985562S1 (en) * | 2021-01-21 | 2023-05-09 | Samsung Electronics Co., Ltd. | Head-mounted display |
USD985564S1 (en) * | 2021-01-06 | 2023-05-09 | Samsung Electronics Co., Ltd. | Head-mounted display |
US11666264B1 (en) | 2013-11-27 | 2023-06-06 | Meta Platforms Technologies, Llc | Systems, articles, and methods for electromyography sensors |
US11669163B2 (en) | 2014-01-21 | 2023-06-06 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US11714284B2 (en) * | 2016-09-20 | 2023-08-01 | Apple Inc. | Display device including foveal and peripheral projectors |
US11797087B2 (en) | 2018-11-27 | 2023-10-24 | Meta Platforms Technologies, Llc | Methods and apparatus for autocalibration of a wearable electrode sensor system |
US11868531B1 (en) | 2021-04-08 | 2024-01-09 | Meta Platforms Technologies, Llc | Wearable device providing for thumb-to-finger-based input gestures detected based on neuromuscular signals, and systems and methods of use thereof |
EP4121813A4 (en) * | 2020-03-20 | 2024-01-17 | Magic Leap Inc | Systems and methods for retinal imaging and tracking |
US11907423B2 (en) | 2019-11-25 | 2024-02-20 | Meta Platforms Technologies, Llc | Systems and methods for contextualized interactions with an environment |
US11921471B2 (en) | 2013-08-16 | 2024-03-05 | Meta Platforms Technologies, Llc | Systems, articles, and methods for wearable devices having secondary power sources in links of a band for providing secondary power in addition to a primary power source |
US11961494B1 (en) | 2019-03-29 | 2024-04-16 | Meta Platforms Technologies, Llc | Electromagnetic interference reduction in extended reality environments |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100056274A1 (en) * | 2008-08-28 | 2010-03-04 | Nokia Corporation | Visual cognition aware display and visual data transmission architecture |
US20120024797A1 (en) * | 2010-04-06 | 2012-02-02 | Kale Aniket | Methods for dewatering wet algal cell cultures |
US20120249797A1 (en) * | 2010-02-28 | 2012-10-04 | Osterhout Group, Inc. | Head-worn adaptive display |
US20150205126A1 (en) * | 2013-11-27 | 2015-07-23 | Magic Leap, Inc. | Virtual and augmented reality systems and methods |
US20160154244A1 (en) * | 2014-01-21 | 2016-06-02 | Osterhout Group, Inc. | Compact optical system for head-worn computer |
-
2016
- 2016-03-15 US US15/070,887 patent/US20160274365A1/en not_active Abandoned
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100056274A1 (en) * | 2008-08-28 | 2010-03-04 | Nokia Corporation | Visual cognition aware display and visual data transmission architecture |
US20120249797A1 (en) * | 2010-02-28 | 2012-10-04 | Osterhout Group, Inc. | Head-worn adaptive display |
US20120024797A1 (en) * | 2010-04-06 | 2012-02-02 | Kale Aniket | Methods for dewatering wet algal cell cultures |
US20150205126A1 (en) * | 2013-11-27 | 2015-07-23 | Magic Leap, Inc. | Virtual and augmented reality systems and methods |
US20160154244A1 (en) * | 2014-01-21 | 2016-06-02 | Osterhout Group, Inc. | Compact optical system for head-worn computer |
Cited By (313)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9965681B2 (en) | 2008-12-16 | 2018-05-08 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US11009951B2 (en) | 2013-01-14 | 2021-05-18 | Facebook Technologies, Llc | Wearable muscle interface systems, devices and methods that interact with content displayed on an electronic display |
US10528135B2 (en) | 2013-01-14 | 2020-01-07 | Ctrl-Labs Corporation | Wearable muscle interface systems, devices and methods that interact with content displayed on an electronic display |
US11921471B2 (en) | 2013-08-16 | 2024-03-05 | Meta Platforms Technologies, Llc | Systems, articles, and methods for wearable devices having secondary power sources in links of a band for providing secondary power in addition to a primary power source |
US11644799B2 (en) | 2013-10-04 | 2023-05-09 | Meta Platforms Technologies, Llc | Systems, articles and methods for wearable electronic devices employing contact sensors |
US11079846B2 (en) | 2013-11-12 | 2021-08-03 | Facebook Technologies, Llc | Systems, articles, and methods for capacitive electromyography sensors |
US11666264B1 (en) | 2013-11-27 | 2023-06-06 | Meta Platforms Technologies, Llc | Systems, articles, and methods for electromyography sensors |
US11782529B2 (en) | 2014-01-17 | 2023-10-10 | Mentor Acquisition One, Llc | External user interface for head worn computing |
US11169623B2 (en) | 2014-01-17 | 2021-11-09 | Mentor Acquisition One, Llc | External user interface for head worn computing |
US11231817B2 (en) | 2014-01-17 | 2022-01-25 | Mentor Acquisition One, Llc | External user interface for head worn computing |
US11507208B2 (en) | 2014-01-17 | 2022-11-22 | Mentor Acquisition One, Llc | External user interface for head worn computing |
US10254856B2 (en) | 2014-01-17 | 2019-04-09 | Osterhout Group, Inc. | External user interface for head worn computing |
US9939934B2 (en) | 2014-01-17 | 2018-04-10 | Osterhout Group, Inc. | External user interface for head worn computing |
US9753288B2 (en) | 2014-01-21 | 2017-09-05 | Osterhout Group, Inc. | See-through computer display systems |
US10007118B2 (en) | 2014-01-21 | 2018-06-26 | Osterhout Group, Inc. | Compact optical system with improved illumination |
US9720234B2 (en) | 2014-01-21 | 2017-08-01 | Osterhout Group, Inc. | See-through computer display systems |
US9720235B2 (en) | 2014-01-21 | 2017-08-01 | Osterhout Group, Inc. | See-through computer display systems |
US10191284B2 (en) | 2014-01-21 | 2019-01-29 | Osterhout Group, Inc. | See-through computer display systems |
US9740012B2 (en) | 2014-01-21 | 2017-08-22 | Osterhout Group, Inc. | See-through computer display systems |
US9740280B2 (en) | 2014-01-21 | 2017-08-22 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US10481393B2 (en) | 2014-01-21 | 2019-11-19 | Mentor Acquisition One, Llc | See-through computer display systems |
US10222618B2 (en) | 2014-01-21 | 2019-03-05 | Osterhout Group, Inc. | Compact optics with reduced chromatic aberrations |
US9766463B2 (en) | 2014-01-21 | 2017-09-19 | Osterhout Group, Inc. | See-through computer display systems |
US11796799B2 (en) | 2014-01-21 | 2023-10-24 | Mentor Acquisition One, Llc | See-through computer display systems |
US9772492B2 (en) | 2014-01-21 | 2017-09-26 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US11796805B2 (en) | 2014-01-21 | 2023-10-24 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US9811152B2 (en) | 2014-01-21 | 2017-11-07 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9811159B2 (en) | 2014-01-21 | 2017-11-07 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9684171B2 (en) | 2014-01-21 | 2017-06-20 | Osterhout Group, Inc. | See-through computer display systems |
US9829703B2 (en) | 2014-01-21 | 2017-11-28 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9836122B2 (en) | 2014-01-21 | 2017-12-05 | Osterhout Group, Inc. | Eye glint imaging in see-through computer display systems |
US10579140B2 (en) | 2014-01-21 | 2020-03-03 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US10698223B2 (en) | 2014-01-21 | 2020-06-30 | Mentor Acquisition One, Llc | See-through computer display systems |
US10866420B2 (en) | 2014-01-21 | 2020-12-15 | Mentor Acquisition One, Llc | See-through computer display systems |
US10890760B2 (en) | 2014-01-21 | 2021-01-12 | Mentor Acquisition One, Llc | See-through computer display systems |
US9885868B2 (en) | 2014-01-21 | 2018-02-06 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US11669163B2 (en) | 2014-01-21 | 2023-06-06 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US11002961B2 (en) | 2014-01-21 | 2021-05-11 | Mentor Acquisition One, Llc | See-through computer display systems |
US11054902B2 (en) | 2014-01-21 | 2021-07-06 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US9933622B2 (en) | 2014-01-21 | 2018-04-03 | Osterhout Group, Inc. | See-through computer display systems |
US9651783B2 (en) | 2014-01-21 | 2017-05-16 | Osterhout Group, Inc. | See-through computer display systems |
US11650416B2 (en) | 2014-01-21 | 2023-05-16 | Mentor Acquisition One, Llc | See-through computer display systems |
US11947126B2 (en) | 2014-01-21 | 2024-04-02 | Mentor Acquisition One, Llc | See-through computer display systems |
US9971156B2 (en) | 2014-01-21 | 2018-05-15 | Osterhout Group, Inc. | See-through computer display systems |
US11099380B2 (en) | 2014-01-21 | 2021-08-24 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US10001644B2 (en) | 2014-01-21 | 2018-06-19 | Osterhout Group, Inc. | See-through computer display systems |
US11353957B2 (en) | 2014-01-21 | 2022-06-07 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US10012838B2 (en) | 2014-01-21 | 2018-07-03 | Osterhout Group, Inc. | Compact optical system with improved contrast uniformity |
US11619820B2 (en) | 2014-01-21 | 2023-04-04 | Mentor Acquisition One, Llc | See-through computer display systems |
US10012840B2 (en) | 2014-01-21 | 2018-07-03 | Osterhout Group, Inc. | See-through computer display systems |
US11622426B2 (en) | 2014-01-21 | 2023-04-04 | Mentor Acquisition One, Llc | See-through computer display systems |
US11487110B2 (en) | 2014-01-21 | 2022-11-01 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US20160018651A1 (en) * | 2014-01-24 | 2016-01-21 | Osterhout Group, Inc. | See-through computer display systems |
US20160018652A1 (en) * | 2014-01-24 | 2016-01-21 | Osterhout Group, Inc. | See-through computer display systems |
US20160018645A1 (en) * | 2014-01-24 | 2016-01-21 | Osterhout Group, Inc. | See-through computer display systems |
US11443493B2 (en) * | 2014-01-24 | 2022-09-13 | Mentor Acquisition One, Llc | Modification of peripheral content in world-locked see-through computer display systems |
US20160170207A1 (en) * | 2014-01-24 | 2016-06-16 | Osterhout Group, Inc. | See-through computer display systems |
US20160018653A1 (en) * | 2014-01-24 | 2016-01-21 | Osterhout Group, Inc. | See-through computer display systems |
US11900554B2 (en) * | 2014-01-24 | 2024-02-13 | Mentor Acquisition One, Llc | Modification of peripheral content in world-locked see-through computer display systems |
US20160171769A1 (en) * | 2014-01-24 | 2016-06-16 | Osterhout Group, Inc. | See-through computer display systems |
US20160085072A1 (en) * | 2014-01-24 | 2016-03-24 | Osterhout Group, Inc. | See-through computer display systems |
US11145132B2 (en) * | 2014-01-24 | 2021-10-12 | Mentor Acquisition One, Llc | Modification of peripheral content in world-locked see-through computer display systems |
US20220358732A1 (en) * | 2014-01-24 | 2022-11-10 | Mentor Acquisition One, Llc | Modification of peripheral content in world-locked see-through computer display systems |
US20160018654A1 (en) * | 2014-01-24 | 2016-01-21 | Osterhout Group, Inc. | See-through computer display systems |
US9841602B2 (en) | 2014-02-11 | 2017-12-12 | Osterhout Group, Inc. | Location indicating avatar in head worn computing |
US9843093B2 (en) | 2014-02-11 | 2017-12-12 | Osterhout Group, Inc. | Spatial location presentation in head worn computing |
US9928019B2 (en) | 2014-02-14 | 2018-03-27 | Osterhout Group, Inc. | Object shadowing in head worn computing |
US10191279B2 (en) | 2014-03-17 | 2019-01-29 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US11104272B2 (en) | 2014-03-28 | 2021-08-31 | Mentor Acquisition One, Llc | System for assisted operator safety using an HMD |
US11727223B2 (en) | 2014-04-25 | 2023-08-15 | Mentor Acquisition One, Llc | Language translation with head-worn computing |
US9651787B2 (en) | 2014-04-25 | 2017-05-16 | Osterhout Group, Inc. | Speaker assembly for headworn computer |
US10853589B2 (en) | 2014-04-25 | 2020-12-01 | Mentor Acquisition One, Llc | Language translation with head-worn computing |
US10634922B2 (en) | 2014-04-25 | 2020-04-28 | Mentor Acquisition One, Llc | Speaker assembly for headworn computer |
US11880041B2 (en) | 2014-04-25 | 2024-01-23 | Mentor Acquisition One, Llc | Speaker assembly for headworn computer |
US11474360B2 (en) | 2014-04-25 | 2022-10-18 | Mentor Acquisition One, Llc | Speaker assembly for headworn computer |
US9746686B2 (en) | 2014-05-19 | 2017-08-29 | Osterhout Group, Inc. | Content position calibration in head worn computing |
US10877270B2 (en) | 2014-06-05 | 2020-12-29 | Mentor Acquisition One, Llc | Optical configurations for head-worn see-through displays |
US11960089B2 (en) | 2014-06-05 | 2024-04-16 | Mentor Acquisition One, Llc | Optical configurations for head-worn see-through displays |
US9841599B2 (en) | 2014-06-05 | 2017-12-12 | Osterhout Group, Inc. | Optical configurations for head-worn see-through displays |
US11402639B2 (en) | 2014-06-05 | 2022-08-02 | Mentor Acquisition One, Llc | Optical configurations for head-worn see-through displays |
US11360318B2 (en) | 2014-06-09 | 2022-06-14 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11327323B2 (en) | 2014-06-09 | 2022-05-10 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US10976559B2 (en) | 2014-06-09 | 2021-04-13 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US10139635B2 (en) | 2014-06-09 | 2018-11-27 | Osterhout Group, Inc. | Content presentation in head worn computing |
US11022810B2 (en) | 2014-06-09 | 2021-06-01 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11663794B2 (en) | 2014-06-09 | 2023-05-30 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US10663740B2 (en) | 2014-06-09 | 2020-05-26 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11790617B2 (en) | 2014-06-09 | 2023-10-17 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US9720241B2 (en) | 2014-06-09 | 2017-08-01 | Osterhout Group, Inc. | Content presentation in head worn computing |
US10649220B2 (en) | 2014-06-09 | 2020-05-12 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11887265B2 (en) | 2014-06-09 | 2024-01-30 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US10684692B2 (en) | 2014-06-19 | 2020-06-16 | Facebook Technologies, Llc | Systems, devices, and methods for gesture identification |
US10012829B2 (en) | 2014-06-25 | 2018-07-03 | Thalmic Labs Inc. | Systems, devices, and methods for wearable heads-up displays |
US10067337B2 (en) | 2014-06-25 | 2018-09-04 | Thalmic Labs Inc. | Systems, devices, and methods for wearable heads-up displays |
US10054788B2 (en) | 2014-06-25 | 2018-08-21 | Thalmic Labs Inc. | Systems, devices, and methods for wearable heads-up displays |
US9874744B2 (en) | 2014-06-25 | 2018-01-23 | Thalmic Labs Inc. | Systems, devices, and methods for wearable heads-up displays |
US9766449B2 (en) | 2014-06-25 | 2017-09-19 | Thalmic Labs Inc. | Systems, devices, and methods for wearable heads-up displays |
US10564426B2 (en) | 2014-07-08 | 2020-02-18 | Mentor Acquisition One, Llc | Optical configurations for head-worn see-through displays |
US11409110B2 (en) | 2014-07-08 | 2022-08-09 | Mentor Acquisition One, Llc | Optical configurations for head-worn see-through displays |
US11940629B2 (en) | 2014-07-08 | 2024-03-26 | Mentor Acquisition One, Llc | Optical configurations for head-worn see-through displays |
US10775630B2 (en) | 2014-07-08 | 2020-09-15 | Mentor Acquisition One, Llc | Optical configurations for head-worn see-through displays |
US9798148B2 (en) | 2014-07-08 | 2017-10-24 | Osterhout Group, Inc. | Optical configurations for head-worn see-through displays |
US11269182B2 (en) | 2014-07-15 | 2022-03-08 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11103122B2 (en) | 2014-07-15 | 2021-08-31 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11786105B2 (en) | 2014-07-15 | 2023-10-17 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US9829707B2 (en) | 2014-08-12 | 2017-11-28 | Osterhout Group, Inc. | Measuring content brightness in head worn computing |
US10908422B2 (en) | 2014-08-12 | 2021-02-02 | Mentor Acquisition One, Llc | Measuring content brightness in head worn computing |
US11360314B2 (en) | 2014-08-12 | 2022-06-14 | Mentor Acquisition One, Llc | Measuring content brightness in head worn computing |
US11630315B2 (en) | 2014-08-12 | 2023-04-18 | Mentor Acquisition One, Llc | Measuring content brightness in head worn computing |
US10078224B2 (en) | 2014-09-26 | 2018-09-18 | Osterhout Group, Inc. | See-through computer display systems |
US10684687B2 (en) | 2014-12-03 | 2020-06-16 | Mentor Acquisition One, Llc | See-through computer display systems |
US11262846B2 (en) | 2014-12-03 | 2022-03-01 | Mentor Acquisition One, Llc | See-through computer display systems |
US9684172B2 (en) | 2014-12-03 | 2017-06-20 | Osterhout Group, Inc. | Head worn computer display systems |
US11809628B2 (en) | 2014-12-03 | 2023-11-07 | Mentor Acquisition One, Llc | See-through computer display systems |
USD792400S1 (en) | 2014-12-31 | 2017-07-18 | Osterhout Group, Inc. | Computer glasses |
US10062182B2 (en) | 2015-02-17 | 2018-08-28 | Osterhout Group, Inc. | See-through computer display systems |
US10191283B2 (en) | 2015-02-17 | 2019-01-29 | North Inc. | Systems, devices, and methods for eyebox expansion displays in wearable heads-up displays |
US10613331B2 (en) | 2015-02-17 | 2020-04-07 | North Inc. | Systems, devices, and methods for splitter optics in wearable heads-up displays |
US10031338B2 (en) | 2015-02-17 | 2018-07-24 | Thalmic Labs Inc. | Systems, devices, and methods for eyebox expansion in wearable heads-up displays |
US9958682B1 (en) | 2015-02-17 | 2018-05-01 | Thalmic Labs Inc. | Systems, devices, and methods for splitter optics in wearable heads-up displays |
US9989764B2 (en) | 2015-02-17 | 2018-06-05 | Thalmic Labs Inc. | Systems, devices, and methods for eyebox expansion in wearable heads-up displays |
US11574383B2 (en) | 2015-04-20 | 2023-02-07 | Intel Corporation | Apparatus and method for non-uniform frame buffer rasterization |
US11024007B2 (en) | 2015-04-20 | 2021-06-01 | Intel Corporation | Apparatus and method for non-uniform frame buffer rasterization |
US11748843B2 (en) | 2015-04-20 | 2023-09-05 | Intel Corporation | Apparatus and method for non-uniform frame buffer rasterization |
US11263725B2 (en) * | 2015-04-20 | 2022-03-01 | Intel Corporation | Apparatus and method for non-uniform frame buffer rasterization |
US10175488B2 (en) | 2015-05-04 | 2019-01-08 | North Inc. | Systems, devices, and methods for spatially-multiplexed holographic optical elements |
US10133075B2 (en) | 2015-05-04 | 2018-11-20 | Thalmic Labs Inc. | Systems, devices, and methods for angle- and wavelength-multiplexed holographic optical elements |
US10197805B2 (en) | 2015-05-04 | 2019-02-05 | North Inc. | Systems, devices, and methods for eyeboxes with heterogeneous exit pupils |
US10078219B2 (en) | 2015-05-28 | 2018-09-18 | Thalmic Labs Inc. | Wearable heads-up display with integrated eye tracker and different optical power holograms |
US10139633B2 (en) | 2015-05-28 | 2018-11-27 | Thalmic Labs Inc. | Eyebox expansion and exit pupil replication in wearable heads-up display having integrated eye tracking and laser projection |
US10488661B2 (en) | 2015-05-28 | 2019-11-26 | North Inc. | Systems, devices, and methods that integrate eye tracking and scanning laser projection in wearable heads-up displays |
US10073268B2 (en) | 2015-05-28 | 2018-09-11 | Thalmic Labs Inc. | Display with integrated visible light eye tracking |
US10114222B2 (en) | 2015-05-28 | 2018-10-30 | Thalmic Labs Inc. | Integrated eye tracking and laser projection methods with holographic elements of varying optical powers |
US10180578B2 (en) | 2015-05-28 | 2019-01-15 | North Inc. | Methods that integrate visible light eye tracking in scanning laser projection displays |
US10078220B2 (en) | 2015-05-28 | 2018-09-18 | Thalmic Labs Inc. | Wearable heads-up display with integrated eye tracker |
US20170098330A1 (en) * | 2015-07-14 | 2017-04-06 | Colopl, Inc. | Method for controlling head mounted display, and program for controlling head mounted display |
US10115235B2 (en) * | 2015-07-14 | 2018-10-30 | Colopl, Inc. | Method for controlling head mounted display, and system for implemeting the method |
US10338451B2 (en) | 2015-08-03 | 2019-07-02 | Facebook Technologies, Llc | Devices and methods for removing zeroth order leakage in beam steering devices |
US10534173B2 (en) | 2015-08-03 | 2020-01-14 | Facebook Technologies, Llc | Display with a tunable mask for augmented reality |
US10437061B2 (en) | 2015-08-03 | 2019-10-08 | Facebook Technologies, Llc | Near-ocular display based on hologram projection |
US10552676B2 (en) | 2015-08-03 | 2020-02-04 | Facebook Technologies, Llc | Methods and devices for eye tracking based on depth sensing |
US10459305B2 (en) | 2015-08-03 | 2019-10-29 | Facebook Technologies, Llc | Time-domain adjustment of phase retardation in a liquid crystal grating for a color display |
US10345599B2 (en) | 2015-08-03 | 2019-07-09 | Facebook Technologies, Llc | Tile array for near-ocular display |
US10162182B2 (en) * | 2015-08-03 | 2018-12-25 | Facebook Technologies, Llc | Enhanced pixel resolution through non-uniform ocular projection |
US10274730B2 (en) | 2015-08-03 | 2019-04-30 | Facebook Technologies, Llc | Display with an embedded eye tracker |
US10451876B2 (en) | 2015-08-03 | 2019-10-22 | Facebook Technologies, Llc | Enhanced visual perception through distance-based ocular projection |
US10297180B2 (en) | 2015-08-03 | 2019-05-21 | Facebook Technologies, Llc | Compensation of chromatic dispersion in a tunable beam steering device for improved display |
US10705342B2 (en) | 2015-09-04 | 2020-07-07 | North Inc. | Systems, articles, and methods for integrating holographic optical elements with eyeglass lenses |
US10488662B2 (en) | 2015-09-04 | 2019-11-26 | North Inc. | Systems, articles, and methods for integrating holographic optical elements with eyeglass lenses |
US10718945B2 (en) | 2015-09-04 | 2020-07-21 | North Inc. | Systems, articles, and methods for integrating holographic optical elements with eyeglass lenses |
US10877272B2 (en) | 2015-09-04 | 2020-12-29 | Google Llc | Systems, articles, and methods for integrating holographic optical elements with eyeglass lenses |
US10890765B2 (en) | 2015-09-04 | 2021-01-12 | Google Llc | Systems, articles, and methods for integrating holographic optical elements with eyeglass lenses |
US10656822B2 (en) | 2015-10-01 | 2020-05-19 | North Inc. | Systems, devices, and methods for interacting with content displayed on head-mounted displays |
US9904051B2 (en) | 2015-10-23 | 2018-02-27 | Thalmic Labs Inc. | Systems, devices, and methods for laser eye tracking |
US10606072B2 (en) | 2015-10-23 | 2020-03-31 | North Inc. | Systems, devices, and methods for laser eye tracking |
US10228558B2 (en) | 2015-10-23 | 2019-03-12 | North Inc. | Systems, devices, and methods for laser eye tracking |
US10705262B2 (en) | 2015-10-25 | 2020-07-07 | Facebook Technologies, Llc | Liquid crystal half-wave plate lens |
US10247858B2 (en) | 2015-10-25 | 2019-04-02 | Facebook Technologies, Llc | Liquid crystal half-wave plate lens |
US10416454B2 (en) | 2015-10-25 | 2019-09-17 | Facebook Technologies, Llc | Combination prism array for focusing light |
US11270506B2 (en) | 2015-10-29 | 2022-03-08 | Sony Computer Entertainment Inc. | Foveated geometry tessellation |
US20170124760A1 (en) * | 2015-10-29 | 2017-05-04 | Sony Computer Entertainment Inc. | Foveated geometry tessellation |
US10726619B2 (en) * | 2015-10-29 | 2020-07-28 | Sony Interactive Entertainment Inc. | Foveated geometry tessellation |
US10802190B2 (en) | 2015-12-17 | 2020-10-13 | Covestro Llc | Systems, devices, and methods for curved holographic optical elements |
US10670928B2 (en) | 2015-12-21 | 2020-06-02 | Facebook Technologies, Llc | Wide angle beam steering for virtual reality and augmented reality |
US10670929B2 (en) | 2015-12-21 | 2020-06-02 | Facebook Technologies, Llc | Enhanced spatial resolution using a segmented electrode array |
US10303246B2 (en) | 2016-01-20 | 2019-05-28 | North Inc. | Systems, devices, and methods for proximity-based eye tracking |
US10126815B2 (en) | 2016-01-20 | 2018-11-13 | Thalmic Labs Inc. | Systems, devices, and methods for proximity-based eye tracking |
US10241572B2 (en) | 2016-01-20 | 2019-03-26 | North Inc. | Systems, devices, and methods for proximity-based eye tracking |
US10451881B2 (en) | 2016-01-29 | 2019-10-22 | North Inc. | Systems, devices, and methods for preventing eyebox degradation in a wearable heads-up display |
US10437067B2 (en) | 2016-01-29 | 2019-10-08 | North Inc. | Systems, devices, and methods for preventing eyebox degradation in a wearable heads-up display |
US10151926B2 (en) | 2016-01-29 | 2018-12-11 | North Inc. | Systems, devices, and methods for preventing eyebox degradation in a wearable heads-up display |
US10365549B2 (en) | 2016-04-13 | 2019-07-30 | North Inc. | Systems, devices, and methods for focusing laser projectors |
US10365548B2 (en) | 2016-04-13 | 2019-07-30 | North Inc. | Systems, devices, and methods for focusing laser projectors |
US10365550B2 (en) | 2016-04-13 | 2019-07-30 | North Inc. | Systems, devices, and methods for focusing laser projectors |
US10684478B2 (en) | 2016-05-09 | 2020-06-16 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US11226691B2 (en) | 2016-05-09 | 2022-01-18 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US10824253B2 (en) | 2016-05-09 | 2020-11-03 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US11500212B2 (en) | 2016-05-09 | 2022-11-15 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US11320656B2 (en) | 2016-05-09 | 2022-05-03 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US20230161167A1 (en) * | 2016-06-01 | 2023-05-25 | Mentor Acquisition One, Llc | Modular systems for head-worn computers |
US11460708B2 (en) | 2016-06-01 | 2022-10-04 | Mentor Acquisition One, Llc | Modular systems for head-worn computers |
US11977238B2 (en) | 2016-06-01 | 2024-05-07 | Mentor Acquisition One, Llc | Modular systems for head-worn computers |
US11022808B2 (en) | 2016-06-01 | 2021-06-01 | Mentor Acquisition One, Llc | Modular systems for head-worn computers |
US11754845B2 (en) * | 2016-06-01 | 2023-09-12 | Mentor Acquisition One, Llc | Modular systems for head-worn computers |
US10466491B2 (en) | 2016-06-01 | 2019-11-05 | Mentor Acquisition One, Llc | Modular systems for head-worn computers |
US11586048B2 (en) | 2016-06-01 | 2023-02-21 | Mentor Acquisition One, Llc | Modular systems for head-worn computers |
US11032471B2 (en) * | 2016-06-30 | 2021-06-08 | Nokia Technologies Oy | Method and apparatus for providing a visual indication of a point of interest outside of a user's view |
US10230929B2 (en) | 2016-07-27 | 2019-03-12 | North Inc. | Systems, devices, and methods for laser projectors |
US10250856B2 (en) | 2016-07-27 | 2019-04-02 | North Inc. | Systems, devices, and methods for laser projectors |
US10277874B2 (en) | 2016-07-27 | 2019-04-30 | North Inc. | Systems, devices, and methods for laser projectors |
US10459221B2 (en) | 2016-08-12 | 2019-10-29 | North Inc. | Systems, devices, and methods for variable luminance in wearable heads-up displays |
US10459223B2 (en) | 2016-08-12 | 2019-10-29 | North Inc. | Systems, devices, and methods for variable luminance in wearable heads-up displays |
US10459222B2 (en) | 2016-08-12 | 2019-10-29 | North Inc. | Systems, devices, and methods for variable luminance in wearable heads-up displays |
US11604358B2 (en) | 2016-09-08 | 2023-03-14 | Mentor Acquisition One, Llc | Optical systems for head-worn computers |
US11366320B2 (en) | 2016-09-08 | 2022-06-21 | Mentor Acquisition One, Llc | Optical systems for head-worn computers |
US10534180B2 (en) | 2016-09-08 | 2020-01-14 | Mentor Acquisition One, Llc | Optical systems for head-worn computers |
US9910284B1 (en) | 2016-09-08 | 2018-03-06 | Osterhout Group, Inc. | Optical systems for head-worn computers |
US11714284B2 (en) * | 2016-09-20 | 2023-08-01 | Apple Inc. | Display device including foveal and peripheral projectors |
US10215987B2 (en) | 2016-11-10 | 2019-02-26 | North Inc. | Systems, devices, and methods for astigmatism compensation in a wearable heads-up display |
US10345596B2 (en) | 2016-11-10 | 2019-07-09 | North Inc. | Systems, devices, and methods for astigmatism compensation in a wearable heads-up display |
US10409057B2 (en) | 2016-11-30 | 2019-09-10 | North Inc. | Systems, devices, and methods for laser eye tracking in wearable heads-up displays |
US10459220B2 (en) | 2016-11-30 | 2019-10-29 | North Inc. | Systems, devices, and methods for laser eye tracking in wearable heads-up displays |
US10365492B2 (en) | 2016-12-23 | 2019-07-30 | North Inc. | Systems, devices, and methods for beam combining in wearable heads-up displays |
US10663732B2 (en) | 2016-12-23 | 2020-05-26 | North Inc. | Systems, devices, and methods for beam combining in wearable heads-up displays |
USD947186S1 (en) * | 2017-01-04 | 2022-03-29 | Mentor Acquisition One, Llc | Computer glasses |
USD864959S1 (en) * | 2017-01-04 | 2019-10-29 | Mentor Acquisition One, Llc | Computer glasses |
USD918905S1 (en) | 2017-01-04 | 2021-05-11 | Mentor Acquisition One, Llc | Computer glasses |
US20180196512A1 (en) * | 2017-01-10 | 2018-07-12 | Samsung Electronics Co., Ltd. | Method for outputting image and electronic device supporting the same |
US10331208B2 (en) * | 2017-01-10 | 2019-06-25 | Samsung Electronics Co., Ltd. | Method for outputting image and electronic device supporting the same |
US10718951B2 (en) | 2017-01-25 | 2020-07-21 | North Inc. | Systems, devices, and methods for beam combining in laser projectors |
US10437073B2 (en) | 2017-01-25 | 2019-10-08 | North Inc. | Systems, devices, and methods for beam combining in laser projectors |
US10437074B2 (en) | 2017-01-25 | 2019-10-08 | North Inc. | Systems, devices, and methods for beam combining in laser projectors |
JP7075939B2 (en) | 2017-03-03 | 2022-05-26 | マイクロソフト テクノロジー ライセンシング,エルエルシー | MEMS scanning display device |
JP2020510873A (en) * | 2017-03-03 | 2020-04-09 | マイクロソフト テクノロジー ライセンシング,エルエルシー | MEMS scanning display device |
WO2018160507A1 (en) * | 2017-03-03 | 2018-09-07 | Microsoft Technology Licensing, Llc | Mems scanning display device |
CN110352374A (en) * | 2017-03-03 | 2019-10-18 | 微软技术许可有限责任公司 | MEMS scanning display equipment |
RU2770138C2 (en) * | 2017-03-03 | 2022-04-14 | МАЙКРОСОФТ ТЕКНОЛОДЖИ ЛАЙСЕНСИНГ, ЭлЭлСи | Scanning mems-display device |
IL268454B (en) * | 2017-03-03 | 2022-10-01 | Microsoft Technology Licensing Llc | Mems scanning display device |
IL268453B2 (en) * | 2017-03-03 | 2023-09-01 | Microsoft Technology Licensing Llc | Mems scanning display device |
IL268453B1 (en) * | 2017-03-03 | 2023-05-01 | Microsoft Technology Licensing Llc | Mems scanning display device |
JP2020510870A (en) * | 2017-03-03 | 2020-04-09 | マイクロソフト テクノロジー ライセンシング,エルエルシー | MEMS scanning display device |
CN110352375A (en) * | 2017-03-03 | 2019-10-18 | 微软技术许可有限责任公司 | MEMS scanning display equipment |
AU2018264863B2 (en) * | 2017-03-03 | 2022-07-21 | Microsoft Technology Licensing, Llc | MEMS scanning display device |
US10365709B2 (en) | 2017-03-03 | 2019-07-30 | Microsoft Technology Licensing, Llc | MEMS scanning display device |
US10317670B2 (en) | 2017-03-03 | 2019-06-11 | Microsoft Technology Licensing, Llc | MEMS scanning display device |
WO2018208361A3 (en) * | 2017-03-03 | 2018-12-20 | Microsoft Technology Licensing, Llc | Mems scanning display device |
AU2018227679B2 (en) * | 2017-03-03 | 2022-05-26 | Microsoft Technology Licensing, Llc | MEMS scanning display device |
KR20190118670A (en) * | 2017-03-03 | 2019-10-18 | 마이크로소프트 테크놀로지 라이센싱, 엘엘씨 | MEMS scanning display device |
KR102582491B1 (en) * | 2017-03-03 | 2023-09-22 | 마이크로소프트 테크놀로지 라이센싱, 엘엘씨 | MEMS scanning display device |
IL268454B2 (en) * | 2017-03-03 | 2023-02-01 | Microsoft Technology Licensing Llc | Mems scanning display device |
JP7075940B2 (en) | 2017-03-03 | 2022-05-26 | マイクロソフト テクノロジー ライセンシング,エルエルシー | MEMS scanning display device |
US11656468B2 (en) | 2017-03-27 | 2023-05-23 | Avegant Corp. | Steerable high-resolution display having a foveal display and a field display with intermediate optics |
WO2018183405A1 (en) * | 2017-03-27 | 2018-10-04 | Avegant Corp. | Steerable foveal display |
CN110520781A (en) * | 2017-03-27 | 2019-11-29 | 阿维甘特公司 | Central fovea display can be turned to |
US10514546B2 (en) | 2017-03-27 | 2019-12-24 | Avegant Corp. | Steerable high-resolution display |
US11163164B2 (en) | 2017-03-27 | 2021-11-02 | Avegant Corp. | Steerable high-resolution display |
US11776196B2 (en) | 2017-04-01 | 2023-10-03 | Intel Corporation | Temporal data structures in a ray tracing architecture |
US11398069B2 (en) * | 2017-04-01 | 2022-07-26 | Intel Corporation | Temporal data structures in a ray tracing architecture |
CN110506230A (en) * | 2017-04-03 | 2019-11-26 | 微软技术许可有限责任公司 | Wide visual field scans display |
US10417975B2 (en) * | 2017-04-03 | 2019-09-17 | Microsoft Technology Licensing, Llc | Wide field of view scanning display |
WO2018187096A1 (en) * | 2017-04-03 | 2018-10-11 | Microsoft Technology Licensing, Llc | Wide field of view scanning display |
US10499021B2 (en) * | 2017-04-11 | 2019-12-03 | Microsoft Technology Licensing, Llc | Foveated MEMS scanning display |
CN108737724A (en) * | 2017-04-17 | 2018-11-02 | 英特尔公司 | The system and method for capturing and showing for 360 videos |
US11164352B2 (en) * | 2017-04-21 | 2021-11-02 | Intel Corporation | Low power foveated rendering to save power on GPU and/or display |
US11587273B2 (en) | 2017-04-21 | 2023-02-21 | Intel Corporation | Low power foveated rendering to save power on GPU and/or display |
US11131861B2 (en) | 2017-05-29 | 2021-09-28 | Eyeway Vision Ltd | Image projection system |
EP3631558A4 (en) * | 2017-05-29 | 2020-12-30 | Eyeway Vision Ltd. | Image projection system |
JP2020021083A (en) * | 2017-06-30 | 2020-02-06 | エルジー ディスプレイ カンパニー リミテッド | Display device and gate driver circuit of the same |
JP2019012258A (en) * | 2017-06-30 | 2019-01-24 | エルジー ディスプレイ カンパニー リミテッド | Display device and gate driving circuit thereof |
US10504442B2 (en) | 2017-06-30 | 2019-12-10 | Lg Display Co., Ltd. | Display device and gate driving circuit thereof, control method and virtual reality device |
US11550157B2 (en) | 2017-07-24 | 2023-01-10 | Mentor Acquisition One, Llc | See-through computer display systems |
US11789269B2 (en) | 2017-07-24 | 2023-10-17 | Mentor Acquisition One, Llc | See-through computer display systems |
US11668939B2 (en) | 2017-07-24 | 2023-06-06 | Mentor Acquisition One, Llc | See-through computer display systems with stray light management |
US11226489B2 (en) | 2017-07-24 | 2022-01-18 | Mentor Acquisition One, Llc | See-through computer display systems with stray light management |
US11567328B2 (en) | 2017-07-24 | 2023-01-31 | Mentor Acquisition One, Llc | See-through computer display systems with adjustable zoom cameras |
US11960095B2 (en) | 2017-07-24 | 2024-04-16 | Mentor Acquisition One, Llc | See-through computer display systems |
US11971554B2 (en) | 2017-07-24 | 2024-04-30 | Mentor Acquisition One, Llc | See-through computer display systems with stray light management |
US10578869B2 (en) * | 2017-07-24 | 2020-03-03 | Mentor Acquisition One, Llc | See-through computer display systems with adjustable zoom cameras |
US11409105B2 (en) * | 2017-07-24 | 2022-08-09 | Mentor Acquisition One, Llc | See-through computer display systems |
US10422995B2 (en) | 2017-07-24 | 2019-09-24 | Mentor Acquisition One, Llc | See-through computer display systems with stray light management |
US11042035B2 (en) | 2017-07-24 | 2021-06-22 | Mentor Acquisition One, Llc | See-through computer display systems with adjustable zoom cameras |
US20190033595A1 (en) * | 2017-07-26 | 2019-01-31 | Thalmic Labs Inc. | Systems, devices, and methods for narrow waveband laser diodes |
US10971889B2 (en) * | 2017-07-26 | 2021-04-06 | Google Llc | Systems, devices, and methods for narrow waveband laser diodes |
US11500207B2 (en) | 2017-08-04 | 2022-11-15 | Mentor Acquisition One, Llc | Image expansion optic for head-worn computer |
US10969584B2 (en) | 2017-08-04 | 2021-04-06 | Mentor Acquisition One, Llc | Image expansion optic for head-worn computer |
US11947120B2 (en) | 2017-08-04 | 2024-04-02 | Mentor Acquisition One, Llc | Image expansion optic for head-worn computer |
US11122256B1 (en) * | 2017-08-07 | 2021-09-14 | Apple Inc. | Mixed reality system |
US11695913B1 (en) | 2017-08-07 | 2023-07-04 | Apple, Inc. | Mixed reality system |
EP3687465A4 (en) * | 2017-09-27 | 2021-06-16 | University of Miami | Digital therapeutic corrective spectacles |
US11635736B2 (en) | 2017-10-19 | 2023-04-25 | Meta Platforms Technologies, Llc | Systems and methods for identifying biological structures associated with neuromuscular source signals |
US11300788B2 (en) | 2017-10-23 | 2022-04-12 | Google Llc | Free space multiple laser diode modules |
US10901216B2 (en) | 2017-10-23 | 2021-01-26 | Google Llc | Free space multiple laser diode modules |
USD906404S1 (en) | 2017-12-07 | 2020-12-29 | Amazon Technologies, Inc. | Wearable device |
USD877237S1 (en) * | 2017-12-07 | 2020-03-03 | Amazon Technologies, Inc. | Wearable device |
US10613323B1 (en) * | 2017-12-13 | 2020-04-07 | Facebook Technologies, Llc | Transition feature for framing multizone optics |
US10656426B2 (en) * | 2017-12-22 | 2020-05-19 | North Inc. | Grating waveguide combiner for optical engine |
US20190196206A1 (en) * | 2017-12-22 | 2019-06-27 | North Inc. | Grating waveguide combiner for optical engine |
US20190272801A1 (en) * | 2018-03-01 | 2019-09-05 | Beijing Boe Optoelectronics Technology Co., Ltd. | Processing method and processing device for display data, and display device |
US10706814B2 (en) * | 2018-03-01 | 2020-07-07 | Beijing Boe Optoelectronics Technology Co., Ltd. | Processing method and processing device for display data, and display device |
US10614734B2 (en) | 2018-03-06 | 2020-04-07 | Varjo Technologies Oy | Display apparatus and method of displaying using controllable scanning mirror |
WO2019170959A1 (en) * | 2018-03-06 | 2019-09-12 | Varjo Technologies Oy | Display apparatus and method of displaying using controllable scanning mirror |
CN112105982A (en) * | 2018-05-04 | 2020-12-18 | 哈曼国际工业有限公司 | Head-up display without mirror |
DE102018209886B4 (en) * | 2018-06-19 | 2020-02-27 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Device for projecting a laser beam to generate an image on the retina of an eye and glasses device with two such devices |
DE102018209886A1 (en) * | 2018-06-19 | 2019-12-19 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Device for projecting a laser beam to generate an image on the retina of an eye and glasses device with two such devices |
WO2019243395A1 (en) | 2018-06-19 | 2019-12-26 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Device apparatus for projecting a laser beam for generating an image on the retina of an eye |
US11867916B2 (en) | 2018-06-19 | 2024-01-09 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Device apparatus for projecting a laser beam for generating an image on the retina of an eye |
US10474296B1 (en) * | 2018-07-12 | 2019-11-12 | Microvision, Inc. | Laser scanning devices and methods with touch detection |
US10719127B1 (en) * | 2018-08-29 | 2020-07-21 | Rockwell Collins, Inc. | Extended life display by utilizing eye tracking |
CN109189357A (en) * | 2018-08-30 | 2019-01-11 | Oppo广东移动通信有限公司 | Information display method, device, intelligent glasses and storage medium |
US11941176B1 (en) | 2018-11-27 | 2024-03-26 | Meta Platforms Technologies, Llc | Methods and apparatus for autocalibration of a wearable electrode sensor system |
US11797087B2 (en) | 2018-11-27 | 2023-10-24 | Meta Platforms Technologies, Llc | Methods and apparatus for autocalibration of a wearable electrode sensor system |
US11927762B2 (en) | 2018-12-07 | 2024-03-12 | Avegant Corp. | Steerable positioning element |
US10777087B2 (en) * | 2018-12-07 | 2020-09-15 | International Business Machines Corporation | Augmented reality for removing external stimuli |
US11169383B2 (en) | 2018-12-07 | 2021-11-09 | Avegant Corp. | Steerable positioning element |
US11126261B2 (en) | 2019-01-07 | 2021-09-21 | Avegant Corp. | Display control system and rendering pipeline |
US11650663B2 (en) | 2019-01-07 | 2023-05-16 | Avegant Corp. | Repositionable foveal display with a fast shut-off logic |
WO2020190519A1 (en) * | 2019-03-15 | 2020-09-24 | Microsoft Technology Licensing, Llc | Holographic image generated based on eye position |
CN113574471A (en) * | 2019-03-15 | 2021-10-29 | 微软技术许可有限责任公司 | Holographic image generated based on eye position |
US11068052B2 (en) | 2019-03-15 | 2021-07-20 | Microsoft Technology Licensing, Llc | Holographic image generated based on eye position |
US11961494B1 (en) | 2019-03-29 | 2024-04-16 | Meta Platforms Technologies, Llc | Electromagnetic interference reduction in extended reality environments |
US11586049B2 (en) | 2019-03-29 | 2023-02-21 | Avegant Corp. | Steerable hybrid display using a waveguide |
US11907423B2 (en) | 2019-11-25 | 2024-02-20 | Meta Platforms Technologies, Llc | Systems and methods for contextualized interactions with an environment |
US11624921B2 (en) | 2020-01-06 | 2023-04-11 | Avegant Corp. | Head mounted system with color specific modulation |
EP4121813A4 (en) * | 2020-03-20 | 2024-01-17 | Magic Leap Inc | Systems and methods for retinal imaging and tracking |
US11435578B2 (en) | 2020-06-02 | 2022-09-06 | Trumpf Photonic Components Gmbh | Method for detecting a gaze direction of an eye |
US11513594B2 (en) * | 2020-06-02 | 2022-11-29 | Trumpf Photonic Components Gmbh | Method for operating a pair of smart glasses |
DE102020206821A1 (en) | 2020-06-02 | 2021-12-02 | Robert Bosch Gesellschaft mit beschränkter Haftung | Method for determining a line of sight of an eye |
DE102020206822A1 (en) | 2020-06-02 | 2021-12-02 | Robert Bosch Gesellschaft mit beschränkter Haftung | Procedure for operating data glasses |
USD946572S1 (en) * | 2020-12-02 | 2022-03-22 | Eissa Nick Asemani | Visor mounted eyewear |
USD985563S1 (en) * | 2021-01-06 | 2023-05-09 | Samsung Electronics Co., Ltd. | Head-mounted display |
USD985564S1 (en) * | 2021-01-06 | 2023-05-09 | Samsung Electronics Co., Ltd. | Head-mounted display |
USD985562S1 (en) * | 2021-01-21 | 2023-05-09 | Samsung Electronics Co., Ltd. | Head-mounted display |
US11868531B1 (en) | 2021-04-08 | 2024-01-09 | Meta Platforms Technologies, Llc | Wearable device providing for thumb-to-finger-based input gestures detected based on neuromuscular signals, and systems and methods of use thereof |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20160274365A1 (en) | Systems, devices, and methods for wearable heads-up displays with heterogeneous display quality | |
US10606072B2 (en) | Systems, devices, and methods for laser eye tracking | |
US10459220B2 (en) | Systems, devices, and methods for laser eye tracking in wearable heads-up displays | |
US10114222B2 (en) | Integrated eye tracking and laser projection methods with holographic elements of varying optical powers | |
US10197805B2 (en) | Systems, devices, and methods for eyeboxes with heterogeneous exit pupils | |
US20180103194A1 (en) | Image capture systems, devices, and methods that autofocus based on eye-tracking | |
US10852817B1 (en) | Eye tracking combiner having multiple perspectives | |
US20180299680A1 (en) | Systems, devices, and methods for eyebox expansion in wearable heads-up displays | |
US9285872B1 (en) | Using head gesture and eye position to wake a head mounted device | |
US11314323B2 (en) | Position tracking system for head-mounted displays that includes sensor integrated circuits | |
US10725302B1 (en) | Stereo imaging with Fresnel facets and Fresnel reflections | |
US9934583B2 (en) | Expectation maximization to determine position of ambient glints | |
US20240012246A1 (en) | Methods, Apparatuses And Computer Program Products For Providing An Eye Tracking System Based On Flexible Around The Lens Or Frame Illumination Sources |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: THALMIC LABS INC., CANADA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BAILEY, MATTHEW;ALEXANDER, STEFAN;SIGNING DATES FROM 20161214 TO 20170123;REEL/FRAME:044900/0433 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |
|
AS | Assignment |
Owner name: GOOGLE LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NORTH INC.;REEL/FRAME:054113/0907 Effective date: 20200916 |
|
AS | Assignment |
Owner name: NORTH INC., CANADA Free format text: CHANGE OF NAME;ASSIGNOR:THALMIC LABS INC.;REEL/FRAME:054414/0100 Effective date: 20180830 |