WO2016105281A1 - Dispositif d'affichage proche de l'oeil - Google Patents

Dispositif d'affichage proche de l'oeil Download PDF

Info

Publication number
WO2016105281A1
WO2016105281A1 PCT/TR2014/000512 TR2014000512W WO2016105281A1 WO 2016105281 A1 WO2016105281 A1 WO 2016105281A1 TR 2014000512 W TR2014000512 W TR 2014000512W WO 2016105281 A1 WO2016105281 A1 WO 2016105281A1
Authority
WO
WIPO (PCT)
Prior art keywords
slm
display device
light
eye display
user
Prior art date
Application number
PCT/TR2014/000512
Other languages
English (en)
Inventor
Hakan Urey
Erdem ULUSOY
Original Assignee
Koc University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koc University filed Critical Koc University
Priority to PCT/TR2014/000512 priority Critical patent/WO2016105281A1/fr
Priority to EP14850109.1A priority patent/EP3237957A1/fr
Priority to CN201480084628.2A priority patent/CN107438796B/zh
Publication of WO2016105281A1 publication Critical patent/WO2016105281A1/fr
Priority to US15/632,164 priority patent/US10571696B2/en
Priority to US15/633,721 priority patent/US10444507B2/en
Priority to US15/633,753 priority patent/US10444508B2/en

Links

Classifications

    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H1/00Holographic processes or apparatus using light, infrared or ultraviolet waves for obtaining holograms or for obtaining an image from them; Details peculiar thereto
    • G03H1/22Processes or apparatus for obtaining an optical image from holograms
    • G03H1/2294Addressing the hologram to an active spatial light modulator
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • G02B27/0172Head mounted characterised by optical features
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/0093Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H1/00Holographic processes or apparatus using light, infrared or ultraviolet waves for obtaining holograms or for obtaining an image from them; Details peculiar thereto
    • G03H1/04Processes or apparatus for producing holograms
    • G03H1/08Synthesising holograms, i.e. holograms synthesized from objects or objects from holograms
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B27/0103Head-up displays characterised by optical features comprising holographic elements
    • G02B2027/0105Holograms with particular structures
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • G02B2027/0178Eyeglass type
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0179Display position adjusting means not related to the information to be displayed
    • G02B2027/0187Display position adjusting means not related to the information to be displayed slaved to motion of at least a part of the body of the user, e.g. head, eye
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H1/00Holographic processes or apparatus using light, infrared or ultraviolet waves for obtaining holograms or for obtaining an image from them; Details peculiar thereto
    • G03H1/22Processes or apparatus for obtaining an optical image from holograms
    • G03H1/2202Reconstruction geometries or arrangements
    • G03H2001/2223Particular relationship between light source, hologram and observer
    • G03H2001/2231Reflection reconstruction
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H1/00Holographic processes or apparatus using light, infrared or ultraviolet waves for obtaining holograms or for obtaining an image from them; Details peculiar thereto
    • G03H1/22Processes or apparatus for obtaining an optical image from holograms
    • G03H1/2202Reconstruction geometries or arrangements
    • G03H2001/2223Particular relationship between light source, hologram and observer
    • G03H2001/2234Transmission reconstruction
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H1/00Holographic processes or apparatus using light, infrared or ultraviolet waves for obtaining holograms or for obtaining an image from them; Details peculiar thereto
    • G03H1/22Processes or apparatus for obtaining an optical image from holograms
    • G03H1/2249Holobject properties
    • G03H2001/2252Location of the holobject
    • G03H2001/226Virtual or real
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H2225/00Active addressable light modulator
    • G03H2225/60Multiple SLMs
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H2227/00Mechanical components or mechanical aspects not otherwise provided for
    • G03H2227/02Handheld portable device, e.g. holographic camera, mobile holographic display
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H2227/00Mechanical components or mechanical aspects not otherwise provided for
    • G03H2227/05Support holding the holographic record
    • G03H2227/06Support including light source
    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03HHOLOGRAPHIC PROCESSES OR APPARATUS
    • G03H2270/00Substrate bearing the hologram
    • G03H2270/55Substrate bearing the hologram being an optical element, e.g. spectacles

Definitions

  • the present invention relates generally to optical systems, and more specifically to near-to-eye display devices.
  • Head-worn displays typically employ a microdisplay on which a two dimensional (2D) regular image is displayed. Since the physical distance between the microdisplay and the eye is typically much smaller than 25 cm (the closest distance at which the human eye can normally focus), a blurred image forms on the retina unless relay optics are placed in between.
  • the relay optics typically consist of several lenses which serve to form a magnified virtual image of the microdisplay beyond 25 cm (mostly at infinity) on which the eye can then focus and form a sharp retinal image.
  • Lightweight HWD designs that employ microdisplays are mostly restricted to systems having small fields of view (FOV), since weight and bulk increase for large FOV designs due to additional components inserted to compensate for aberrations.
  • FOV small fields of view
  • the recently emerging Google Glass which has a quite thin form factor
  • high end Military-type displays may support a FOV approaching 150 degrees or more, but weigh as much as 5 kg or more and may contain more than 10 different lenses, most of which are present to compensate for aberrations that emerge due to the enlarged FOV.
  • Microdisplay based HWD designs also fall short of providing the ultimate three dimensional (3D) visual experience. These HWD designs typically provide only stereoscopic images, which invoke 3D perception essentially only through binocular disparity. Monocular cues, especially accommodation, are typically not supported, or are incorrect. Users of stereoscopic systems typically suffer from visual fatigue caused by the so called accommodation-convergence conflict, in which eyes converge truly to the apparent position of a 3D object while accommodation is set incorrectly to the screen so as to make retinal images sharp. The fatigue is especially severe when virtual objects are closer than 50 cm.
  • Figure 1 shows a perspective view of a near-to-eye display device
  • Figure 2 shows a top view of the near-to-eye display device of Figure 1;
  • Figure 3 shows a handheld near-to-eye display device
  • Figure 4 shows a cross section of a spatial light modulator (SLM) being illuminated and generating a virtual scene wave;
  • SLM spatial light modulator
  • Figure 5 shows the cross section of Figure 4 depicting the virtual scene as seen by a user
  • Figure 6 shows a spatial light modulator with a pixelated structure
  • Figure 7 shows a cross section of an SLM that generates noise beams and multiple diffraction orders
  • Figure 8 shows the cross section of Figure 7 with a user's eye pupil filtering out unwanted noise beams and diffraction orders
  • Figures 9, 10, and 11 show multiple diffraction orders on an exit pupil plane with a useful portion
  • Figure 12 shows an optical architecture in which the SLM is placed on a converging beam path
  • Figure 13 shows an optical architecture in which the SLM is illuminated by a diverging wavefront
  • Figure 14 shows an optical architecture with a point light source and SLM, with no other components with refractive power
  • Figure 15 shows an optical architecture in which an SLM is illuminated in a time sequential manner by an array of point light sources;
  • Figure 16 shows an optical architecture with multiple light sources and apertures to the associated emission cones;
  • Figure 17 shows an optical architecture in which a reflective SLM is placed directly in front of the user's eye
  • Figures 18, 19, and 20 show optical architectures in which real world vision is not blocked by the SLM
  • Figure 21 shows an optical architecture in which a reflective SLM is placed in front of the user's eye
  • Figures 22-28 show optical architectures in which real world vision is not blocked by the SLM
  • Figure 29 shows an optical architecture in which an SLM and reflector are combined
  • Figure 30 shows a reflector based solution for modulation in a single direction
  • Figure 31 shows a virtual reality (VR) architecture with SLM tiling
  • Figure 32 shows a VR architecture with SLM tiling
  • Figure 33 shows a VR display architecture
  • Figure 34 shows two tiled SLMs to double resolution
  • Figure 35 shows a near-to-eye display device with a high resolution portion and a low resolution portion
  • Figure 36 shows a high resolution image for foveal vision and lower resolution for peripheral vision
  • Figure 37 shows the high resolution image being steered to a user's pupil position
  • Figures 38 and 39 show a display system with a rotating hologram module to create a steerable high resolution image
  • Figure 40 shows a portion of a near-to-eye display device having a moving platform
  • Figure 41 shows a moving platform upon which an SLM is mounted
  • Figure 42 shows a portion of a near-to-eye display device having a moving platform with an array of bars;
  • Figure 43 shows a moving platform having an array of bars upon which SLMs and microdisplays are mounted;
  • Figure 44 shows a moving platform that moves in two dimensions to increase resolution;
  • Figure 45 shows a near-to-eye display device with a transducer to interact with a user for calibration
  • Figure 46 shows a flowchart of calibration methods in accordance with various embodiments of the invention.
  • Figure 47 shows example images shown to a user during calibration
  • Figure 48 shows a flowchart of calibration methods in accordance with various embodiments of the invention.
  • Figure 49 shows example images shown to a user during calibration
  • Figure 50 shows a flowchart of calibration methods in accordance with various embodiments of the invention.
  • Figure 51 shows example images shown to a user during calibration
  • Figure 52 shows a flowchart of calibration methods in accordance with various embodiments of the invention.
  • Figure 53 shows a near-to-eye display device with actuators for calibration
  • Figure 54 shows images of a user's eyes used for calibration
  • Figure 55 shows the near-to-eye display device of Figure 53 with actuation for calibration
  • Figure 56 shows a flowchart representing computation of SLM data
  • Figure 57-64 show a number of space-angle (or space-frequency) distributions that illustrate the basics of the computation procedure;
  • Figure 65 shows fast hologram computation
  • Figure 66 illustrates the fundamentals of the method for delivering speckle free images to the retina of a user
  • Figure 67 shows a perspective drawing of a back light unit that generates a two dimensional converging beam
  • Figure 68 shows a cross section of the back light unit of Figure 67 showing a scattering point and linearly arranged micromirror array
  • Figure 69 shows a cross section of the back light unit of Figure 67 showing a light scattering apparatus and a reflective optical element arranged as a Fresnel mirror
  • Figure 70 shows a cross section of the back light unit of Figure 67 showing a light scattering apparatus and a reflective optical element arranged as a free form concave reflector
  • Figure 71 shows a cross section of the back light unit of Figure 67 showing a scattering point and nonlinearly arranged micromirror array
  • Figure 72 shows a back light unit with an external light source
  • Figure 73 shows a cross section of a back light unit with transmissive
  • Figure 74 shows a cross section of a back light unit with a reflective SLM
  • Figure 75 shows a cross section of back light unit with cross polarizers
  • Figure 76 shows a cross section of back light unit with a mirror
  • Figure 77 shows a cross section of a back light unit with a fiber
  • Figure 78 shows a perspective view of a back light unit that generates a one dimensional converging beam
  • Figure 79 shows a perspective view of a back light unit that generates a collimated beam
  • Figure 80 shows a perspective view of a back light unit that generates a diverging beam
  • Figure 81 shows a cross section of a slab waveguide, a wedge, and a component with a micromirror array
  • Figure 82 shows a top view of the apparatus of Figure 81 ;
  • Figure 83 shows a cross section of a slab, wedge, component with micromirror array, and SLM positioned along the slab;
  • Figure 84 shows a cross section of a slab waveguide, a wedge, a component with a micromirror array, and an SLM between the wedge and the component with the micromirror array;
  • Figure 85 shows a cross section of slab waveguide, wedge, component with a micromirror array, and an SLM below the wedge;
  • Figure 86 shows a cross section of a slab waveguide, wedge, component with micromirror array, and an SLM at entrance to the slab;
  • Figure 87 shows a cross section of a slab waveguide, wedge, compensating wedge with micromirror array, and SLM below the wedge
  • Figure 88 shows a cross section of a slab waveguide with a 90 degree bend, wedge, optical component with a micromirror array, and an SLM
  • Figure 89 shows a cross section of a slab waveguide, wedge, and camera for eye tracking
  • Figure 90 shows a near-to-eye display device with a slab waveguide, wedge, component with micromirror array, SLM, and camera for eye tracking;
  • Figure 91 shows a slab waveguide, a curved wedge, and a compensation plate
  • Figure 92 shows a slab waveguide, curved wedge, and SLM in a converging beam
  • Figure 93 shows a slab waveguide, curved wedge, and SLM on top of the slab
  • Figure 94 shows a slab waveguide, curved wedge, and SLM at the entrance to the slab waveguide
  • Figure 95 shows a slab waveguide, curved wedge, and camera for eye tracking
  • Figure 96 shows a perspective view of the apparatus of Figure 91
  • Figure 97 shows a near-to-eye display device with a slab waveguide, curved wedge, SLM, and camera for eye tracking;
  • Figure 98 shows a near-to-eye display device with a moving platform assembly
  • Figure 99 shows a cross section of a moving platform assembly
  • Figure 100 shows a perspective view of a moving platform assembly
  • Figure 101 shows a side view of contact lens placed on an eye
  • Figure 102 shows a front view of the contact lens of Figure 101
  • Figure 103 shows a cross section of a contact lens on an eye and a moving platform assembly
  • Figure 104 shows a near-to-eye display device with a moving platform assembly
  • Figure 105 shows a perspective view of a near-to-eye display device with a rotating bar
  • Figures 106-108 show front views of near-to-eye display devices with rotating bars;
  • Figures 109 and 110 show rotating bar actuation embodiments;
  • Figure 111 shows a front view of a near-to-eye display device with a platform that moves in two dimensions
  • Figure 112 shows an external display with no contact lens
  • Figure 113 shows a perspective view of near-to-eye display device that includes a LED array.
  • Figure 114 shows a two dimensional LED array
  • Figures 115 and 116 show a top view of pupil tracking using multiple
  • Figure 117 shows a near-to-eye display device that includes a rotating
  • Figures 118 and 119 show a top view of pupil tracking using a rotating
  • Figure 120 shows a perspective view of a near-to-eye display device that includes rotating SLMs and LED arrays.
  • Figure 121 shows a flowchart showing rotation for small angles and LED selection for larger angles
  • Figure 122 shows a flowchart showing rotation for small angles and diffraction order selection for larger angles
  • Figure 123 shows a near-to-eye display device that includes active grating
  • Figures 124 and 125 show top views of pupil tracking using an SLM and an active grating
  • Figure 126 shows a near-to-eye display device with a combination of an active grating and an LED array
  • Figure 127 shows a flowchart showing grating actuation for small angles and LED selection for larger angles
  • Figure 128 shows a flowchart showing grating actuation for small angles and diffraction order selection for larger angles
  • Figures 129 and 130 show augmented reality views demonstrating a virtual scene at different depths
  • Figure 131 shows a block diagram of a near-to-eye display device in accordance with various embodiments of the present invention.
  • Figure 132 shows a near-to-eye display device with transparent touch sensitive layers.
  • Figure 1 shows a perspective view of a near-to-eye display device.
  • Near- to-eye display device 100 includes a frame 102 in the shape of an eyeglass frame.
  • Near-to-eye display device 100 also includes spatial light modulators (SLM) 110, point light source 120, electronics module 160, and cable 170.
  • SLM spatial light modulators
  • the near-to-eye display device is a head- worn device (HWD), although this is not a limitation of the present invention.
  • HWD head- worn device
  • near-to-eye display devices are not head-worn.
  • Various examples of non-head-worn near-to-eye display devices are discussed further below.
  • Point light source 120 may include any type of light source.
  • point light source 120 includes a laser light source.
  • point light source 120 includes an ultraviolet (UV) light source, an infrared (IR) light source, or other source of visible or nonvisible light.
  • UV ultraviolet
  • IR infrared
  • near-to-eye display device 100 displays a monochrome or full color video of a 2D or 3D virtual scene 150 that appears to be located on the outer side of the eyeglass to the user.
  • point light source 120 For each frame of the displayed video, point light source 120 generates a coherent light wave of a single wavelength that illuminates a spatial light modulator (SLM 110) that is mounted on the front section of the eyeglass.
  • SLM spatial light modulator
  • the SLM spatially modulates the phase and/or amplitude of the incident wave and reflects it towards the eye of the user, shown generally at 130.
  • near-to-eye display device 100 is a monochrome display device, and point light source 120 only generates a coherent light wave of a single color.
  • near-to-eye display device 100 is a full color display device, and point light source 120 generates coherent light waves of different wavelengths in a time sequential manner.
  • the data on the SLM is a computer generated holographic image of the virtual scene.
  • the data on the SLM is computed and fed by a computer unit, which can be mounted on frame 102 as electronics module 160, or can be connected to the display device by cable 170 or wireless links (not shown).
  • Electronics module 160 may include any suitable components.
  • electronics module 160 includes driver circuits to drive point light source 120, and digital processing components to store SLM data and to drive the SLMs 110 with that data.
  • electronics module 160 may include a processor and memory, or any other suitable electronics components.
  • SLM data is computed real-time as it is displayed.
  • electronics module 160 computes the SLM data and drives SLMs 110 with the SLM data to create virtual scene 150 in real-time.
  • the real-time SLM data may be a function of head tracking data, pupil tracking data, environmental data (e.g., ambient light, objects in the user's field of view, etc.).
  • SLM data is precomputed and stored for retrieval at display time.
  • SLM data for an entire virtual environment may be precomputed and stored.
  • the appropriate SLM data is retrieved and displayed.
  • portions of the SLM data are precomputed, and portions of the SLM data are computed real-time.
  • Point light source 120 is shown on an outer portion of frame 102, and
  • SLM 110 is reflective.
  • point light source is located differently, and the SLM is transmissive.
  • point light source 120 is mounted on the nose bridge between the two SLMs, and the light generated by point light source 120 is fed to a back light unit that illuminates the SLM from the opposite side than shown in Figure 1.
  • Figure 2 shows a top view of the near-to-eye display device of Figure 1.
  • the wave 130 reflected by SLM 110 propagates towards the user's eye 210 and forms a light wave distribution on the exit pupil plane 220, which is defined as the plane that lies just in front of the user's eye, and corresponds to the expected location of the user's eye pupil 212.
  • Part of the light wave distribution formed on the exit pupil plane is intercepted by the user's eye pupil 212 and propagates to the retina, where a 3D image of the virtual scene is formed.
  • a real world view is superimposed on the virtual scene, and in other embodiments, the real world view is blocked, and the only image formed on the retina is the virtual scene.
  • VR virtual reality
  • AR augmented reality
  • Figure 3 shows a handheld near-to-eye display device.
  • the term “near-to-eye display device” refers to any device that produces a light wave distribution of a virtual scene on an exit pupil plane from a physical distance that is less than the typical minimal distance at which the human eye can normally focus (e.g., 25 cm).
  • a near-to-eye display device may be handheld as in Figure 3, or may head- worn as in Figure 1.
  • a near-to-eye display device may also be stationary for applications in which a user is expected to place their head against or near the near-to-eye display device (e.g., VR demonstrators).
  • the example handheld near-to-eye display device of Figure 3 is in the shape of a pair of binoculars, but this is not a limitation of the present invention. Any type of near-to-eye display device: head- worn, handheld (e.g., electronic viewfinders in cameras, foldable image viewer, smartphones), or otherwise, may include embodiments of the present invention.
  • Any near-to-eye display device may include any of the invention embodiments described herein.
  • any near-to-eye display device may include any of the embodiments that produce a light wave distribution on the exit pupil plane 220 described herein.
  • FIG. 4 shows a cross section of an SLM being illuminated and generating a virtual scene wave.
  • SLM 410 is shown as a transmissive SLM.
  • Illumination optics module 440 produces, and illuminates SLM 410 with, a coherent illumination wave 442.
  • SLM 410 modulates the light and creates virtual scene wave 412.
  • Encoded in virtual scene wave 412 is a 3D virtual scene that is imaged on the user's retina 424. Only the portion of the virtual scene wave that intersects the user's pupil 212 on the exit pupil plane 220 creates an image on the retina.
  • Other information in the virtual scene wave that falls outside the user's pupil is filtered out and does not enter the user's eye.
  • Various invention embodiments that employ pupil filtering are discussed in more detail below.
  • Illumination optics module 440 shown here creates a converging illumination wave. In some embodiments, this is accomplished with light sources and optical components such as mirrors, micromirror arrays, lenses, and the like. Various embodiments of illumination optics modules are described in more detail below.
  • the illumination optics module does not necessarily generate a converging illumination wave.
  • one simple example of an illumination optics module is a point light source 120 ( Figure 1). In that case, the illumination wave is a diverging wave. Yet, in other embodiments shown below, the illumination wave is generated by arrays containing multiple point light sources. However, in any case, the illumination wave must possess a certain degree of spatial coherency over sufficiently large areas of the SLM.
  • SLMs are basically dynamically programmable diffractive optical elements.
  • SLMs based on nematic liquid crystals (LC) make use of the electrically controlled refractive index of anisotropic LCs to modulate polarization, intensity or phase of incident light.
  • the type of modulation depends on the mode of the LC that is used. Twisted nematic LCs rotate the polarization of incident light by some controlled amount, and are used along with polarizers on both sides to constitute intensity modulators suitable for incoherent light applications, most commonly, 2D displays.
  • Parallel aligned nematic (PAN) (or electrically controlled birefringence (ECB)) mode LCs are most suitable for coherent light applications, and they can be used as multilevel phase only SLMs.
  • Transmissive SLMs based on LCs have large pixel pitch due to the fact that electronic circuitry associated with each pixel must be embedded within the pixel aperture.
  • Reflective SLMs based on Liquid Crystal on Silicon (LCoS) technology can be made to have much smaller pixel pitches, since electronics can be buried under the pixel.
  • One advantage of SLMs based on nematic LCs is the multilevel modulation these devices can perform. However, their performance is limited by pixel crosstalk and low frame rates, which may be problematic in color field sequential holographic applications.
  • SLMs based on ferroelectric LCs have much higher frame rates at the cost of merely binary modulation at each pixel.
  • MEMS Microelectromechanical systems
  • SLM Digital Micromirror Device
  • DMD Digital Micromirror Device
  • Other MEMS SLMs can be implemented using deformable membrane structures, piston motion micromirror arrays, programmable diffraction gratings such as grating light valve devices, electro- wetting and magneto-optic Kerr effect devices, or MEMS laser scanners.
  • FIG. 5 shows the cross section of Figure 4 depicting the virtual scene as seen by a user.
  • Virtual scene 150 includes one virtual object: a 3D
  • SLM 410 converts the illumination wave to the virtual scene wave that would be emanated by virtual scene 150.
  • the SLM is illuminated by a coherent wavefront, which is generated by a group of optical components and light sources that are part of illumination optics module 440.
  • the computer generated holographic image displayed on the SLM helps convert the
  • the SLM is the device where information about the virtual scene is fed to the light wave that is delivered to the eye of the user.
  • the SLM is able to synthesize only a portion of the wave emanated by the virtual scene, and the incident wave is only partially converted to the wave emanated by the virtual scene.
  • real SLMs have finite spatial sizes, which restrict the size of a virtual scene that is displayed (or, the field of view (FOV) within which the virtual scene is visible), and finite spatial bandwidths (pixel pitches usually several multiples of wavelength), which limits the portion of the wave emanating from each virtual scene point that can be reconstructed.
  • FOV field of view
  • SLMs also generate higher diffraction orders as a result of their pixelated structure. These orders correspond to shifted replicas of virtual scenes that are to be displayed, which appear as ghost image replicas if they are allowed to enter the eye and propagate to the retina.
  • conjugate beam In the case of SLMs that essentially provide real valued modulation, such as binary SLMs or amplitude only SLMs, a conjugate beam will be formed.
  • This conjugate beam which corresponds to the wave emitted by a second virtual scene which is the mirror image of the actual virtual scene with respect to the exit pupil plane, should also be prevented from entering the eye.
  • Figure 6 shows an SLM with a pixelated structure.
  • the pixelated structure of SLMs is intimately linked with sampling and interpolation of light waves.
  • the final analog optical mask structure that is implemented on the SLM can be considered to be obtained by sampling and re-interpolating the ideal holographic image that is intended to be displayed on the SLM. Rate of sampling is determined by pixel pitch of the SLM, while the pixel aperture function of the SLM constitutes the interpolating function. It is well known that when a signal is sampled in the space domain, its spectrum is periodically replicated in the spatial frequency domain.
  • HDO higher diffraction orders
  • Figure 6 shows an SLM having a pixel pitch P at an eye relief distance D from a user's eye. The distance X between diffraction orders on the exit pupil plane can be approximated for small angles as
  • is the wavelength of light incident on the SLM.
  • various embodiments of the present invention select values for pixel pitch, expected eye relief distance, wavelength, and other parameters, such that the user's eye pupil forms an effective spatial filter.
  • the hatch pattern shown in the SLM of Figure 6 and other figures is not to any particular scale.
  • the hatch pattern, when included, is used as a visual aid to differentiate SLMs from other optical components in the figures, and is not meant to imply an actual pixel pitch scale.
  • Figure 7 shows a cross section of an SLM that generates noise beams and multiple diffraction orders.
  • SLM 710 may be either transmissive or reflective.
  • Figure 7 show the light modulated by the SLM, but does not show the illumination wave.
  • the illumination wave may come from any direction.
  • the light wave distribution falling on the exit pupil plane 220 includes the virtual scene wave (the 0 order), higher diffraction orders (HDO), and noise beams.
  • the useful portion of the exit pupil plane is that portion that ideally includes the virtual scene wave and nothing else. As shown in Figure 7, noise beams and HDOs are not included in the useful portion of the exit pupil plane. As described further below, when a user's eye pupil is substantially aligned with the useful portion of the exit pupil plane, the correct virtual scene represented by the virtual scene wave is imaged on the user's retina.
  • Figure 8 shows the cross section of Figure 7 with a user's eye pupil filtering out unwanted noise beams and diffraction orders.
  • Embodiments represented by Figure 8 eliminate HDOs, noise beams, DC beams, conjugate beams, and other possibly disturbing beams by using the eye pupil of the user as a spatial filter.
  • no attempt to eliminate the undesired beams is made (optically or computationally) within the near-to-eye display device before these beams reach the exit pupil plane.
  • the optical architecture of the system is designed and the holographic image on the SLM is computed such that on the exit pupil plane, there is a useful portion within which only the virtual scene wave exists, and all other undesired beams fall outside this region.
  • this useful portion is at least equal to the size of the pupil of the user, and (if needed) this useful portion is steered to follow the pupil movements of the user, then the undesired beams are automatically eliminated by the user's pupil and do not propagate to the retina.
  • This technique which we refer to as "pupil filtering," has the benefit of reducing the bulk within optical designs, but demands the SLM pixel pitch to be sufficiently small, or
  • optical filters such as 4f filters
  • 4f filters may be used within the system to eliminate HDOs and noise beams before they reach the exit pupil plane.
  • the bulk within the system increases.
  • Most embodiments of the invention described herein employ pupil filtering, and therefore benefit from reduced bulk and weight.
  • Figures 9, 10, and 11 show multiple diffraction orders on an exit pupil plane having a useful portion. Each of Figures 9, 10, and 11 show multiple diffraction orders as black dots.
  • the centermost diffraction order is the virtual scene wave, which includes the information desired to be propagated to the retina.
  • Figure 9 also shows a schematic representation of the noise beams surrounding the virtual scene wave. In operation, the noise beams have a finite distribution not shown in the figures.
  • the ideal useful portion of the exit pupil plane includes all of the virtual scene wave and nothing else.
  • Pupil filtering works when the user's pupil is substantially aligned with the useful portion of the exit pupil plane such that the virtual scene wave is allowed to propagate to the retina while everything else is filtered out.
  • ideal pupil filtering may not always be achieved.
  • the user's pupil substantially overlaps the useful portion of the exit pupil plane ( Figure 10).
  • Figure 10 provide less than perfect pupil filtering.
  • Some embodiments generate a useful portion of the exit pupil plane large enough so that it is at least the size of an expected pupil size. Physically, the minimum pupil width is typically assumed to be 2mm. However, what is of concern to us is the physical size of the image of the pupil in front of the cornea (i.e.
  • some embodiments of the present invention create a useful portion having a width no smaller than about 3 mm. If the width of the useful portion is smaller than 3 mm, some part of the undesired beams may enter through the pupil, degrading the image quality at the retina. Further, some embodiments maintain the amount of average light power that is delivered to the eye above a certain threshold, in order to guarantee that the user's pupil size stays at the lower size limit when the display device is in use.
  • Figures 12-34 show various optical architectures suitable for use in near- to-eye display devices described herein. Some employ transmissive SLMs and some employ reflective SLMs. Some block the real world views to create a virtual reality, and some superimpose the real world view on the virtual scene to create an augmented reality. No near-to-eye display device described herein is limited to any one (or any set) of the optical architectures. In general, subsets of each of the optical architectures may be considered as part of an illumination optics module (440, Figure 4). Further, the optical architectures in many of the figures below are shown for a single eye. In some embodiments, they are replicated to create two sides of a display. Further, in some embodiments, when they are replicated, they are mirrored to provide symmetry.
  • Figure 12 shows an optical architecture in which the SLM is placed on a converging beam path, where the converging beam is obtained from a point light source 120 by an optical component with a positive refractive power (here shown as a positive lens 1210) placed between the point light source 120 and the transmissive SLM 410.
  • the point light source is actually imaged on the exit pupil plane 220. Therefore, the point light source is optically at a conjugate plane of the exit pupil plane.
  • the SLM under the assumption that it is closer to the eye than the normal closest point of the human eye (25 cm), is not at a plane that is conjugate to the retina.
  • This architecture is that the directivity patterns of the light waves emerging from each pixel of the SLM are made to almost completely overlap on the exit pupil plane.
  • the SLM acts as the optical mask that transforms the converging illumination beam to the part of the virtual scene wave that propagates to and fills the useful portion of the exit pupil plane.
  • the spatial bandwidth requirement of the SLM is directly proportional to the width of the useful portion of the exit pupil plane. In order for the pupil filtering technique to work, SLM bandwidth must be sufficiently large so that the useful portion is greater than at least the expected minimum size of user's eye pupil.
  • the pixel pitch of the SLM must at least be smaller than the multiplication of the wavelength of light produced by the point light source and the distance between the SLM and the exit pupil plane divided by minimum size of the eye pupil.
  • a smaller pixel pitch is needed, so that some of the additional SLM bandwidth can be used to distribute the noise beam.
  • the SLM provides real valued modulation (such as binary amplitude or phase modulation, or intensity modulation)
  • the pixel pitch must be halved, since half of the bandwidth will be occupied by the conjugate beam.
  • the useful portion can be located at a slightly off axis eye position so that the DC beam can also be filtered by the eye pupil.
  • the optical component that focuses the diverging light from the point light source to the exit pupil plane might represent a reflective element, such as an elliptical mirror, a spherical mirror, etc. Such a component both acts as a lens and also changes the optical axis.
  • Figure 13 shows an optical architecture in which the SLM is illuminated by a diverging wavefront.
  • the light modulated by the SLM which has an overall diverging character, is then collected by an eyepiece lens 1310 and directed towards the eye.
  • the point light source and the exit pupil plane are again conjugate planes.
  • the SLM might or might not be at a plane that is conjugate to the retina depending on its position.
  • the eyepiece lens basically forms an image of the SLM, which might be virtual or real depending on the position of the SLM.
  • This image of the SLM is referred to herein as the "effective SLM" and it appears to be illuminated by a converging wave.
  • the architecture is equivalent to the architecture shown in Figure 12.
  • the pupil filtering technique works if the pixel pitch of the effective SLM is sufficiently small as discussed in Figure 12.
  • a reflective surface such as an elliptical, spherical, etc. mirror may be the optical equivalent of the eyepiece lens illustrated here.
  • This architecture constitutes a convenient option for designing augmented reality displays, especially in cases where the SLM is reflective and non transparent.
  • the SLM might be placed on the side of the eyeglass frame, and the light from the SLM can be directed toward the eye by a semitransparent reflective surface, which is the optical equivalent of the eyepiece lens illustrated here.
  • Such architectures are illustrated in subsequent figures.
  • Figure 14 shows an optical architecture with a point light source and SLM, with no other components with refractive power.
  • the point light source is not at an optical conjugate plane of the exit pupil plane, since it is not imaged on the exit pupil plane.
  • the SLM is not at an optical conjugate plane of the retina.
  • the greatest advantage of this architecture is its simplicity, thus the potential for realizing near-to-eye display deviices with quite thin form factors, since no components other than the SLM and point light source are present.
  • the SLM is illuminated with diverging light, and the light from the SLM retains its overall diverging character at the exit pupil plane, directivity patterns of light waves from each pixel of the SLM do not overlap on the exit pupil plane.
  • Some embodiments use SLMs with lower fill factors. In these embodiments, though there is a loss in light efficiency, the directivity patterns of SLM pixels become uniform, i.e., SLM pixels optically behave closer to isotropic point light sources, and the intensity variation described above no longer exists. Further, in embodiments where the SLM generates an
  • unmodulated DC beam that beam is not focused to a single spot on the exit pupil plane, but spreads over a large area. Hence, some part of it enters into the useful portion. However, since the energy is spread out, only a little portion of the unmodulated DC beam is intercepted, and the related background noise on the retina is quite low if not perceivable at all.
  • Figure 15 shows an architecture in which an SLM is illuminated in a time sequential manner by an array of point light sources.
  • five point light sources PSl to PS5 are illustrated where PS3 is assumed to be on.
  • the architecture is the same with the architecture in Figure 14, and the non-uniform brightness problem discussed in Figure 14 is present.
  • the point light source that is switched on changes, the part of the SLM that contributes to the useful portion with the highest power changes.
  • the power contributed by a particular section of the SLM to the useful portion changes as the point light source that is turned on changes.
  • the number and positions of point light sources are arranged such that when time averaged, every part of the SLM sends equal power to the useful portion. Therefore, the point light source array enables us to obtain a uniform variation of brightness in the field of view by time integration of retinal images created by different point light sources.
  • Embodiments represented by Figure 15 demand a higher frame rate SLM than previously described embodiments.
  • the higher frame rate is driven in synchronism with the point light sources, and the deployment of multiple point light sources.
  • the holographic image on the SLM must be updated according to the new position of the illumination wave. Therefore, multiple holographic images need to be computed for each video frame of the virtual scene.
  • point light sources should be turned on one at a time only if all light sources significantly illuminate every part of the SLM and no crosstalk at all among reconstructions by different point light sources is tolerable.
  • point light sources PSl, PS3, and PS5 may form the first group
  • PS2 and PS4 may form the second group.
  • the crosstalk between the point light sources in any of these groups is weak due to the fact there is sufficient separation between the light sources and the light power received from a part of the SLM is dominated by one of the sources. In this way, the demand on SLM frame rate is decreased.
  • the holographic image on each region of the SLM is computed according to the point light source through which the highest power is delivered to the useful portion from that region.
  • Figure 16 shows an architecture, which is similar to the architecture illustrated in Figure 15 with the difference that all point light sources are simultaneously turned on, and the directivity angles of point light sources are constrained, possibly by apertures placed in front of the point light sources.
  • the SLM surface is divided into a number of nonoverlapping regions (labeled 1-5 for example purposes), where each of these regions are essentially illuminated by only one of the point sources. Therefore, the light wave in the useful portion is formed by the superposition of the waves from multiple point light sources.
  • the holographic image on each of the regions of the SLM is computed according to the corresponding light source, and the final holographic image displayed on the SLM is obtained by concatenating these individual holographic images.
  • the first and second groups of light sources are turned on in a time sequential manner. Object points that lie close to the one set of boundaries might be skipped when the corresponding group of light sources are turned on, and they may be displayed only when the other group of light sources are turned on, with doubled intensity so that average power stays the same. In this way, diffraction artifacts and resolution loss for virtual scene points that lie close to the boundary regions can be avoided, however, twice a frame rate is demanded from the SLM.
  • Figure 17 shows an optical architecture in which a reflective SLM is placed directly in front of the user's eye.
  • a reflective SLM 110 is placed directly in front of the eye and is illuminated by a point light source 120 mounted on the side of the eyeglass.
  • the system is optically equivalent to the system depicted in Figure 14, and constitutes a non see through display since the SLM blocks the vision of real world.
  • Figure 18 shows an architecture in which the SLM is placed such that real world vision is not blocked.
  • a reflective SLM 110 is placed at a position such that real world vision is not blocked.
  • the SLM is illuminated by a point light source 120 mounted on the side of the eyeglass.
  • the light reflected from the SLM 110 is directed to the user's eye by a beamsplitter 1810.
  • the system is optically equivalent to the system depicted in Figure 14, and constitutes a see through display.
  • a transmissive SLM 410 is placed directly in front of the eye such that real world vision is not blocked, however, as the real world light passes through the SLM, the image of the real world might be slightly corrupted.
  • the SLM is illuminated by a point light source 120 mounted on the side of the eyeglass at a location that is further to the eye than the SLM.
  • the system is optically equivalent to the system depicted in Figure 14, and constitutes a see through display with some degradation of real world view.
  • a transmissive SLM 410 is placed at a position so that real world vision is not affected by its presence.
  • the SLM is illuminated by a point light source 120 mounted on the side of the eyeglass.
  • the light transmitted by the SLM is directed to the eye by a beamsplitter 1810.
  • the system is optically equivalent to the system depicted in Figure 14, and constitutes a see through display with no degradation of real world view.
  • Figure 21 shows an optical architecture in which a reflective SLM 110 is placed in front of the user's eye.
  • a look at display is implemented with a reflective SLM.
  • a positive lens 2110 is placed in front of the SLM.
  • the focal length of the positive lens is equal to eye relief distance.
  • the lens converts the diverging wave from the point light source 120 to a collimated beam, which hits the SLM with a slight angle, gets modulated and reflected, and passes once again through the same lens which now acts as an eyepiece lens and directs the light towards the pupil.
  • the system is optically equivalent to the system in Figure 13.
  • Figures 22-28 show optical architectures in which real world vision is not blocked by the SLM.
  • the reflective SLM 110 is placed to the side of the eyeglass frame so that the reflective SLM does not block the real world view.
  • An additional beamsplitter 1810 is used to direct SLM light to the eye pupil of the user.
  • the system is optically equivalent to the system in Figure 12, and constitutes a see through display.
  • a see through display is implemented with a transmissive SLM 410.
  • the diverging light wave from a point light source 120 is converted to a converging wave by a positive lens 1210.
  • the converging wave passes through the SLM and gets modulated.
  • the SLM wave is directed towards the eye with a beamsplitter 1810. Though the SLM is transmissive, the lens and the SLM are both placed before the beamsplitter so that real world view is not affected by their presence.
  • the system is optically equivalent to the system in Figure 12.
  • FIG 25 a see through display with a reflective SLM 110 is illustrated.
  • the system is optically equivalent to the system in Figure 13, where the eyepiece lens is replaced by the semi-transparent reflector 2510 placed in front of the eye.
  • the reflector 2510 can either be a single piece curved component, such as an elliptical or spherical mirror, or it can be a flat component with an array of micromirrors with different tilt angles.
  • FIG 26 a see through display with a reflective SLM is illustrated.
  • the system is optically equivalent to the system in Figure 13.
  • the beamsplitter on the right and the lens form a virtual image of the point light source, and SLM is illuminated by a diverging spherical wave which seems to emerge from the said virtual image of the point light source. This wave gets modulated, and then is bent towards the eye pupil with the combination of lens and curved mirror.
  • the architecture is advantageous in that it is compact and provides undistorted see through vision.
  • FIG 27 a see through display with a transmissive SLM 410 is illustrated.
  • the system is optically equivalent to the system in Figure 13 and different from the system in Figure 25 only in that the SLM is transmissive.
  • FIG 28 a see through display with a transmissive SLM 410 is illustrated.
  • the system is optically equivalent to the system in Figure 13 and different from the system in Figure 27 only in that beamsplitter 1810 is included.
  • Figure 29 shows an optical architecture in which an SLM and reflector are combined.
  • the SLM is fabricated directly on the semitransparent reflector.
  • the diverging light from the point light source 120 illuminates the SLM, which is directly fabricated on top of a semi-transparent reflector.
  • the SLM-reflector combination can be considered as a single device, which is similar to LCoS SLMs, but fabricated on a transparent substrate.
  • Figure 30 shows a reflector based solution for modulation in a single direction.
  • Figure 30 illustrates an embodiment of the invention in which a transmissive SLM 410 is placed between a semi-transparent reflector 2510 and the eye to constitute a see through display.
  • the reflector and the SLM are separate devices, with considerable space in between. If the polarizers 3020 and 3010 were not present, the wave emanated from the point light source 120 would get modulated by the transmissive SLM twice: firstly during the initial passage, secondly after getting reflected from the semi- transparent reflector. This double modulation is undesired especially when some of the incident light rays are modulated by different sections of the SLM.
  • a polarizer 2920 As the transmissive SLM, a liquid crystal SLM in Parallel Aligned Nematic (PAN) mode may be used, where the LC director axis of the liquid crystal is orthogonal to the axis of the polarizer 3020 that is placed in front of the point light source. Then, the light emanated by the point source does not get modulated by the SLM during the first passage.
  • PAN Parallel Aligned Nematic
  • the light wave After passing the SLM, the light wave passes through a 45 degree polarization rotator 3010, then gets reflected from the semi-transparent reflector 2510, and then passes once again through the 45 degree polarization rotator 3010 after which its polarization becomes parallel to the LC director of the SLM. Then the wave enters the SLM once again, and gets modulated this time. In this manner, double modulation is avoided and the incident light wave is modulated by the SLM only during its second passage.
  • Figure 31 shows a virtual reality (VR) architecture with SLM tiling.
  • Light from point light source 120 is collimated by collimation lens 3150, passed through a polarizer 3130, and split into two with beamsplitter 3114. One portion is fed to the first reflective SLM 3110, and the other portion is fed to the second reflective SLM 3120. Modulated light coming from the SLMs are joined by beamsplitter 3112 and then passed through a common eyepiece lens 3140 and directed to the eye.
  • the architecture is particularly useful when it is not possible to place SLMs side by side due to their external frames that contain the electronic control circuitry.
  • the SLMs used in the architecture can be identical.
  • FIG 32 shows a VR architecture with SLM tiling.
  • This architecture has a smaller form factor than the architecture in Figure 31, but it requires the LC director axis of the SLMs to be perpendicular to each other.
  • Figure 33 shows a VR display architecture.
  • a concave mirror 3310 such as used in telescopes, has an opening.
  • the diverging waves emitted by two point light sources 120 are converted to two pieces of converging waves by the mirror.
  • the converging waves illuminate the reflective SLM 110.
  • the light modulated by the SLM propagates to the exit pupil plane through the opening between the mirrors.
  • Figure 34 shows two tiled SLMs to double resolution.
  • Two identical reflective SLMs 110A and 110B are placed facing opposite surfaces of beamsplitter 3420.
  • the SLMs are illuminated by collimated light sent from an illumination optics module 3430.
  • the light emerging at 3450 is equivalent to the light generated by a single SLM that is obtained by adding the complex transmittances of the two SLMs.
  • the SLMs are positioned such that they are offset on the transverse plane by half a pixel pitch with respect to each other during the addition.
  • the equivalent SLM 3410 then has a pixel pitch that is half the pixel pitch of each reflective SLM.
  • the pixel aperture function of the equivalent SLM is the same as the pixel aperture function of the reflective SLM. Since the effective SLM has a higher pixel pitch, its bandwidth and the angular separation between diffraction orders are increased. Such a structure can be used to enlarge the size of the useful portion that can be obtained.
  • Figure 35 shows a near-to-eye display device with a high resolution portion and a low resolution portion.
  • the high resolution portion is provided by inset 3510 and low resolution portion is provided by peripheral imaging device 3520.
  • peripheral imaging device 3520 includes a microdisplay device such as an organic light emitting diode (OLED) display, a liquid crystal display (LCD), or a reflective LCD.
  • OLED organic light emitting diode
  • LCD liquid crystal display
  • reflective LCD a reflective LCD
  • the high resolution inset is an SLM that provides a virtual scene to the user as described above.
  • the SLM has a fixed location and so does the high resolution inset within the resulting display.
  • near-to-eye display device 3500 includes an SLM that provides about 30-40 degrees high resolution central foveal vision with natural depth cues, and a regular 2D display that provides a low resolution peripheral image. The idea presented here depends on the promise that though the human eyes have a large FOV - around 170 degrees - a very small portion of this FOV (around 6 degrees) constitutes sharp foveal vision at a time. Humans typically enlarge the FOV for sharp foveal vision to about 30-40 degrees by eye motion before resorting to head motion.
  • a display that supports a high quality foveal vision within a FOV of 30-40 degrees, and supplements this with a low quality peripheral vision will be an economical solution for large FOV designs.
  • the image provided by the SLM carries all natural depth cues in addition to being high resolution.
  • the eye can focus on the virtual objects seen through the SLM as in natural vision.
  • the peripheral image provided by the regular 2D display is not focused on the retina and is low resolution. However, it still establishes a degree of peripheral awareness.
  • Figure 36 shows a high resolution image for foveal vision and lower resolution for peripheral vision.
  • Image 3600 represents an image seen by a user using near-to-eye display device 3500.
  • the part of the virtual scene that falls in the central vision part of the FOV appears as a high resolution image, while the part that falls in the peripheral vision part appears as a low resolution and defocused image.
  • Figure 37 shows the high resolution image being steered to a user's pupil position. Some embodiments provide for the high resolution image to be moved within the field of view. Examples of these embodiments are described with reference to figures that follow.
  • Image 3700 represents an image seen by a user when the user's pupil is tracked as the user looks to the right within the FOV.
  • the high resolution inset is steered to follow the user's eye movement.
  • Figures 38 and 39 show a display system with a rotating hologram module to create a steerable high resolution image.
  • only the 6-10 degree portion of the FOV is provided by the SLM at a single time. In other embodiments, more than 6-10 degrees is provided at a time.
  • the rest of the FOV is covered by a regular 2D display. Pupil movements of the user are tracked, and the hologram module 3810 is rotated based on those movements to steer the SLM light towards the pupil.
  • Part of the 2D display image that lies within the central vision region is temporarily blackened, so that the central vision is formed only by the SLM and thus is high resolution.
  • the reflector is designed such that the SLM light is directed towards the eye pupil for any position of the eye pupil.
  • Rotating hologram module 3810 is shown with an SLM, lens, beamsplitter, and light source. Any of the optical architectures described herein may be employed within rotating hologram module 3810 without departing from the scope of the present invention.
  • LCD 3820 is used as peripheral imaging device 3520 ( Figure 35), and rotating hologram module 3810 illuminates a portion of LCD 3820 to create the high resolution inset 3510 ( Figure 35).
  • Rotating hologram module 3810 may be physically location on the frame of near-to-eye display device 35.
  • rotating hologram module 3810 may be co- located with a point light source 120.
  • Figure 39 shows pupil tracker 3910 tracking movement of the user's eye 210 and actuator 3920 used to rotate rotating hologram module 3810.
  • pupil tracker 3910 sends a signal to actuator 3920 to cause the hologram module to rotate.
  • Pupil tracker 3910 may include any suitable components capable of performing as described.
  • pupil tracker 3910 may include one or more cameras, one more light sources (e.g., infrared), and a processing element to interpret the pupil tracking data and to command actuator 3920.
  • Actuator 3920 may include any type of component capable of performing as described.
  • actuator 3920 may be a stepper motor or series of stepper motors coupled to rotating hologram module 3810.
  • Figure 40 shows a portion of a near-to-eye display device having a moving platform.
  • Moving platform 4010 moves within the field of view of the user.
  • Moving platform 4010 is actuated by circuits (not shown) mounted on the near-to-eye display device, or connected to the near-to-eye display device with cabling or wirelessly.
  • moving platform includes light sources and/or SLMs.
  • the light sources and/or SLMs are driven by circuits (not shown) mounted on the near-to-eye display device, or connected to the near-to-eye display device with cabling or wirelessly.
  • Figure 41 shows a moving platform upon which an SLM bar that covers about 30-40 degrees of central FOV is mounted, along with two LED bars each of which covers about 30 degrees of peripheral FOV.
  • the SLM bar includes a plurality of pixels, the spacing of which satisfies the criteria described herein with respect to the useful portion of the exit pupil plane.
  • the LED bars may include any number of pixels. In some embodiments, the resolution of the LED bars is less than the resolution of the SLM bar.
  • the entire platform 4010 can move up and down periodically to scan the vertical direction. The display is considered see-through since the moving platform does not continuously block any part of the user's FOV, but does so only for a short duration of time. Both the SLM bar and the LED bar have high refresh rates.
  • Figure 42 shows a portion of a near-to-eye display device having a moving platform with an array of bars.
  • Moving platform 4210 includes more than one bar that moves up and down in the vertical direction to fill the FOV.
  • Moving platforms that include a plurality of bars, such as platform 4210 are also referred to herein as "slotted platforms.”
  • Moving platform 4210 is actuated synchronously with the SLM data being driven on the various SLM elements on moving bar 4210.
  • the idea is similar to Figure 40, except for the fact that an array of bars are used so that each of the bars need to scan a smaller vertical range, relieving the frame rate constraint on the SLM bar and LED bar.
  • Portions of moving platform 4210 are considered to include a
  • microdisplay For example, the portions of bars 4010 that include LEDs and the LED bars above and below bars 4010 constitute a microdisplay.
  • microdisplays on moving bars have a lower resolution than SLM bars. Also in some embodiments, microdisplays on moving bars have a greater pixel pitch than SLM bars.
  • Figure 43 shows a moving platform having an array of bars upon which SLMs and microdisplays are mounted.
  • Moving platform 4210 includes a plurality of bars equivalent to 4010, and a plurality of bars that only include LEDs.
  • the SLM bar is mounted only on the bars in the middle so that 30-40 degrees of FOV is covered in the vertical direction as well.
  • the top and bottom bars only consist of LEDs, since they are not responsible for central foveal vision but only peripheral vision.
  • Figure 44 shows an even simpler design where only a small SLM is mounted on the middle bar for central vision, while two more LED bars are placed to provide peripheral vision.
  • the SLM bar only covers about 6-7 degrees of horizontal FOV.
  • the bar For a fixed position of eye pupil, the bar only scans in the vertical direction to cover 6-7 degrees of vertical FOV as well.
  • the SLM bar also moves in the horizontal direction to cover the portion of the FOV for central vision.
  • all bars shown move as described, and in other embodiments, only the middle bar with the SLM moves as described.
  • Figure 45 shows a near-to-eye display device with a transducer to interact with a user for calibration.
  • Near-to-eye display device 4500 is similar to near-to- eye display device 100 ( Figure 1) with the addition of adjustment knob 4510.
  • Adjustment knob 4510 is an example of a transducer that allows the user to interact with the near-to-eye display device. For example, in some
  • near-to-eye display device 4510 may perform calibration actions in which the user is asked to provide feedback using the transducer.
  • calibration actions are now described.
  • Figures 46, 48, 50, and 52 show flowcharts of calibration methods in accordance with various embodiments of the present invention.
  • these methods, or portions thereof are performed by a near-to-eye display device, embodiments of which are shown in, and described with reference to, the figures of this disclosure.
  • these methods are performed by a computer or an electronic system.
  • the various calibration methods are not limited by the particular type of apparatus performing the method. Further the disclosed actions in the calibration methods may be performed in the order presented, or may be performed in a different order. Also, in some embodiments, some actions listed in the figures are omitted while performing method embodiments.
  • a user is prompted to identify a type of any visual disorder.
  • An example image to prompt a user is shown at 4710 ( Figure 47).
  • the near-to-eye display device may display a chart from which the user may make a selection.
  • the user has selected myopia, and the system presents a chart prompting the user to select the smallest letter the user can comfortably read.
  • a user may make a selection by looking at it, in which case built in pupil tracking hardware can interpret the selection.
  • a user may interact with a transducer, such as adjustment knob 4510 ( Figure 45) to make the selection, and in still further embodiments, a user may interact with a touch sensitive portion of the display area on the near-to-eye display device.
  • a transducer such as adjustment knob 4510 ( Figure 45) to make the selection
  • a user may interact with a touch sensitive portion of the display area on the near-to-eye display device.
  • a light wave distribution is modified to present the user with at least one test image intended to determine a degree of the visual disorder suffered by the user. For example, in some embodiments, a single image such as that shown at 4730 ( Figure 47) is shown to the user. In other embodiments, multiple images such as those shown at 4920 ( Figure 9) are shown to the user.
  • feedback is received from the user regarding the at least one test image.
  • this corresponds to a user selecting an image using a transducer.
  • this corresponds to a user turning an adjustment knob.
  • the image at 4730 may be focused at different distances until the user's myopia has been overcome.
  • the light wave distribution is modified to correct for the visual disorder suffered by the user. This is shown at 4740 ( Figure 47).
  • the different images displayed are generated using an SLM as described above. Visual disorders may be corrected using the computation of the SLM data.
  • a light wave distribution is modified to present the user with a plurality of test images intended to determine a type of visual disorder (if any) suffered by the user. For example, in some embodiments, images such as that shown at 4910 ( Figure 49) is shown to the user.
  • feedback is received from the user regarding the plurality of test images. In some embodiments, this corresponds to a user selecting an image using a transducer. In other embodiments, this corresponds to a user turning an adjustment knob. In still further embodiments, this corresponds to a user interacting with a touch sensitive portion of the display.
  • the type of visual disorder suffered by the user is determined based on the feedback received.
  • the user has selected an image corresponding to myopia.
  • the light wave distribution is modified to present the user with a second plurality of test images intended to determine a degree of the visual disorder suffered by the user. For example, in some embodiments, multiple images such as those shown at 4920 ( Figure 9) are shown to the user.
  • additional feedback is received from the user regarding the second plurality of test images.
  • this corresponds to a user selecting an image using a transducer.
  • this corresponds to a user turning an adjustment knob or interacting with a touch sensitive display.
  • 4840 and 4850 are performed more than once to determine the proper correction to be applied to correct the user's visual disorder.
  • the light wave distribution is modified to correct for the visual disorder suffered by the user. This is shown at 4930 ( Figure 49).
  • the different images displayed are generated using an SLM as described above. Visual disorders may be corrected using the computation of the SLM data.
  • the user is prompted to enter the type and degree of the visual disorder at 5010.
  • Example images to prompt a user are shown at 5110 and 5120 ( Figure 51).
  • the near-to-eye display device modifies a light wave distribution to present the user with at least one test image intended to correct for the visual disorder suffered by the user at 5020. This is shown at 5130.
  • feedback is received from the user regarding the at least one test image.
  • this corresponds to a user selecting an image using a transducer.
  • this corresponds to a user turning an adjustment knob or interacting with a touch sensitive display.
  • the image at 5130 may be focused at different distances until the user's myopia has been overcome
  • the light wave distribution is modified to correct for the visual disorder suffered by the user. This is shown at 5140 ( Figure 51).
  • the different images displayed are generated using an SLM as described above. Visual disorders may be corrected using the computation of the SLM data.
  • user profiles are stored within the near-to-eye display device for later retrieval.
  • the calibration methods described also provide actions to allow for brightness, contrast, and color correction. Any type of visual setting may be applied and any type image enhancement may be incorporated without departing from the scope of the present invention.
  • Figure 52 shows a flowchart of calibration methods in accordance with various embodiments of the invention. Methods represented by Figure 52 differ from the previously described calibration methods in that methods represented by Figure 52 interact with one or more actuators on the near-to-eye display device to correct for anomalies. At 5210, at least one test image is displayed. This is shown in Figure 53. Note that in all calibration embodiments, test images are not actually displayed on an eyeglass lens, but rather, test images are made to be part of a virtual scene using the SLM and pupil filtering methods described above.
  • the user is shown multiple test images that are at different depths and transverse positions.
  • two cameras that are mounted on the HWD and well calibrated with respect to each other
  • the cameras may be infrared cameras.
  • Example images that may be recorded are shown in Figure 54.
  • the recorded images are analyzed for user characteristics such as interpupil distance (IPD), and convergence when viewing 3D images.
  • IPD interpupil distance
  • images of the eyes captured by the cameras may be analyzed to determine the convergence points of the eyes.
  • the images from the infrared cameras are processed to precisely understand the eye relief distance, orientation of the display with respect to the head and eyes of the user, inter pupillary distance of the user, etc.
  • the sizes of the iris, limbus, and locations of and distances between Purkinje images are detected and analyzed to get the best prediction about the values of the relative positional parameters between the eyeglass and the user's eyes.
  • the near-to-eye display device may cause the actuators 5310, 5320, and 5330 to change physical characteristics of the device to accommodate a user's physical characteristics.
  • the near-to-eye display deviice may adjust SLM position, position of light sources, distance between the two SLMs, and the like. Examples are shown in Figure 55.
  • the various embodiments of the present invention provide for precise calibration between left and right display units, which enables correct and comfortable 3D vision.
  • the display Once the display is calibrated for a user, she can save the parameters in a user profile stored in the system. When she wears the display the next time - after another user - she can select her own profile. In this way, display can be shared by multiple users in a convenient manner. Further, when a user wears the display, the cameras may take a quick photo of the eyes, and use iris recognition algorithms to recognize a user who used the display before. Then, automatic re-calibration of the display for that user can be performed.
  • Figure 56 shows a flowchart representing computation of SLM data.
  • method 5600 or portions thereof, is performed by a near-to- eye display device, embodiments of which are shown in previous figures.
  • method 5600 is performed by a computer or an electronic system.
  • Method 5600 is not limited by the particular type of apparatus performing the method.
  • the various actions in method 5600 may be performed in the order presented, or may be performed in a different order. Further, in some embodiments, some actions listed in Figure 56 are omitted from method 5600.
  • the display system delivers, to the useful portion, the light wave that would be generated by a virtual scene.
  • an SLM is the component through which the information about the virtual scene is imposed on the light waves generated by the light source or sources within the display. Therefore, the SLM transforms the non-information-bearing wave illuminating it to the information-bearing wave part of which is delivered to the useful portion.
  • the main computational steps to determine SLM data for a frame of video are as follows:
  • the user sees a virtual scene by intercepting the light wave delivered to the useful portion of the exit pupil plane. Therefore, the computational procedure starts by the computation of the light wave that should be delivered to the useful portion, which is referred to herein as the "2D complex valued profile of virtual scene wave on useful portion". This wave carries all the information about the virtual scene. Ideally, the display system delivers only this wave within the useful portion and nothing else. However, as discussed above, due to the restrictions of the SLMs, several undesired beams are also generated by the SLM and they propagate to the exit pupil plane, but hopefully fall outside the useful portion.
  • a point cloud model is used to represent a virtual scene.
  • the virtual scene is assumed to consist of a finite number of point light sources.
  • some of these virtual scene points are visible, while some others are non-visible, due to occlusion.
  • the "2D complex valued profile of virtual scene wave on useful portion" can be computed (for a single wavelength) by superposing the diverging spherical waves emitted by each of the visible virtual scenes on the useful portion as follows:
  • U vs (x, y) is the 2D complex valued profile of virtual scene wave on useful portion
  • M is number of virtual scene points that are visible
  • is the phase of the the virtual scene point
  • yi , Zj is the location of the i virtual scene point
  • is the wavelength
  • the computation according to the formula above also applies only for a certain position of the pupil position.
  • providing a motion parallax upon changing positions of eye pupils of the user is not critical or demanded at all.
  • the computation in the above equation does not need to be updated, and it is sufficient to only steer the display light to the new position of the eye pupil, using mechanisms that are described more fully below.
  • the system it is desired for the system to provide a motion parallax.
  • the position of the eye pupil changes, which corresponds to a change in perspective from which the virtual scene is viewed, one needs to identify the new set of virtual scene points that become visible, and the new set of occluded virtual scene points. Then, the "2D complex valued profile of virtual scene wave on useful portion" should be recalculated as described above.
  • the polygon mesh models might be more suitable, especially when system compatibility with or expandability on existing computer graphics hardware and software is of concern.
  • most of the computer graphics related tasks such as occlusion, shading, etc. can be accomplished on a graphical processing unit (GPU).
  • GPU graphical processing unit
  • the virtual scene, which is represented by polygon mesh models can be rendered by a GPU, which handles any occlusion, shading etc. effects as currently done with standard GPUs embedded in computers, smart phones etc.
  • the final rendered RGB image, as well as the depth or z buffer data (which is a standard piece of data stored in GPUs and represents the distance of each rendered virtual scene point to viewpoint under consideration) can be imported from the GPU.
  • a spherical wave term emitted from the depth location of the point as inferred from the z or depth buffer, can be superposed to compute the "2D complex valued profile of virtual scene wave on useful portion", as described above.
  • This step involves numerically back-propagating the "2D complex valued profile of virtual scene wave on useful portion" through any optical components that lie in the pathway between the exit pupil plane and the SLM, including sections of free space, and determining the "ideal 2D complex valued profile of wave at the exit of SLM".
  • the identifier "ideal” stresses that if this field were present at the exit of the SLM, the light wave distribution on the exit pupil plane would not consist of any undesired components.
  • the SLM is placed directly in front of the user's eye. Therefore, there are no optical components in between except for a section of free space.
  • the "ideal 2D complex valued profile of wave at the exit of SLM" can be found as:
  • W U p(x,y) is the aperture function of useful portion
  • h ) ⁇ x,y) is the impulse response of free space propagation.
  • detailed wave optics models to account for the behavior of the components should be developed. These models should relate the light wave distribution at the entrance side to the light wave distribution at the exit side of a component. Note that when the model for each component is sufficiently detailed; aberrations, distortions, diffraction effects, and other effects introduced by the components between the eye and the SLM are incorporated into the computations automatically.
  • This step involves development of detailed wave optics models for components that lie in the pathway between the point light source and the SLM, and propagating the light wave emitted by the point light source and passing it through the related components. Aberrations introduced by components between the point light source and the SLM are incorporated in the computations in this manner.
  • Ujii (*, >) is the 2D complex valued profile of wave illuminating the SLM.
  • This may be performed using any suitable method, including, but not limited to, Iterative Fourier Transform Algorithm, Gerschberg-Saxton algorithm, Fineup Algorithm with Don't Care, Error Diffusion, Direct Binary Search, etc. - These are all existing commonly known and used algorithms.
  • the noise generated as a result of encoding is distributed in spatial frequency domain to regions that are outside the support of the "2D ideal complex valued discrete SLM image"
  • the noise beams after getting generated by the SLM plane and propagate to the exit pupil plane, get distributed outside the useful portion.
  • Computations for embodiments that include moving SLM bar are the same with the exception that the SLM is partitioned into a number of slices, and the entire "2D actual digital SLM image" is displayed slice by slice in a time sequential manner depending on the scan location of the SLM bar. The scan is completed in the frame time reserved for the "2D actual digital SLM image".
  • Figures 57-64 show a number of space-angle (or space-frequency) distributions that illustrate the basics of the computation procedure.
  • Space angle distributions are a well known concept in the literature and illustrate the distribution of rays as a function of transverse spatial position over various planes in an optical system, and provide insight.
  • the SLM is illuminated by a converging spherical wave obtained from a point light source and a positive lens combination, as in Figure 12.
  • x is assumed to denote the transverse coordinates
  • z denotes the longitudinal propagation axis.
  • the angle of each ray is measured from z axis towards +x axis.
  • Figure 65 shows a method for fast hologram computation.
  • the computational steps detailed above in general comprise wave propagation simulations that require large storage memory and high computational power.
  • One such case is that of embodiments in which the SLM is placed directly in front of the eye and is illuminated by a converging spherical wave.
  • the mathematical relation between the "2D ideal analog complex valued SLM transmittance" and each planar cross section of a virtual scene becomes simpler and a direct one.
  • the converging illumination wave is generated by a collimated beam of light and a thin positive lens with a focal length given by D (the distance between SLM and eye), that is placed immediately before the SLM.
  • D the distance between SLM and eye
  • the orders of the SLM and the lens can be changed, since both are assumed to be thin multiplicative components.
  • collimated light illuminates the SLM and then passes through an eyepiece lens to get directed towards the eye.
  • the computation procedure can be carried out with a larger step size, and with simulation windows of a common and smaller size, lowering the memory requirements drastically. Also, the computation procedure for each planar cross section is completed in parallel through the well known Angular Spectrum method, hence no loops over the points on a planar cross section are needed.
  • the two dimensional complex valued profile of a virtual scene wave on a useful portion of an exit pupil plane may in some embodiments be determined by: partitioning the virtual scene into a plurality of spherical surfaces concentric at the center of the useful portion of the exit pupil plane with different radius;
  • each element of the matrix is associated with a specific angular location on the sphere, and each element is filled with the complex amplitude of the point source at that angular location on the sphere;
  • Figure 66 illustrates the fundamentals of the method for delivering speckle free images to the retina of a user.
  • Speckle noise is observed when real objects are imaged by a user under coherent light, due to the fact that surface irregularities of real objects lead to a random phase distribution over the object.
  • each point of a real object point is imaged as an Airy disk on the retina due to diffraction from eye pupil.
  • individual images of object points overlap on the retina. Since the illumination is coherent, these individual images do interfere with each other.
  • the random phase on the real object points causes this interference pattern to exhibit rapid constructive and destructive interference patterns, which are seen as the speckle noise over the image of a real object.
  • speckle noise depends on the relative position between eyes of an observer and the object that is imaged, and the speckle noise pattern changes as the eye moves.
  • the various embodiments of the present invention are not imaging real objects under coherent illumination. Rather, they are delivering images of virtual objects to the retina using coherent illumination. This case is different than the case of real objects in that we have total control over the phase distribution that we can assign to virtual object points.
  • Figure 66 the image on the retina of a virtual object that consists of a number of point sources is illustrated for two cases. In both cases, the final continuous intensity function that forms on the retina is obtained as an interpolated version of the discrete images of virtual object points.
  • ⁇ wpx , wpy are the dimensions of the eye pupil - assumed to be
  • the intensity function exhibits rapid intensity variations between the discrete images of virtual object points. The user perceives these rapid changes as the speckle noise.
  • an appropriate phase distribution has been assigned to the virtual object points.
  • the intensity function that forms on the retina is a smoothly interpolated version of discrete images of virtual object points.
  • the image delivered by the system resembles the image that would be seen under incoherent illumination, free of speckle noise.
  • the interpolation that forms on the retina becomes smooth.
  • An equivalent condition is that light waves from each object point to arrive at the pupil of the user in phase. Therefore, if a virtual object point that has a distance of R to the center of the pupil is assigned a phase of e ⁇ ⁇ with k denoting the wave number, the light from all virtual object points arrive on the pupil of the user in phase, and form a speckle free retinal image.
  • the proposed phase assignment is specific to a certain pupil position and wavelength. Hence, it must be updated when the pupil location changes, and when the object wave within the useful portion is calculated for a different color.
  • BLUs Back Light Units
  • BLUs described below are suitable for use in an illumination optics module such as illumination optics module 440 ( Figure 4).
  • Various BLU embodiments create coherent light beams that may be converging, diverging, or collimated.
  • BLUs are also described as being part of near-to-eye display devices.
  • the BLUs may be incorporated in any near-to-eye described herein, including for example those described with reference to Figures 1, 3, 35, and 53.
  • Figure 67 shows a perspective view of a back light unit that generates a two dimensional converging beam. The rays emanate from a transparent substrate and focus on the convergence point.
  • Back light unit 6700 includes first face 6710 from which a converging light beam emanates.
  • Back light unit 6700 also includes second face 6720. In some embodiments, faces 6710 and 6720 are parallel, but this is not a limitation of the present invention.
  • Apparatus 6700 is referred to as a "back light unit” in part because it can be used to "back light” an SLM with a converging beam (or other type of beam).
  • back light unit 6700 is equivalent to the combination of point light source 120 and lens 1210 as shown in Figure 12; however, back light unit 6700 provides a significant space savings as compared to the system shown in Figure 12.
  • FIG. 68 shows a cross section of a back light unit.
  • Back light unit 6800 corresponds to back light unit 6700 ( Figure 67) implemented with a light scattering apparatus 6830 and a reflective optical element arranged as a planar micromirror array 6810.
  • the term "planar micromirror array” as used herein refers to the individual mirrors being arranged on a plane, and is not meant to infer that each mirror has the same tilt angle.
  • the light emanating from light scattering apparatus 6830 hits the micromirror array and then focuses on the convergence point.
  • the position of each individual micromirror in the array 6810 is arranged such that it reflects the incoming ray from light scattering apparatus 6830 to the convergence point.
  • the micromirror array 6810 is buried in a refractive index matched medium.
  • the reflectivity of the micromirror array can be provided either by notch coating, semi-reflective thin metal coating, or the like.
  • Light scattering apparatus 6830 scatters light away from the first face
  • micromirror array 6810 reflects the light from scattering apparatus 6830 to first face 6710 and creates the converging light beam.
  • light scattering apparatus 6830 receives light from an external light source (not shown), and in other embodiments, light scattering apparatus 6830 is co-located with one or more light sources embedded within the back light unit, and scatters light away from the first face 6710.
  • an organic light emitting diode OLED
  • RGB red, green, and blue OLEDs are included in back light unit 6800.
  • a fluorescent molecule such as a quantum dot is embedded in the substrate as a light source.
  • the light scattering apparatus 6830 includes a diffusive material such as silver epoxy or epoxy with embedded microparticles. Further, in some embodiments, the same scattering apparatus is provided for all the colors. Some embodiments include multiple scattering apparatus (a "source array") in order to increase FOV.
  • Figure 69 shows a cross section of a back light unit.
  • Back light unit 6900 corresponds to back light unit 6700 ( Figure 67) implemented with light scattering apparatus 6830 and reflective optical element 6910 arranged as a Fresnel mirror.
  • Light scattering apparatus 6830 scatters light away from the first face 6710
  • Fresnel mirror 6910 reflects the light from scattering apparatus 6830 to first face 6710 and creates the converging light beam.
  • Figure 70 shows a cross section of a back light unit.
  • Back light unit 7000 corresponds to back light unit 6700 ( Figure 67) implemented with light scattering apparatus 6830 and a reflective optical element arranged as a free form concave reflector 7010.
  • Light scattering apparatus 6830 scatters light away from the first face 6710, and reflector 7010 reflects the light from scattering apparatus 6830 to first face 6710 and creates the converging light beam.
  • Figure 71 shows a cross section of a back light unit.
  • Back light unit 7100 corresponds to back light unit 6700 ( Figure 67) implemented with light scattering apparatus 6830 and a reflective optical element arranged as a nonplanar micromirror array 7110.
  • Light scattering apparatus 6830 scatters light away from the first face 6710
  • nonplanar micromirror array 7110 reflects the light from scattering apparatus 6830 to first face 6710 and creates the converging light beam.
  • Nonplanar micromirror array 7110 reduces the shadowing effects in between the individual mirrors of a planar micromirror array.
  • Figure 72 shows a cross section of a back light unit and an external point light source.
  • Back light unit 7200 corresponds to back light unit 6700 ( Figure 67) implemented with light scattering apparatus 6830 and a reflective optical element arranged as a planar micromirror array 6810.
  • Light scattering apparatus 6830 scatters light away from the first face 6710
  • planar micromirror array 6810 reflects the light from scattering apparatus 6830 to first face 6710 and creates the converging light beam.
  • the light emanates from an external point light source 120, and is focused on light scattering apparatus 6830 inside the transparent substrate.
  • FIG. 73 shows a near-to-eye display device that includes back light unit 7300 and transmissive SLM 410.
  • Back light unit 7300 corresponds to back light unit 6700 ( Figure 67) implemented with light scattering apparatus 6830 and a reflective optical element arranged as a planar micromirror array 6810.
  • Light scattering apparatus 6830 scatters light away from the first face 6710
  • micromirror array 6810 reflects the light from scattering apparatus 6830 to first face 6710 to create the converging light beam.
  • the converging beam at the output of the back light unit passes through transmissive SLM 410 and then focuses on the eye pupil.
  • the SLM has a computer-generated hologram written on it in order to construct the desired light field on the retina.
  • a reflective SLM 110 can be used in the near-to-eye display device instead of the transparent SLM as can be seen in Figure 74.
  • Figure 74 shows a near-to-eye display device that includes back light unit 7400 and reflective SLM 110.
  • Back light unit 7400 corresponds to back light unit 6700 ( Figure 67) implemented with light scattering apparatus 6830 and a linearly arranged transreflective micromirror array 7410.
  • Light scattering apparatus
  • transreflective micromirror array 7410 reflects the light from scattering apparatus 6830 to first face 6710 where it is modulated and reflected by reflective SLM 110.
  • the modulated virtual scene wave passes back through transreflective micromirror array 7410 and emanates from the second face 6720 as a converging beam that focuses on the eye pupil.
  • the SLM has a computer-generated hologram written on it in order to construct the desired light field on the retina.
  • Back light unit 7500 corresponds to back light unit 6700 ( Figure 67) implemented with light scattering apparatus 6830 and micromirror array 6810.
  • Light scattering apparatus 6830 scatters light away from the first face 6710
  • micromirror array 6810 reflects the light from scattering apparatus 6830 to first face 6710 and creates the converging light beam.
  • Back light unit 7500 also includes cross polarizers 7510.
  • cross polarizers 7510 are two orthogonally polarized optical elements to block the passage of light. When cross polarizers 7510 are included, the bright spot referred to above is not present on the retina.
  • Back light unit 7600 corresponds to back light unit 6700 ( Figure 67) implemented with light scattering apparatus 6830 and a reflective optical element arranged as a planar micromirror array 6810.
  • Light scattering apparatus 6830 scatters light away from the first face 6710
  • planar micromirror array 6810 reflects the light from scattering apparatus 6830 to first face 6710 and create the converging light beam.
  • Back light unit 7600 also includes mirror 7610.
  • Mirror 7610 blocks light reflected from micromirror array 6810 that would otherwise create a bright spot on the retina. When mirror 7610 is included, the bright spot referred to above is not present on the retina.
  • Figure 77 shows a cross section of a back light unit with a fiber.
  • the light carried by the fiber hits the 45° angled mirror 7730 and is directed to light scattering apparatus 6830, which is used for increasing the solid angle of the ray bundle for fully covering the micromirror array.
  • Light scattering apparatus 6830 scatters the light away from first face 6710 and towards the micromirror array 6810. The scattered light is then reflected off the micromirror array 6810 and emanates from first face 6710 as a converging beam.
  • light scattering apparatus 6830 can be realized by using high refractive index transparent nanoparticles.
  • One advantage of this system can be explained as follows: the different colors can be coupled into the same fiber and directed to the same scattering apparatus. Therefore, the effective positions of the different colored light sources do not change with respect to the micromirror array, which reduces the chromatic aberrations.
  • the end face of fiber 7710 is polished with a 45° angle and coated with metal, which is used instead of mirror 7730.
  • the back light unit can be arranged such that the output beam has a profile different than the converging beam. For example, by arranging the position of the individual mirrors in the micromirror array, a one-dimensional converging beam can be generated as shown in Figure 78. Similarly, collimated and diverging beams can be generated as can be seen in Figure 79 and Figure 80, respectively.
  • wedge based back light units are now described. Many of the wedge based back light units may be used in illumination optics modules such as illumination optics module 440 ( Figure 4). Wedge based back light units are also described as being part of near-to-eye display devices. The wedge based back light units may be incorporated in any near-to-eye display device described herein, including for example those described with reference to Figures 1, 3, 35, and 53.
  • Figure 81 shows a cross section of a slab waveguide, a wedge, and a component with a micromirror array.
  • Apparatus 8100 includes slab waveguide 8110, wedge 8120, and component 8130 with micromirror array 8132.
  • Slab waveguide 8110 includes input end 8112, output end 8114, first surface 8118, and second surface 8116. First surface 8118 and second surface 8116 are parallel to each other to cause light to propagate from input end 8112 to output end 8114 by total internal reflection.
  • Wedge 8120 is coupled to the output end 8114 of slab waveguide 8110.
  • Wedge 8120 includes first surface 8128 and slanted surface 8126 that are not parallel to each other.
  • First surface 8128 and slanted surface 8126 form a continuously decreasing thickness to cause light received from slab waveguide 8110 to exit wedge 8120 from slanted surface 8126.
  • first surface 8128 is parallel to first surface 8118, and in other embodiments, slanted surface 8126 is parallel to first surface 8118.
  • Optical component 8130 includes face 8138 that is oriented parallel to slanted surface 8126. Further, optical component 8130 includes micromirror array 8132 to reflect light received through face 8138 back out through the same face 8138 and through wedge 8120. Micromirror array 8132 may be any type of micromirror array including but not limited to those shown in, and described with reference to, Figures 68-76.
  • optical component 8130 has a shape that performs as a compensating wedge for see through capability.
  • optical component 8130 is referred to a compensating wedge.
  • optical component 8130 When functioning as a compensating wedge, optical component 8130 has a wedge shape that complements the shape of wedge 8120 such that light traveling through both the wedge and compensating wedge travel through the same amount of material. This eliminates any prism effect that would otherwise be perceived by a user.
  • Optical component 8130 is positioned to provide a uniform air gap 8140 between slanted surface 8126 and face 8138.
  • point light source 120 creates a diverging light beam.
  • the diverging light beam enters slab waveguide 8110 at input end 8112 and propagates by total internal reflection within slab waveguide 8110 to output end 8114, at which point it enters wedge 8120.
  • the internal angle of incidence changes due to the decreasing thickness, allowing the light beam to exit almost collimated from the slanted surface 8126 of wedge 8120.
  • the light then enters into optical component 8130 at face 8138 and hits micromirror array 8132.
  • the light reflected from the micromirror array goes through wedge 8120, exiting at surface 8128 as a converging wave, and then focuses onto exit pupil plane 220.
  • Figure 82 shows a top view of the apparatus of Figure 81.
  • the light entering into the slab waveguide 8110 expands in the lateral direction and is confined in the vertical direction by total internal reflection. Then the light enters into the wedge region and the rays start to exit from the wedge since the incidence angles are reduced at each reflection.
  • Figures 83-88 combine wedge-based back light units with SLMs to form near-to-eye display devices. In operation, these perform the functions of both illumination optics module 440 and SLM 410 ( Figure 4). Direct applications to near-to-eye display devices are also described.
  • Figure 83 shows a cross section of a slab waveguide, wedge, optical component with micromirror array, and SLM positioned along the slab waveguide. In this configuration, the light field hits reflective SLM 110 while it is propagating in slab waveguide 8110.
  • SLM 110 is shown on surface 8118 of slab waveguide 8110 in Figure 83, this is not a limitation of the present invention. In some embodiments, the SLM is placed on surface 8116.
  • the computer-generated hologram on the SLM modulates the light as it propagates in slab waveguide 8110, and the desired virtual scene is generated at the useful portion of exit pupil plane 220 as described above.
  • Figure 84 shows a cross section of a slab waveguide, a wedge, a component with a micromirror array, and an SLM between the wedge and the component with the micromirror array.
  • a transmissive SLM 410 is placed in between wedge 8120 and optical component 8130.
  • the collimated light at the output of wedge 8120 passes through transmissive SLM 410 which has a computer generated hologram on it, and hits micromirror array 8132.
  • the light field reflects from micromirror array 8132, passes through transmissive SLM 410 again and then converges on exit pupil plane 220.
  • the light that enters from the eye pupil then constructs the desired content on the retina.
  • Figure 85 shows a cross section of a slab waveguide, wedge, component with a micromirror array, and an SLM below the wedge.
  • Figure 85 is similar to Figure 83 except that the SLM is below the wedge and it is transmissive.
  • the combination of Figure 85 can be used as a near-to-eye display device.
  • Figure 86 shows a cross section of a slab waveguide, wedge, component with micromirror array, and an SLM at entrance to the slab.
  • Figure 86 is similar to Figure 85 except that the SLM is at the input end of the slab waveguide.
  • the combination of Figure 86 can be used as a near-to-eye display device.
  • Figure 87 shows a cross section of a slab waveguide, wedge
  • SLM 110 is reflective, and micromirror array 8732 is transreflective.
  • Light first exiting wedge 8120 enters optical component 8730 and is reflected off micromirror array 8732 as a converging beam. The converging beam then passes back through wedge 8120 to be reflected and modulated by reflective SLM 110. The light reflected off SLM 110 passes back through optical component 8730, and converges at the exit pupil plane 220.
  • Figure 88 shows a cross section of slab waveguide with a 90 degree bend, wedge, optical component with a micromirror array, and an SLM.
  • the near-to-eye display device of Figure 88 is similar to the near-to-eye display device of Figure 85 with the exception that slab waveguide 8810 includes a 90 degree bend in Figure 88.
  • Light rays propagating in slab waveguide 8810 couple into wedge 8120 by means of a turning mirror 8820. Placing at least a portion of the slab waveguide perpendicular to the major axis of the wedge as shown in Figure 88 can reduce the form factor of the wedge-based near-to-eye display device.
  • a wedge-based eye tracker can be constructed as can be seen in Figure 89.
  • a near infrared (NIR) illumination provided by light source 8950 is coupled into the slab after passing through a beam splitter 8910 and the rays exit from the wedge.
  • a light turning film 8940 is placed on the wedge for directing the light beam towards the eye. The light reflected back from the eye is coupled back into the wedge 8120 and forms the image of the eye onto the camera, which can be used for eye tracking.
  • NIR near infrared
  • Camera 8930 is shown at the input to slab waveguide 8110 and coupled with a beamsplitter 8910. In some embodiments, camera 8930 is positioned along slab waveguide on either surface 8116 or 8118 similar to SLM 110 in Figure 83.
  • Figure 90 shows a near-to-eye display device with a slab waveguide, wedge, component with micromirror array, SLM, and camera for eye tracking.
  • Near-to-eye display device 9000 is in the form of a head-worn device, and more specifically in the shape of a pair of eyeglasses, but this is not a limitation of the present invention.
  • near-to-eye display device 9000 is a handheld device, and in other embodiments, near-to-eye display device 9000 is a fixed device that a user rests against to create a constant eye relief.
  • Near-to-eye display device 9000 includes slab waveguides 8810, wedges 8120, optical components 8130, optical components 9010, cameras 9020, and light sources 120. Near-to-eye display device 9000 also shows reflective SLM 110 on the slab waveguide 8810, although this is not a limitation of the present invention. Any SLM, either transmissive or reflective may be positioned anywhere as shown above in the previous figures without departing from the scope of the present invention. For example, in some embodiments, a reflective SLM is placed in optical component 9010, and in other embodiments, a transmissive SLM is placed at display area 9030.
  • near-to-eye display device 9000 is an augmented reality device that allows real world light to pass through optical components 9010, 8130, and wedge 8120.
  • the real world view is superimposed on any virtual scene created by the near-to-eye display device to create an augmented reality for the user of near-to-eye display device 9000.
  • near-to-eye display device 9000 includes electronics to provide SLM data to the SLMs.
  • the electronics may include a processor and memory, or may include cabling and transmission circuits to receive data from external sources.
  • the manner in which data is provided to the SLMs is not a limitation of the present invention.
  • Figure 91 shows a slab waveguide, a curved wedge and a compensation plate.
  • Apparatus 9100 includes slab waveguide 8110, curved wedge 9120, and curved compensation plate 9130.
  • Slab waveguide 8110 includes input end 8112, output end 8114, first surface 8118, and second surface 8116. As described above with reference to Figure 81, first surface 8118 and second surface 8116 are parallel to each other to cause light to propagate from input end 8112 to output end 8114 by total internal reflection.
  • Curved wedge 9120 is coupled to the output end 8114 of slab waveguide 8110.
  • Curved wedge 9120 includes first curved surface 9128 and second curved surface 9126 that form a continuously decreasing thickness.
  • curved wedge is constructed from a refractive material having a graded refractive index (GRIN).
  • GRIN graded refractive index
  • the curvature of curved wedge 9120 and the gradient of the refractive index in the GRIN material are selected such that light received from slab waveguide 81 10 exits curved wedge 9120 from curved surface 9128 as a converging beam that focuses on exit pupil plane 220.
  • Compensating wedge 9130 includes surface 9138 having substantially the same curvature as surface 9126, and is positioned to provide a uniform air gap 9140 between curved surface 9126 and surface 9138.
  • Compensating wedge 9130 has a wedge shape that complements the shape of wedge 9120 such that light traveling through both the curved wedge and the compensating wedge travel through an equivalent amount of like-refractive material. This eliminates any prism effect that would otherwise be perceived by a user. Undistorted real world views are provided because of the combination of the curved wedge and compensating wedge 9130.
  • a light beam enters slab waveguide 8110 at input end 8112 and propagates by total internal reflection within slab waveguide 8110 to output end 8114, at which point it enters wedge 9120.
  • the internal angle of incidence changes due to the decreasing thickness, and the critical angle changes due to the graded refractive index, allowing the light beam to exit curved surface 9128 of curved wedge 9120 as a converging wave that focuses onto exit pupil plane 220.
  • Figure 92 shows a slab waveguide, curved wedge, and SLM in a converging beam.
  • Apparatus 9200 includes slab waveguide 8110, curved wedge 9120 and transmissive SLM 410.
  • Transmissive SLM 410 is placed in the converging beam path and modulates the beam to create a virtual scene light wave distribution on exit pupil plane 220.
  • Apparatus 92 may also include a compensating wedge such as compensating wedge 9130 ( Figure 91).
  • Figure 93 shows a slab waveguide, curved wedge, and SLM on top of the slab.
  • the light field hits reflective SLM 110 while it is propagating in slab waveguide 8110.
  • SLM 110 is shown on surface 8116 of slab waveguide 8110 in Figure 93, this is not a limitation of the present invention.
  • the SLM is placed on surface 8118.
  • the computer-generated hologram on the SLM modulates the light as it propagates in slab waveguide 8110, and the desired virtual scene is generated at the useful portion of exit pupil plane 220 as described above.
  • Figure 94 shows a slab waveguide, curved wedge, and SLM at the entrance to the slab waveguide.
  • Figure 94 is similar to Figure 93 except that SLM 410 is at the input end of the slab waveguide, and SLM 410 is
  • Figure 95 shows a slab waveguide, curved wedge, and camera for eye tracking.
  • An NIR illumination provided by light source 8950 is coupled into the slab after passing through a beam splitter 8910. The operation is similar to that described with respect to Figure 89 in which the light reflected back from the eye is coupled back into the wedge 9120 and forms the image of the eye onto the camera, which can be used for eye tracking.
  • Camera 8930 is shown at the input to slab waveguide 8110 and coupled with a beamsplitter 8910. In some embodiments, camera 8930 is positioned along slab waveguide on either surface 8116 or 8118 similar to SLM 110 in Figure 83.
  • Figure 96 shows a perspective view of the apparatus of Figure 91.
  • Figure 91 shows light representing a real world view passing through both
  • the real world view may be superimposed on any modulated light distribution and presented at the exit pupil plane to form an augmented reality display.
  • Figure 97 shows a near-to-eye display device with a slab waveguide, curved wedge, compensating wedge, SLM, and camera for eye tracking.
  • Near- to-eye display device 9700 is in the form of a head-worn device, and more specifically in the shape of a pair of eyeglasses, but this is not a limitation of the present invention.
  • near-to-eye display device 9700 is a handheld device, and in other embodiments, near-to-eye display device 9700 is a fixed device that a user rests against to create a constant eye relief.
  • Near-to-eye display device 9700 includes slab waveguides 8810, curved wedges 9120, compensating wedges 9130, cameras 9020, and light sources 120. Near-to-eye display device 9700 is shown with slab waveguides 8810 including a 90 degree bend as described above with reference to Figures 88 and 93. Near- to-eye display device 9700 also shows reflective SLM 110 on the slab waveguide 8810, although this is not a limitation of the present invention. Any SLM, either transmissive or reflective may be positioned anywhere as shown above in the previous figures without departing from the scope of the present invention.
  • near-to-eye display device 9700 is a virtual reality device that blocks the real world view and provides the user with a virtual scene at the useful portion of the exit pupil plane.
  • near-to-eye display device 9700 is an augmented reality device that allows real world light to pass through the compensating wedge 9130 and the curved wedge 9120.
  • the real world view is superimposed on any virtual scene created by the near-to-eye display device to create an augmented reality for the user of near-to-eye display device 9700.
  • near-to-eye display device 9700 includes electronics to provide SLM data to the SLMs.
  • the electronics may include a processor and memory, or may include cabling and transmission circuits to receive data from external sources.
  • the manner in which data is provided to the SLMs is not a limitation of the present invention.
  • Figure 98 shows a near-to-eye display device with a moving platform assembly.
  • Near-to-eye display device 9800 includes moving platform assembly 9802 and electronics module 160.
  • Near-to-eye display device 9800 may include many more components such as wiring, cabling, camera, and the like. These components are intentionally omitted for clarity.
  • near-to-eye display device 9800 is shown with a moving platform assembly 9802 on only one side, whereas in practice, near-to-eye display device 9800 may have two moving bar assemblies 9802 - one on each side.
  • Moving platform assembly 9802 includes moving platform 9804 and coils 9840.
  • Moving platform 9804 includes LED array 9810, LED drivers 9820, and magnets 9830 for actuation.
  • LED drivers 9820 may be integrated circuits affixed to moving platform 9804.
  • LED drivers 9820 cause individual LEDs in LED array 9810 to be illuminated in response to electrical signals received from electronics module 160.
  • LED array 9810 may be a one dimensional array of red, green, and blue LEDs.
  • LED array 9810 may include one row of red LEDs, one row of green LEDs, and one row of blue LEDs.
  • LED array 9810 may be a two dimensional array of red, green, and blue LEDs.
  • LED array 9810 may include multiple rows of red LEDs, multiple rows of green LEDs, and multiple rows of blue LEDs.
  • moving platform 9804 moves vertically across a user's field of view.
  • Moving platform 9804 carries two permanent magnets 9830.
  • Two linear arrays of electromagnectic coils 9840 are attached to the moving platform assembly 9802 outside the display area. Current can be passed through any given subset of the coils 9840 to actuate moving platform 9804.
  • Electronics module 160 actuates moving platform 9804 and drives LED drivers 9820 synchronously such that a transparent image is created for a user.
  • moving platform assembly 9802 effectively creates an image on a transparent screen.
  • the area occupied by the transparent screen is referred to herein as the "display area.”
  • Figure 99 shows a cross section of moving platform assembly 9802 and a polarizing film 9910.
  • Moving platform 9804 is shown with a cross section of a one dimensional array of LEDs. Further, the actuation in the direction of the arrows is accomplished by energizing coils 9840 in sequence so that magnet 9830 is either attracted or repulsed. The timing of coil energizing is
  • Polarizing film 9910 is oriented such that environmental light viewed by a user of near-to-eye display device 9800 passes through polarizing film 9910, and further oriented such that light produced by the plurality of light sources does not pass through the polarizing film.
  • light from LED array 9810 is also polarized.
  • light passing through the polarizer is polarized in a first orientation and light emitted from the LEDs is polarized in a second orientation orthogonal to the first orientation.
  • polarizing film 9910 is omitted.
  • Figure 100 shows a perspective view of a moving platform assembly.
  • Moving platform assembly 9802 is shown with frame 11010, coils 9840 and moving platform 9804.
  • Frame 11010 and moving platform 9804 are shown interconnected by flex cable 11020.
  • Flex cable 11020 carries signals from electronics module 160 ( Figure 98) to LED drivers 9820 on moving platform 9804.
  • moving platform 9804 includes one moving bar that has an array of light sources mounted thereon.
  • Figure 101 shows a side view of a contact lens placed on an eye.
  • Contact lens 10100 includes two concentric portions, a peripheral portion 10120, and a central portion 10110.
  • Central portion 10110 has a high diopter lens to allow a user to focus at a plane of the plurality of light sources on moving platform 9804 when wearing near-to-eye display device 9800.
  • Peripheral portion 10120 of the contact lens admits only light polarized in a first orientation
  • central portion 10110 of the contact lens admits only light polarized in a second orientation, orthogonal to the first orientation.
  • central portion 10110 admits the polarized light emitted from LED array 9810
  • peripheral portion 10120 admits the polarized light that has passed through polarizing film 9910.
  • Figure 102 shows a front view of the contact lens of Figure 101.
  • Figure 102 shows three different variations of contact lens 10100.
  • Variation A has central portion 10110 split into two different parts; variation B has central portion 10110 split into three different parts; and variation C has central portion 10110 split into four different parts.
  • each different part of central portion 10110 has a different color filter to separate different color components of the light emanating from the plurality of light sources.
  • Figure 103 shows a cross section of a contact lens on an eye and a moving platform assembly.
  • Moving platform assembly 9802 includes moving platform 9804 which carries a plurality of light sources to form a transparent display for the user.
  • the light from the surroundings shown generally at 10310, is polarized in a first orientation by polarizing film 9910.
  • the light from the plurality of light sources is polarized in a second orientation, orthogonal to the first orientation.
  • the peripheral portion 10120 of the contact lens is constructed so that it only admits light with the first orientation.
  • the central portion 10110 of the contact lens is constructed so that it only admits light with the second orientation.
  • the central portion 101 10 of the contact lens is split into multiple parts, each having a separate color filter to separate different color components of the light emanating from the plurality of light sources.
  • the portion of the light from the plurality of light sources that passes through the high diopter lens in the central portion 10110 of the contact lens is properly focused in a user's eye. This allows a user to focus at a plane of the plurality of light sources.
  • the portion of the light from the surroundings that passes through the outer portion 10120 of the contact lens allows a user to see the surroundings with the user's normal eye sight.
  • Figure 104 shows a near-to-eye display device with a moving platform assembly.
  • near-to-eye display device 10400 includes moving platform assembly 10402, which in turn includes a moving platform with multiple bars.
  • the multiple moving bars move vertically together across the user's field of view as the moving platform moves.
  • Each bar may contain a one dimensional or two dimensional array of light sources.
  • Actuation is the same as described above with reference to Figures 98-100.
  • Figure 105 shows a perspective view of a near-to-eye display device with a rotating bar.
  • Rotating bar 10510 includes a plurality of light sources and rotates about pivot point 10520.
  • Rotating bar 10510 is actuated synchronously with signals that drive the light source to create an effective transparent display for the user.
  • Figures 106-108 show front views of near-to-eye display devices with rotating bars.
  • Figure 106 shows a front view of near-to-eye display device 10500 with the detail shown for the right eye rather than the left eye.
  • the rotating bar 10510 rotates about pivot point 10520 and sweeps through the display area shown at 10610.
  • the rotating bar includes a plurality of light sources as described above with reference to previous figures.
  • Figure 107 shows a front view of a near-to-eye display device with a rotating bar rotating around two pivot points.
  • the rotating bar 10710 carries a plurality of light sources.
  • a rotating arm 10720 is rotating around a first pivot point.
  • the rotating arm 10720 is connected to the rotating bar 10710 at the second pivot point.
  • the rotating bar 10710 is kept at a fixed orientation throughout the motion so that the display can make a more efficient use of the motion.
  • the dotted line outlines the potential display area.
  • Figure 108 shows front view of a near-to-eye display device with a moving bar moving vertically across a user's field of view.
  • the moving bar 10830 is actuated by a rotating arm 10820 that rotates around a pivot point
  • the rotating arm 10820 is attached to a groove on the moving bar 10830.
  • the rotating arm 10820 can move along the groove.
  • the moving bar 10830 is constrained by two mechanical guides 10810 to produce a vertical motion.
  • the moving bar 10830 carries a plurality of light sources. The dotted line outlines the potential display area.
  • Figure 109 shows a rotating bar actuation embodiment where a permanent magnet 10910 is placed inside of an electromagnetic coil 10930.
  • the permanent magnet 10910 is attached to the rotating bar 10510 and is suspended so that there is a pivot point 10520 inside of the electromagnetic coil 10930.
  • a small stabilization magnet 10920 is attached to the electromagnetic coil 10930 to keep the rotating bar 10510 stable when not actuated.
  • Figure 110 shows a rotating bar actuation embodiment where a permanent magnet 10910 is placed be ween two electromagnetic coils 11030.
  • the permanent magnet 10910 is attached to the rotating bar 10510 and is suspended so that the rotating bar 10510 will rotate around the pivot point 10520.
  • the rotating bar 10510 will rotate around the pivot point 10520.
  • the various embodiments of the present invention are not limited to magnetic actuation.
  • piezoelectric actuation is employed, and in other embodiments, actuation using a rotary or linear motor of any sort is employed.
  • Figure 111 shows a front view of a near-to-eye display device with a moving bar that moves in two dimensions.
  • the moving bar 11130 moves periodically in the vertical direction to form a transparent display for a user, and it simultaneously moves periodically a shorter distance in the horizontal direction.
  • the purpose of the horizontal motion is to increase the horizontal display resolution above the resolution dictated by the spacing of the light sources.
  • Figure 112 shows an external near-to-eye display device 11200 with no contact lens.
  • a moving bar (not shown) is moving across an otherwise transparent area in a near-to-eye display deviice.
  • the plurality of light sources is arranged so that light from the display can reach observers other than the user of the device. If an observer views the transparent display from a distance where the observer's eyes can focus on the content on the transparent display, the observer sees image 11210. Two examples of image 11210 are shown in Figure
  • image 11210 is generated with light sources that face away from the user of near-to-eye display device 11200, the user does not see image 11210.
  • Figure 113 shows a perspective view of a near-to-eye display device that includes a LED array.
  • Near-to-eye display device 1130 includes SLM 11320, infrared (IR) camera and light source 11302, and LED array 11310.
  • Near-to-eye display device 11300 may also include additional components, such as an electronics module, battery, cabling, and the like. These additional components are intentionally omitted from the figure so as to not obscure the components that are shown.
  • near-to-eye display device 11300 like many other near-to- eye display devices depicted herein, shows most components for only one side (one eye) of the device. In some embodiments, all components are duplicated and mirrored to create a near-to-eye display device for both eyes.
  • the IR light sources are used to illuminate a user's pupils and the cameras are used to detect the position of the user's pupils.
  • the cameras are positioned on the frame as shown in Figure
  • IR light sources are co-located with the cameras, although this is not a limitation of the present invention.
  • IR light sources are co-located with point light sources used to illuminate an SLM.
  • an IR light source may be co-located with LED array 11310.
  • the user's eyes are illuminated with infrared light, which is not visible to the user.
  • the cameras capture infrared images of the user's eyes, and existing computer vision, pattern recognition, and image processing algorithms are used to detect the pupil positions.
  • LED array 11310 includes a two dimensional array of color light sources, where each light source includes a red, a green, and a blue LED.
  • LED array 11310 also includes LED drivers 11410. When different LEDs are selected to provide light to illuminate SLM 11320, the resulting virtual scene wave moves slightly on the exit pupil plane. As described below, this phenomenon is exploited to steer the useful portion of the exit pupil plane to follow eye motion.
  • Figures 115 and 116 show a top view of pupil tracking using multiple
  • Figures 115 and 116 depict the salient components from Figure 113, and are not necessarily to scale. Further, Figures 115 and 116 show a one dimensional array of three LEDs for simplicity, however in a practical system many more LEDs may be used, and a two dimensional array such as that shown in Figure 114 may be used.
  • SLM 11320 is a stationary SLM that includes a reflector 11510 to reflect modulated light as a converging beam.
  • SLM 11320 is a transmissive SLM in a converging or diverging light path.
  • SLM 11320 is a reflective SLM in a converging or diverging light path.
  • SLM 11320 may be oriented as shown in any of Figures 17- 28.
  • Figure 115 represents the case in which the user is looking straight ahead, and the center LED is turned and used as the point light source to illuminate the SLM.
  • Figure 116 represents the case in which the user has moved her eye to look a few degrees to the right.
  • Pupil tracker 11610 detects the new pupil position and commands LED driver 11410 to use a different LED to illuminate the SLM so that the useful portion of the exit pupil plane follows the user's pupil.
  • Pupil tracker 11610 may include light sources, cameras, a processor, instructions stored in a memory, and many other components.
  • pupil tracker 1160 is a combination of components, that when taken together, functions to track the position of the user's pupil. As the user's pupil is tracked, pupil tracker 11610 takes one or more actions to steer the useful portion of the exit pupil plane to follow the user's pupil. In the case of near-to- eye display device 11300, pupil tracker 11610 commands different LEDs to illuminate the SLM to steer the useful portion of the exit pupil plane to track the user's pupils.
  • Figure 117 shows a near-to-eye display device that includes a rotating SLM.
  • Near-to-eye display device 11700 includes rotating SLM 11720, actuator 11710, camera 11302, and point light source 120.
  • Actuator 11710 when actuated, causes SLM 11720 to rotate.
  • actuator 11710 may be a stepper motor or a like device capable of controlling the amount of rotation of the SLM.
  • actuator 11710 is commanded to operate by an electronic module (not shown) that is part of a pupil tracker such as pupil tracker 11610.
  • FIGS 118 and 119 show a top view of pupil tracking using a rotating SLM.
  • the views in Figures 118 and 119 depict the salient components from Figure 117, and are not necessarily to scale.
  • Rotating SLM 11720 includes a reflector 11510 to reflect modulated light as a converging beam.
  • SLM 11720 is a transmissive SLM in a converging or diverging light path.
  • SLM 11720 is a reflective SLM in a converging or diverging light path.
  • SLM 11720 may be oriented as shown in any of Figures 17-28.
  • Figure 118 represents the case in which the user is looking straight ahead, and the rotating SLM 11720 is oriented so that the useful portion of the exit pupil plane overlaps the user's pupil.
  • Figure 119 represents the case in which the user has moved her eye to look a few degrees to the left.
  • Pupil tracker 11610 detects the new pupil position and commands actuator 11710 to rotate SLM 11720 so that the useful portion of the exit pupil plane follows the user's pupil.
  • pupil tracker 11610 may take many forms, and many take any appropriate action to ensure that the useful portion of the exit pupil plane tracks the user's pupil.
  • pupil tracker 11610 commands an actuator to rotate the SLM to steer the useful portion of the exit pupil plane to track the user's pupils.
  • Figure 120 shows a perspective view of a near-to-eye display device that includes rotating SLMs and LED arrays.
  • Near-to-eye display device 12000 includes an LED array 11310 and rotating SLM 11720 with actuator 11710.
  • Near-to-eye display device 12000 may rotate the SLM and select different LEDs in any combination to steer the useful portion of the exit pupil plane to the location of the user's pupil.
  • Figures 121 and 122 show flowcharts of methods in accordance with various embodiments of the invention. In some embodiments, the methods of Figures 121 and 122, or portions thereof, are performed by a near-to-eye display device, embodiments of which are shown in previous figures. In other embodiments, the methods are performed by a computer or an electronic system. The various actions in the methods may be performed in the order presented, or may be performed in a different order. Further, in some embodiments, some actions listed in Figures 121 and 122 are omitted.
  • Figure 121 shows a flowchart showing rotation for small angles and LED selection for larger angles.
  • a user's pupil is tracked. In some embodiments, this corresponds to pupil tracker 11610 tracking the position of a user's pupil.
  • a rotatable SLM When a user moves her eye, the eye rotates and the pupil moves through an angle. When the pupil moves through a small angle, a rotatable SLM is rotated to steer the useful portion of the exit pupil plane to the location of the user's pupil at 12120. For larger angles, a different light source is selected to steer the useful portion of the exit pupil plane to the location of the user's pupil at 12130. This process is repeated as the user moves her eye and it is tracked by the near-to-eye display device.
  • Figure 122 shows a flowchart showing rotation for small angles and diffraction order selection for larger angles.
  • a user's pupil is tracked. In some embodiments, this corresponds to pupil tracker 11610 tracking the position of a user's pupil.
  • a rotatable SLM When the pupil moves through a small angle, a rotatable SLM is rotated to steer the useful portion of the exit pupil plane to the location of the user's pupil at 12120. For larger angles, the light wave distribution is recomputed such that a higher diffraction order moves into the useful portion of the exit pupil plane at 12230. This process is repeated as the user moves her eye and it is tracked by the near-to-eye display device.
  • a user's pupil is tracked. In some embodiments, this corresponds to pupil tracker 11610 tracking the position of a user's pupil.
  • a rotatable SLM When a user moves her eye, the eye rotates and the pupil moves through an angle. When the pupil moves through a small angle, a rotatable SLM is rotated to steer the useful portion of the exit pupil plane to the location of the user's pupil at 12120. For larger angles, a different light source is selected to steer the useful portion of the exit pupil plane to the location of the user's pupil at 12130. This process is repeated as the user moves her eye and it is tracked by the near-to-eye display device.
  • Figure 123 shows a near-to-eye display device that includes an active grating.
  • Near-to-eye display device 12300 includes SLM with active grating 12320, actuator 12310, camera 11302, and point light source 120.
  • Actuator 12310 when actuated, causes an active grating within SLM 12320 to change its optical qualities.
  • actuator 12310 may be a driver circuit capable of controlling a voltage applied to the active grating.
  • actuator 12310 is commanded to operate by an electronic module (not shown) that is part of a pupil tracker such as pupil tracker 11610.
  • Figures 124 and 125 show a top view of pupil tracking using an SLM with an active grating.
  • SLM with active grating 12320 includes active grating 12410, and a reflector 11510 to reflect modulated light as a converging beam.
  • active grating 12410 is a custom liquid crystal based device that implements a multi-section prism.
  • Active grating 12410 may be an LC device that merely contain electrodes and no pixels.
  • the SLM, active grating, and reflector are separate devices.
  • SLM 12320 may be a transmissive SLM in a converging or diverging light path. Further, in some embodiments, SLM 12320 is a reflective SLM in a converging or diverging light path. For example, SLM 12320 may be oriented as shown in any of Figures 17-28.
  • Figure 124 represents the case in which the user is looking straight ahead, and active grating 12410 is controlled so that the useful portion of the exit pupil plane overlaps the user's pupil.
  • Figure 125 represents the case in which the user has moved her eye to look a few degrees to the right.
  • Pupil tracker 11610 detects the new pupil position and commands actuator 12310 to energize active grating 12410 so that the useful portion of the exit pupil plane follows the user's pupil.
  • pupil tracker 11610 may take many forms, and many take any appropriate action to ensure that the useful portion of the exit pupil plane tracks the user's pupil.
  • pupil tracker 11610 commands an actuator to energize an active grating to steer the useful portion of the exit pupil plane to track the user's pupils.
  • Figure 126 shows a perspective view of a near-to-eye display device that includes an SLM with an active grating and LED arrays.
  • Near-to-eye display device 12600 includes an LED array 11310 and SLM with active grating 12320 with actuator 12310.
  • Near-to-eye display device 12600 may energize the active grating and select different LEDs in any combination to steer the useful portion of the exit pupil plane to the location of the user's pupil.
  • Figure 127 is provided in Figure 127.
  • Figures 127 and 128 show flowcharts of methods in accordance with various embodiments of the invention.
  • the methods of Figures 127 and 128, or portions thereof are performed by a near-to-eye display device, embodiments of which are shown in previous figures.
  • the methods are performed by a computer or an electronic system.
  • the various actions in the methods may be performed in the order presented, or may be performed in a different order. Further, in some embodiments, some actions listed in Figures 127 and 128 are omitted.
  • Figure 127 shows a flowchart showing grating actuation for small angles and LED selection for larger angles.
  • a user's pupil is tracked. In some embodiments, this corresponds to pupil tracker 11610 tracking the position of a user's pupil.
  • the eye rotates and the pupil moves through an angle.
  • an active grating is actuated to steer the useful portion of the exit pupil plane to the location of the user's pupil at 12720.
  • a different light source is selected to steer the useful portion of the exit pupil plane to the location of the user's pupil at 12730. This process is repeated as the user moves her eye and it is tracked by the near-to-eye display device.
  • Figure 128 shows a flowchart showing grating actuation for small angles and diffraction order selection for larger angles.
  • a user's pupil is tracked. In some embodiments, this corresponds to pupil tracker 11610 tracking the position of a user's pupil.
  • an active grating is energized to steer the useful portion of the exit pupil plane to the location of the user's pupil at 12720.
  • the light wave distribution is recomputed such that a higher diffraction order moves into the useful portion of the exit pupil plane at 12830. This process is repeated as the user moves her eye and it is tracked by the near-to-eye display device.
  • Figures 129 and 130 show augmented reality views demonstrating a virtual scene at different depths.
  • the views in Figures 129 and 130 represent what a user of a near-to-eye display device might see out of one eye at different accommodations.
  • the real world view includes objects in a foreground 12820, and objects in a background 12940.
  • the user's accommodation is set to focus on the foreground, hence the real world foreground 12920 is shown in focus, and the real world background 12940 is shown slightly out of focus.
  • Figure 129 also shows a virtual scene that is superimposed on the real world view.
  • the virtual scene includes three objects: virtual object 12910, virtual object 12930, and virtual object 12950. These virtual objects are simply text, however virtual objects can be anything, and are not limited to text.
  • virtual object 12910 was set at a depth corresponding to the depth of the real-world foreground
  • virtual object 12930 was set at a depth corresponding to the depth of the real world background.
  • both virtual objects 12920 and 12930 are reconstructed over the entire useful portion of the exit pupil plane.
  • Figure 130 shows the same real world view and superimposed virtual scene as Figure 129. The only difference is now the user has accommodated to the depth of the real- world background. As a result both the real world background 12940 and the virtual object 12930 are in focus, and both the real world foreground 12920 and the virtual object 12910 are not in focus.
  • virtual object 12950 is always in focus regardless of the user's accommodation. This is because virtual object 1250 is reconstructed over a smaller region of the useful portion of the exit pupil plane, thereby increasing the depth of field.
  • the virtual scene is computed in such a way that virtual object 12950 only overlaps a one mm section of the pupil.
  • Figures 129 and 130 are an example of an SLM being programmed to display virtual objects appearing at different depths while some objects appear focused in all depths (stay in focus even if the eye accommodates to a different depth). Waves from a first plurality of subsections of the displayed virtual scene are reconstructed over the entire useful portion so that each of the first plurality of subsections appears focused on the retina only when the user accommodates to the depth of that subsection, and waves from a second subsection of the displayed virtual scene are reconstructed over smaller regions of the useful portion so that these parts always appear focused on the retina.
  • the techniques demonstrated in Figures 129 and 130 are combined with binocular disparity to provide realistic 3D visual experiences without causing visual fatigue due to the accommodation- convergence conflict.
  • eyes converge to the apparent position of a virtual 3D object and accommodation of each eye is also set for the depth corresponding to the apparent position of the virtual 3D object. This results in "natural 3D" in which the accommodation-convergence conflict is greatly reduced if not completely eliminated, providing a very comfortable 3D experience for the user.
  • Figure 131 shows a block diagram of a near-to-eye display device in accordance with various embodiments of the present invention.
  • Near-to-eye display device 13100 includes processor 13102, memory 13110, light sources 13160, SLMs 13162, light bars 13164, cameras 13166, actuators 13168, transducers 13170, global positioning system (GPS) receiver 13172,
  • GPS global positioning system
  • Near-to-eye display device 13100 may be any near-to-eye display device described herein.
  • mobile device 300 may be a near-to-eye display device that performs pupil filtering, pupil tracking, speckle reduction, or any other function described herein.
  • Processor 13102 may be any type of processor capable of executing instructions store in memory 13110 and capable of interfacing with the various components shown in Figure 131.
  • processor 13102 may be a microprocessor, a digital signal processor, an application specific processor, or the like.
  • processor 13102 is a component within a larger integrated circuit such as a system on chip (SOC) application specific integrated circuit (ASIC).
  • SOC system on chip
  • ASIC application specific integrated circuit
  • Light sources 13160 may include any type of light source capable of illuminating an SLM. Examples include point light source 120 ( Figure 1), illumination optics module 440 ( Figure 4), and the array of point light sources shown in Figures 15 and 16. In operation, processor 13102 may command light sources 13160 to turn on and off.
  • SLMs 13162 are SLMs that impart information to an illumination wave to create the desired light wave distribution in the useful portion of the exit pupil plane.
  • processor 13102 programs SLMs 13162 using data stored in memory 13110.
  • processor 13102 computes the SLM data to be displayed on the SLM and stores it in memory 13110.
  • the SLM data is computed by a separate device, and the SLM data is provided to near-to-eye display device 13100 for later display.
  • Light bars 13164 include any of the light bar and/or moving platform embodiments described herein.
  • processor 13102 may command an actuator to cause one or more light bar to move. Further processor 13102 may also command one or more light sources on a light bar to illuminate.
  • Cameras 13166 may be any type of camera capable of capturing an image and providing the image data to processor 13102.
  • cameras 13166 are cameras used for calibration, and in other embodiments, cameras 13166 are cameras used for pupil tracking.
  • Actuators 13168 are devices that convert one form of energy to another.
  • actuators 13168 may include stepper motors, magnets, electrical coils, and the like.
  • Actuators 13168 may include any of the actuator
  • Transducers 13170 are devices that convert energy from one form to electricity.
  • adjustment knob 4510 ( Figure 45) is an example of a transducer.
  • processor 13102 receives electronic signals when a user interacts with any of transducers 13170.
  • GPS 13172 includes a GPS receiver.
  • processor 13102 receives fine location data from GPS 13172. In some embodiments, this data is used to generate SLM data or to determine what stored SLM data should be displayed. For example, in embodiments represented Figures 120 and 130, GPS data may be used to determine what virtual objects should be included in the virtual scene.
  • Accelerometers 13174 are devices that measure rate of change of motion or the direction of forces applied to near-to-eye display device 13100 due to gravity.
  • processor 13102 receives accelerometer data when near- to-eye display device 13100 is moved or its orientation is changed.
  • Compass 13176 is a device that measures the orientation of near-to-eye display device 13100 relative to magnetic north.
  • processor 13102 receives data from compass 13176 that represents the orientation of near-to-eye display device 13100 with respect to magnetic north.
  • Radios 13178 may include any type of radio that can provide
  • radio 13178 may be a cellular radio, a Bluetooth radio, a NFC radio, a WiFi radio, or the like.
  • GPU 13180 is a device that can accelerate some computations performed during the generation of SLM data.
  • GPU 13180 maybe used to render a virtual scene represented by polygon mesh models.
  • Gyroscopes 13182 provide high resolution data regarding movement of near-to-eye display device.
  • processor 13102 may make use of data provided by gyroscopes 13182 for head tracking applications.
  • Touchscreen 13184 allows user interaction with the display surfaces of near-to-eye display device 13100.
  • An example near-to-eye display device with a touchscreen interface is described below with reference to Figure 132.
  • Touchscreen 13184 is a device that includes a touch sensitive surface, sensor, or set of sensors that accept input from a user. For example, touchscreen 13184 may detect when and where an object touches the screen, and may also detect movement of an object across the screen. Touchscreen 13184 may be manufactured using any applicable display technologies, including for example, liquid crystal display (LCD), active matrix organic light emitting diode
  • touchscreen 13184 may be manufactured using any application touch sensitive input technologies, including for example, capacitive and resistive touch screen technologies, as well as other proximity sensor technologies.
  • Audio circuits 13186 provide an audio interface (input, output, or both) between processor 13102 and a user.
  • one or more applications make use of audio circuits 13186 to provide a multi-sensory experience.
  • tour guide application 13143 may provide interpretive audio as well as an immersive 3D augmented reality experience.
  • audio circuits 13186 include a microphone that allows a user to record audio or to provide audio commands to near-to-eye display device 13100.
  • Memory 13110 may include any type of memory device.
  • memory 13110 may include volatile memory such as static random access memory (SRAM), or nonvolatile memory such as FLASH memory.
  • Memory 13110 is encoded with (or has stored therein) one or more software modules (or sets of instructions), that when accessed by processor 113102, result in processor 13102 performing various functions.
  • the software modules stored in memory 13110 may include an operating system (OS) 13120, near-to-eye modules 13130 and applications 13140.
  • Applications 13140 may include any number or type of applications. Examples provided in Figure 131 include games 13141, maps 13142, a tour guide app 13143, and a video player. An example display from a tour guide app is described above with reference to Figures 129 and 130.
  • Memory 13110 may also include any amount of space dedicated to data storage 13150.
  • Operating system 13120 may be any to of operating system such as an operating system to control a mobile phone, tablet computer, embedded system, or the like. As shown in Figure 131, operating system 13120 includes user interface component 13121 and application installer component 13122.
  • Operating system 13120 may include many other components without departing from the scope of the present invention.
  • User interface component 13121 includes processor instructions that cause near-to-eye display device 13100 to display user interaction components, such as dialog boxes, alerts, and prompts. User interface 13121 also includes instructions to display menus, move icons, and manage other portions of the display environment.
  • Application installer component 13122 installs applications to near-to- eye display device 13100. Any type or number of applications may be installed.
  • Example apps currently installed on near-to-eye display device include games 13141, maps 13142, tour guide app 13143, and video player app 13144.
  • Near-to-eye modules 13130 include calibration 13131, SLM computation 13132, pupil tracking 13133, and speckle reduction 13134.
  • Calibration module 13131 includes instructions that cause processor 13102 to perform calibration embodiments described herein.
  • calibration module 13131 may cause processor 13102 to capture images using cameras 13166, and interact with the user using user interface 13121 and transducers 13170.
  • SLM computation module includes instructions to perform the computations described above with reference to Figure 56.
  • the near-to-eye modules shown in Figure 131 are meant as examples only; many more near-to-eye modules may be included without departing from the scope of the present invention. In general, any method described herein may include a module component within near-to-eye modules 13130.
  • Pupil tracking module 13133 includes instructions that when executed by processor 13102 cause near-to-eye display device 13100 to steer the useful portion of the exit pupil plane to follow a user's pupils.
  • the combination of pupil tracking modulel3133, processor 13102, cameras 13166, and light sources 13160 (for IR light) make up pupil tracker 11610 described above.
  • Speckle reduction module 13134 includes instruction that when executed by processor 13102 causes a virtual scene to be computed with assigned phase terms that reduce speckle.
  • Data storage 13150 stores data that does not include processor instructions.
  • SLM data 13151 is stored in data storage 13150, as are user profiles.
  • SLM data 13151 includes still images, and in other embodiments, SLM data 13151 includes many frames that form video data.
  • SLM data 13151 may represent 2D or 3D virtual scenes used for either or both of virtual reality display applications or augmented reality applications.
  • Each of the above-identified applications and modules correspond to a set of instructions for performing one or more functions described above. These applications (sets of instructions) need not be implemented as separate software programs, procedures or modules, and thus various subsets of these applications may be combined or otherwise re-arranged in various embodiments. For example, SLM computation 13132 may be combined with speckle reduction 13134. Furthermore, memory 13110 may store additional applications (e.g., audio players, camera applications, etc.) and data structures not described above.
  • device 13100 is presented as an example of a near-to-eye display device, and that device 13100 may have more or fewer components than shown, may combine two or more components, or may have a different configuration or arrangement of components.
  • device 13100 may include many more components such as sensors (optical, touch, proximity etc.), or any other components suitable for use in a near-to-eye display device.
  • Memory 13110 represents a computer-readable medium capable of storing instructions, that when accessed by processor 13102, result in the processor performing as described herein. For example, when processor 13102 accesses instructions within pupil tracking module 13133, processor 13102 analyzes images of a user's eyes, determines the pupil location, and then steers the useful portion of the exit pupil plan to overlap with the user's pupil.
  • Figure 132 shows a near-to-eye display device with transparent touch sensitive layers 13210.
  • the front surfaces of the near-to- eye display device are covered with transparent touch sensitive layers that allow for user interaction.
  • a user using near-to-eye display device 13200 can use her fingers to make selections among displayed items 13220 (e.g., some icons/menu items) or to perform actions such as zoom in and out operations, and input text data through virtual keyboards, similar to the usage of touch sensitive screens on existing smart phones, tablets, etc, with the difference that the user sees the displayed content through the backside of the display, while she performs the finger touch based input operations through the front side.
  • displayed items 13220 e.g., some icons/menu items
  • actions such as zoom in and out operations
  • input text data through virtual keyboards
  • a near-to-eye display device comprising:
  • SLM spatial light modulator
  • light produced by the at least one point light source illuminates the SLM and gets modulated to produce modulated light, and the modulated light is directed on an exit pupil plane that includes a useful portion, and wherein a light wave distribution within the useful portion is equal to a computed light distribution from a virtual scene;
  • the useful portion is steerable across the exit pupil plane to follow the motion of a user's eye pupil when the near-to-eye display device is in use so that the user's eye pupil acts as a spatial filter to filter out undesired beams produced by the SLM at the exit pupil plane.
  • a near-to-eye display device comprising:
  • SLM spatial light modulator
  • light produced by the at least one point light source illuminates the SLM and gets modulated to produce modulated light, and the modulated light is directed on an exit pupil plane that includes a useful portion, and wherein a light wave distribution within the useful portion is equal to a computed light distribution from a virtual scene;
  • the useful portion is steerable to an expected location of a user's eye pupil when the near-to-eye display device is in use so that the user's eye pupil acts as a spatial filter to filter out undesired beams produced by the SLM at the exit pupil plane.
  • a near-to-eye display device comprising:
  • SLM spatial light modulator
  • light produced by the at least one point light source illuminates the SLM and gets modulated to produce modulated light, and the modulated light is directed on an exit pupil plane that includes a useful portion, and wherein a light wave distribution within the useful portion is equal to a computed light distribution from a virtual scene;
  • a near-to-eye display device comprising:
  • SLM spatial light modulator
  • light produced by the at least one point light source illuminates the SLM and gets modulated to produce modulated light, and the modulated light is directed on an exit pupil plane that includes a useful portion, and wherein a light wave distribution within the useful portion is equal to a computed light distribution from a virtual scene;
  • a near-to-eye display device comprising:
  • SLM spatial light modulator
  • light produced by the at least one point light source illuminates the SLM and gets modulated to produce modulated light, and the modulated light is directed on an exit pupil plane that includes a useful portion, and wherein a light wave distribution within the useful portion is equal to a computed light distribution from a virtual scene;
  • a near-to-eye display device comprising:
  • SLM spatial light modulator
  • At least one point light source at least one spatial light modulator (SLM) mounted on the near-to-eye display device;
  • SLM spatial light modulator
  • SLM spatial light modulator
  • light produced by the at least one point light source illuminates the SLM and gets modulated to produce modulated light, and the modulated light is directed on an exit pupil plane that includes a useful portion, and wherein a light wave distribution within the useful portion is equal to a computed light distribution from a virtual scene;
  • the SLM data is determined using a computation that compensates for optical aberrations of a light path from the at least one point light source to the exit pupil plane.
  • a near-to-eye display device comprising:
  • the user's eye pupil acts as a spatial filter to filter out undesired beams produced by the SLM at the exit pupil plane;
  • 1 B 1.
  • a near-to-eye display device comprising :
  • a spatial light modulator illuminated by the array of point light sources in a time sequential manner, the spatial light modulator having a plurality of sections that project diverging light toward an exit pupil plane positioned at an expected location of a user's eye pupil when the near-to-eye display device is in use;
  • spatial light modulator and the array of point light sources are positioned such that each of the plurality of sections contributes to the light wave in the useful portion of the exit pupil plane with the highest optical power when the corresponding point light source of the array is turned on.
  • the near-to-eye display device of 1B1 wherein the array of point light sources comprises a plurality of groups of point light sources, with more than one point light source in a group, and the point light sources within each of the plurality of groups can be turned on at the same time.
  • the near-to-eye display device of 1 B 1 wherein the near-to-eye display device comprises a head- worn device.
  • a near-to-eye display device comprising:
  • the near-to-eye display device of 1B4 further comprising a second array of point light sources with restricted emission cones, wherein the array of point light sources and the second array of point light sources partition the SLM differently with nonoverlapping borders, and wherein the array of point light sources and the second array of point light sources are turned on in a time sequential manner
  • a method comprising:
  • determining a plurality of data sets to be programmed in a spatial light modulator (SLM) in a near-to-eye display device that includes an array of point light sources, wherein for a video frame of a virtual scene, a different data set for each of the point light sources in the array is computed; and
  • SLM spatial light modulator
  • a method comprising:
  • SLM spatial light modulator
  • a microdisplay positioned on the near-to-eye display device to generate on a user's retina a defocused peripheral image that surrounds a focused image generated by the spatial light modulator.
  • a near-to-eye display device comprising:
  • SLM spatial light modulator
  • 1C5. The near-to-eye display device of any of 1C 1-1 C2, wherein the modulated light is steerable across the exit pupil plane to follow the motion of a user's eye pupil when the near-to-eye display device is in use.
  • IC 10 The near-to-eye display device of any of 1 C 1 - 1 C2 wherein the SLM includes at least one row of pixels.
  • IC 11 The near-to-eye display device of any of lCl-lC2 wherein the SLM has a vertical dimension of at least 2mm.
  • IC 13 The near-to-eye display device of any of 1 C 1 - 1 C2 wherein the SLM presents a horizontal field of view of about 40 degrees. 1 C 14. The near-to-eye display device of any of lCl-lC2 wherein the microdisplay is mounted on a movable platform.
  • the near-to-eye display device of any of lCl-lC2 wherein the microdisplay can be selected from an organic light emitting diode (OLED) display, a transmissive liquid crystal display (LCD), or a reflective LCD.
  • OLED organic light emitting diode
  • LCD transmissive liquid crystal display
  • reflective LCD reflective LCD
  • a near-to-eye display device comprising:
  • a spatial light modulator capable of modulating reflected light or displaying color pixels
  • a pupil tracking device to track a user's pupil position
  • a spatial light modulator driver circuit responsive to the pupil tracking device to cause the spatial light modulator to modulate reflected light in a central region of the user's gaze and to display color pixels away from the central region of the user's gaze.
  • 1 C 19 The near-to-eye display device of 1 C 18, wherein the near-to-eye display device comprises a head- worn device.
  • a method comprising:
  • a method comprising:
  • 1D8 The method of 1D7 wherein the transducer comprises an adjustment knob. 1 D9. The method of any of lDl-lD3, wherein the user selects an image and then interacts with a transducer to provide feedback.
  • 1D11 The method of any of 1D1-1D3, wherein the feedback from the user is used to adjust for eye relief variations.
  • ID 16 The near-to-eye display device of 1D15 wherein the transducer comprises an adjustment knob.
  • the display calibration component modifies phase values of the light distribution.
  • At least one point light source At least one point light source
  • At least one spatial light modulator mounted on the near-to-eye display device, wherein light produced by the at least one point light source illuminates the SLM and gets modulated to produce modulated light, and the modulated light is directed on an exit pupil plane that includes a useful portion, and wherein a light wave distribution within the useful portion is equal to a computed light distribution from a virtual scene;
  • SLM spatial light modulator
  • a camera to modify data patterns presented to the SLM based on images captured by the camera.
  • a near-to-eye display device comprising:
  • a display calibration component coupled to the at least one actuator to modify the physical characteristics of the near-to-eye display device based on images captured by the camera.
  • the method of 1E8 wherein recording images of a user's eyes comprises recording the images using two cameras.
  • IE 10. The method of 1E8 wherein analyzing the recording images for user characteristics comprises recording a position of the user's pupils.
  • determining a two dimensional complex valued profile of a virtual scene on a useful portion of an exit pupil plane comprises using a point cloud object model to represent a virtual object as a plurality of point light sources.
  • each element of the matrix is associated with a specific angular location on the sphere, and each element is filled with the complex amplitude of the point source at that angular location on the sphere;
  • the method of IF 1 wherein back-propagating comprises incorporating free space propagation and wave optics models of any optical components including aberrations introduced by the components between the spatial light modulator and the useful portion of the exit pupil plane.
  • determining a two dimensional complex valued profile of an illumination wave comprises performing free space propagation and wave optics analysis for components within an illumination module including aberrations.
  • encoding comprises encoding as a phase- only hologram.
  • encoding comprises encoding an an amplitude-only hologram.
  • the size of the useful portion is taken to be smaller than 2mm, a typical value being 1 mm, so that a sharp retinal image of these objects can also be delivered to the retina using the pinhole imaging principle.
  • the images have infinite depth of focus.
  • the method of 1 Gl wherein assigning a phase value to each of the plurality of virtual scene points to reduce speckle comprises assigning phase values to produce a smoothly interpolated version of a plurality of points on a user's retina.
  • the method of 1G1 wherein assigning a phase value to each of the plurality of virtual scene points to reduce speckle comprises assigning phase values to make optical paths from the virtual scene points to the retina differ by integer multiples of a center wavelength of the light source.
  • the method of 1 Gl wherein assigning a phase value to each of the plurality of virtual scene points to reduce speckle comprises assigning phase values to make optical paths from the plurality of virtual scene points to the pupil differ by integer multiples of a center wavelength of the light source.
  • the method of 1G1 wherein determining a two dimensional complex valued profile of an illumination wave comprises performing wave optics analysis for components within an illumination module.
  • An apparatus to create a coherent light beam comprising:
  • a transparent substrate having a face and an embedded light scattering apparatus
  • a light guiding apparatus positioned within the substrate to receive light from outside the substrate and guide the light to the embedded light scattering apparatus;
  • An apparatus that includes a near-to-eye display device comprising: at least one point light source;
  • a transparent substrate having a first face through which a coherent light beam emanates
  • a light scattering apparatus embedded in the substrate to receive light from the at least one point light source and scatter light away from the first face; a reflective optical element to reflect the light from the scattering apparatus to the first face to create the coherent light beam;
  • a light guide within the substrate to guide light from the point light source to the scattering apparatus.
  • OLED organic light emitting diode
  • An apparatus in accordance with any of 2A1-2A27 further comprising a point light source to provide light to the light guiding apparatus.
  • a point light source to provide light to the light guiding apparatus.
  • the at least one point light source comprises a red light source, a green light source, and a blue light source.
  • a light guide within the transparent substrate to guide light from the at least one point light source to the light scattering apparatus.
  • An apparatus comprising :
  • a wedge coupled to receive light from the output end of the slab waveguide, the wedge having a first surface, and a slanted surface nonparallel to the first surface of the wedge to form a continuously decreasing thickness to cause light to exit the wedge from the slanted surface;
  • an optical component having a face parallel to the slanted surface of the wedge, the optical component including a micromirror array to reflect light received through the face back through the wedge.
  • the apparatus of 2B1 further comprising a spatial light modulator positioned on the first surface of the slab waveguide to modulate the light as it propagates by total internal reflection.
  • the apparatus of 2B1 further comprising a spatial light modulator positioned between the wedge and the micromirror array to modulate the light after leaving the slanted surface.
  • the apparatus of 2B1 further comprising a camera for eye tracking.
  • a compensating wedge that includes a micromirror array to reflect light exiting the first wedge.
  • the compensating wedge includes a second slanted surface parallel to the first slanted surface.
  • the apparatus of 2B9 further comprising a point light source to provide light to the input end. 2B11.
  • the apparatus of 2B7 further comprising a spatial light modulator positioned between the first wedge and the micromirror array to modulate the light after leaving the first slanted surface.
  • the apparatus of 2B7 further comprising a camera for eye tracking.
  • a near-to-eye display device comprising:
  • a slab waveguide having an input end, an output end, and first and second surfaces parallel to each other to cause light received from the point light source to propagate from the input end to the output end by total internal reflection; a wedge coupled to receive light from the output end of the slab waveguide, the wedge having a first surface, and a slanted surface nonparallel to the first surface of the wedge to form a continuously decreasing thickness to cause light to exit the wedge from the slanted surface;
  • an optical component having a face parallel to the slanted surface of the wedge, the optical component including a micromirror array to reflect light received through the face back through the wedge to create a converging light beam;
  • a spatial light modulator illuminated by the converging light beam, wherein the spatial light modulator is not in an optical conjugate plane to a retina of a user using the near-to-eye display device.
  • the near-to-eye display device of 2B14 further comprising a spatial light modulator positioned on the first surface of the slab waveguide to modulate the light as it propagates by total internal reflection.
  • the near-to-eye display device of 2B14 further comprising a spatial light modulator positioned between the wedge and the micromirror array to modulate the light after leaving the slanted surface.
  • the near-to-eye display device of 2B14 further comprising a camera for eye tracking.
  • An apparatus comprising :
  • a slab waveguide having an input end, an output end, and first and second surfaces parallel to each other to cause light to propagate from the input end to the output end by total internal reflection;
  • a curved wedge coupled to receive light from the output end of the slab waveguide, the curved wedge having a continuously decreasing thickness to cause light to exit the wedge from one of two surfaces.
  • the apparatus of 2C1 further comprising a spatial light modulator positioned on the first surface of the slab waveguide to modulate the light as it propagates by total internal reflection.
  • the apparatus of 2C 1 further comprising a camera for eye tracking.
  • a slab waveguide having an input end and an output end, and first and second surfaces parallel to each other to cause light to propagate from the input end to the output end by total internal reflection
  • a curved wedge coupled to receive light from the output end of the slab waveguide, the curved wedge having a continuously decreasing thickness to cause light to exit the wedge from one of two surfaces; and a compensating curved wedge that provides a uniform optical path length for light passing through both the curved wedge and the compensating curved wedge.
  • the apparatus of 2C5 further comprising a spatial light modulator positioned along the slab waveguide to modulate the light as it propagates by total internal reflection.
  • the apparatus of 2C5 further comprising a point light source to provide light to the input end.
  • the apparatus of 2C5 further comprising a camera for eye tracking.
  • a near-to-eye display device comprising:
  • a slab waveguide having an input end, an output end, and first and second surfaces parallel to each other to cause light received from the point light source to propagate from the input end to the output end by total internal reflection; a curved wedge coupled to receive light from the output end of the slab waveguide, the wedge having first and second surfaces oriented to form a continuously decreasing thickness to cause light to exit the curved wedge from one of the first and second surface and create a converging light beam;
  • the near-to-eye display device of 2C 10 further comprising a
  • the near-to-eye display device of 2C10 further comprising a camera for eye tracking.
  • a near-to-eye display device comprising:
  • a movable platform that includes a plurality of light sources; and a circuit to modulate the plurality of light sources and to synchronize the modulation with motion of the movable platform.
  • 3A2 The near-to-eye display device of 3A1 further comprising a polarizing film to pass environmental light polarized in a first orientation, wherein the plurality of light sources are positioned to direct light toward an expected location of a user's eye.
  • 3 A3 The near-to-eye display device of 3 Al wherein the plurality of light sources are positioned to direct light away from an expected location of a user's eye.
  • the near-to-eye display device of 3 A4 wherein the array of light sources comprises an array of light emitting diodes.
  • the near-to-eye display device of 3 A4 wherein the array of light sources comprises light sources of at least two different colors.
  • the near-to-eye display device of 3 A4 wherein the array of light sources comprises red, green, and blue light sources.
  • the near-to-eye display device of 3 A4 wherein the array of light sources comprises a two-dimensional array.
  • 3 A10 The near-to-eye display device of 3 A4 wherein the movable platform comprises a bar that moves in one dimension.
  • 3 Al 1.
  • the near-to-eye display device of 3 Al wherein the movable platform comprises a plurality of bars that move in one dimension.
  • a near-to-eye display device that comprises a movable platform that includes a plurality of light sources;
  • 3A15 The combination of 3A14 wherein the near-to-eye display device comprises a head-worn device.
  • 3A17 The combination of 3A16 wherein the plurality of light sources produce light polarized in a second orientation different from the first orientation.
  • 3A18 The combination of 3A14 wherein the first portion of the contact lens has a polarization matching the polarizing film and the second portion has a polarization matching the light produced by the plurality of light sources.
  • 3A19 The combination of 3A14 wherein the second portion of the contact lens includes color filtering.
  • 3 A20 The combination of 3A14 wherein the first and second portions of the contact lens are concentric.
  • 3A21 The combination of 3A14 wherein the plurality of light sources is on a movable platform that sweeps over a viewing area of the near-to-eye display device.
  • 3 A22 The combination of 3 A21 wherein the movable platform comprises a plurality of bars that move in one dimension.
  • 3 A23 The combination of 3 A21 wherein the movable platform moves in one dimension.
  • 3 A24 The combination of 3A21 wherein the movable platform is mounted to the near-to-eye display device at a pivot point.
  • 3A28 The combination of 3A14 wherein the movable platform is magnetically actuated.
  • 3A30 The combination of 3A14 wherein the movable platform is electrically actuated.
  • a near-to-eye display device comprising:
  • a movable platform that includes a spatial light modulator positioned to be illuminated by the point light source such that when the movable platform is swept through a user's field of view, the spatial light modulator projects light on an exit pupil positioned at an expected location of the user's eye pupil when the near-to-eye display device is in use, and wherein the exit pupil plane is at an optical conjugate location of the point light source.
  • 3A33 The near-to-eye display device of 3 A31 wherein the spatial light modulator includes multiple rows of pixels.
  • 3A34 The near-to-eye display device of 3A31 wherein the point light source is mounted on the moving platform.
  • 3A35 The near-to-eye display device of 3A31 wherein the movable platform comprises a plurality of bars that move in one dimension.
  • 3 A36 The near-to-eye display device of 3 A31 wherein the moving platform comprises a plurality of bars that each includes at least one row of spatial light modulator pixels.
  • the near-to-eye display device of 3 A31 further comprising a plurality of light sources of different colors that are time multiplexed when in use.
  • a near-to-eye display device comprising:
  • a reflective optical element rotatably mounted to the near-to-eye display device and positioned to be illuminated by the at least one point light source to project light on an exit pupil plane positioned at an expected location of a user's eye pupil when the near-to-eye display device is in use;
  • a pupil tracking device to determine a position of the user's eye pupil
  • an actuator to rotate the reflective optical element in response to the position of the user's eye pupil.
  • the near-to-eye display device of 4A1 further comprising a plurality of point light sources and a light selection component responsive to the pupil tracking device.
  • a method comprising:
  • tracking comprises measuring an angle, and further comprising driving the spatial light modulator with different data to change the diffraction order that enters the user's eye pupil for angles above a threshold.
  • a near-to-eye display device comprising:
  • an active grating that implements a multi-section prism disposed between the point light source and the spatial light modulator, the active grating being positioned to be illuminated by the point light source to direct light on the spatial light modulator, the spatial light modulator being positioned to be illuminated by the wave directed by the active grating to direct light on an exit pupil plane positioned at an expected location of a user's eye pupil when the near-to-eye display device is in use;
  • a pupil tracking device to determine a position of the user's eye pupil
  • a near-to-eye display device comprising:
  • an optical component that includes a spatial light modulator, a reflector, and an active grating disposed between the reflector and the spatial light modulator, the optical component being positioned to be illuminated by the at least one point light source to project light on an exit pupil plane positioned at an expected location of a user's eye pupil when the near-to-eye display device is in use;
  • a pupil tracking device to determine a position of the user's eye pupil
  • a method comprising:
  • tracking comprises measuring an angle, and further comprising driving the spatial light modulator with different data to change the diffraction order that enters the user's eye pupil for angles above a threshold.
  • tracking comprises measuring an angle, and further comprising selecting a different light source to illuminate the spatial light modulator based on the angle.

Abstract

Un dispositif d'affichage proche de l'oeil comprend un modulateur de lumière spatial. Le modulateur de lumière spatial module une onde d'éclairement pour créer une onde de scène virtuelle qui est orientée vers une partie utile d'un plan de pupille de sortie. Les ordres de diffraction supérieurs et les faisceaux de bruit sont filtrés par la pupille de l'utilisateur agissant en tant que filtre spatial.
PCT/TR2014/000512 2014-12-26 2014-12-26 Dispositif d'affichage proche de l'oeil WO2016105281A1 (fr)

Priority Applications (6)

Application Number Priority Date Filing Date Title
PCT/TR2014/000512 WO2016105281A1 (fr) 2014-12-26 2014-12-26 Dispositif d'affichage proche de l'oeil
EP14850109.1A EP3237957A1 (fr) 2014-12-26 2014-12-26 Dispositif d'affichage proche de l'oeil
CN201480084628.2A CN107438796B (zh) 2014-12-26 2014-12-26 头戴式显示装置、近眼显示装置和方法
US15/632,164 US10571696B2 (en) 2014-12-26 2017-06-23 Near-to-eye display device
US15/633,721 US10444507B2 (en) 2014-12-26 2017-06-26 Near-to-eye display device with spatial light modulator and pupil tracker
US15/633,753 US10444508B2 (en) 2014-12-26 2017-06-26 Apparatus for generating a coherent beam illumination

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/TR2014/000512 WO2016105281A1 (fr) 2014-12-26 2014-12-26 Dispositif d'affichage proche de l'oeil

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
PCT/TR2014/000513 Continuation WO2016105282A1 (fr) 2014-12-26 2014-12-26 Dispositif d'affichage près de l'œil à modulateur spatial de lumière et dispositif de suivi de pupille
PCT/TR2014/000516 Continuation WO2016105285A1 (fr) 2014-12-26 2014-12-26 Dispositif d'affichage près de l'œil à résolution variable

Related Child Applications (3)

Application Number Title Priority Date Filing Date
PCT/TR2014/000514 Continuation WO2016105283A1 (fr) 2014-12-26 2014-12-26 Dispositif d'affichage près de l'œil avec sources de lumière mobile
PCT/TR2014/000516 Continuation WO2016105285A1 (fr) 2014-12-26 2014-12-26 Dispositif d'affichage près de l'œil à résolution variable
US15/632,164 Continuation US10571696B2 (en) 2014-12-26 2017-06-23 Near-to-eye display device

Publications (1)

Publication Number Publication Date
WO2016105281A1 true WO2016105281A1 (fr) 2016-06-30

Family

ID=52774516

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/TR2014/000512 WO2016105281A1 (fr) 2014-12-26 2014-12-26 Dispositif d'affichage proche de l'oeil

Country Status (3)

Country Link
EP (1) EP3237957A1 (fr)
CN (1) CN107438796B (fr)
WO (1) WO2016105281A1 (fr)

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106445167A (zh) * 2016-10-20 2017-02-22 网易(杭州)网络有限公司 单眼视界自适配调整方法及装置、头戴式可视设备
CN106842880A (zh) * 2017-03-27 2017-06-13 深圳市美誉镜界光电科技有限公司 全息图像生成方法、处理器及全息图像显示装置、设备
WO2017150633A1 (fr) * 2016-03-04 2017-09-08 Sharp Kabushiki Kaisha Visiocasque à unité d'éclairage de panneau directionnel
WO2017150636A1 (fr) * 2016-03-04 2017-09-08 Sharp Kabushiki Kaisha Visiocasque utilisant un modulateur spatial de lumière pour générer une image holographique
CN108152955A (zh) * 2016-12-06 2018-06-12 艾菲瑞斯特有限公司 用于近眼显示器的图像引导光学器件
WO2018108800A1 (fr) * 2016-12-15 2018-06-21 tooz technologies GmbH Lunettes intelligentes, verre de lunettes pour des lunettes intelligentes et procédé permettant de générer une image sur la rétine
EP3351999A1 (fr) * 2017-01-19 2018-07-25 Oculus VR, LLC Dispositif d'affichage à surface focale
US10330936B2 (en) 2017-01-19 2019-06-25 Facebook Technologies, Llc Focal surface display
CN110402410A (zh) * 2017-02-28 2019-11-01 赛伊视觉公司 使用空间光调制器的近眼显示装置
US10788791B2 (en) 2016-02-22 2020-09-29 Real View Imaging Ltd. Method and system for displaying holographic images within a real object
US10795316B2 (en) 2016-02-22 2020-10-06 Real View Imaging Ltd. Wide field of view hybrid holographic display
JP2020537767A (ja) * 2017-10-18 2020-12-24 シーリアル テクノロジーズ ソシエテ アノニムSeereal Technologies S.A. 大きな視野を生成するための表示装置および方法
US10877437B2 (en) 2016-02-22 2020-12-29 Real View Imaging Ltd. Zero order blocking and diverging for holographic imaging
AU2017360746B2 (en) * 2016-11-15 2022-07-28 Creal Sa Near-eye sequential light-field projector with correct monocular depth cues
CN115145023A (zh) * 2016-12-31 2022-10-04 鲁姆斯有限公司 用于导出人眼睛的注视方向的设备
US11604352B2 (en) 2020-07-29 2023-03-14 Meta Platforms Technologies LLC Waveguide-based projector
US11663937B2 (en) 2016-02-22 2023-05-30 Real View Imaging Ltd. Pupil tracking in an image display system
US11953689B2 (en) 2020-09-30 2024-04-09 Hes Ip Holdings, Llc Virtual image display system for virtual reality and augmented reality devices

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI666483B (zh) * 2018-03-22 2019-07-21 醫百科技股份有限公司 可調焦之影像顯示眼鏡
CN108762033B (zh) * 2018-05-28 2022-12-09 江苏慧光电子科技有限公司 成像方法和光学系统及其存储介质、芯片与组合
CN109462924B (zh) * 2018-10-11 2020-09-11 北方信息控制研究院集团有限公司 一种多层空间光调制自适应光场的照明方法
CN109061883B (zh) * 2018-10-25 2020-12-04 杭州瞳创医疗科技有限公司 一种自动测量瞳距的头戴式显示设备和方法
JP7320057B2 (ja) * 2019-01-31 2023-08-02 クリアル・ソシエテ・アノニム 観察者に対する正しい単眼奥行き手がかりを持つライトフィールド複合現実システム
CN109725462B (zh) * 2019-03-04 2022-11-04 京东方科技集团股份有限公司 显示器件、显示设备以及显示器件的驱动方法
CN110308566B (zh) * 2019-06-28 2021-12-03 上海慧希电子科技有限公司 显示系统及双目系统
CN110824611B (zh) * 2019-11-21 2021-02-05 歌尔股份有限公司 均匀性补偿方法、光波导系统及增强现实设备
CN111458895A (zh) * 2020-04-28 2020-07-28 迪鹏光电科技股份有限公司 阵列式结构光图案投射装置
CN111624774B (zh) * 2020-06-30 2023-04-11 京东方科技集团股份有限公司 增强现实显示光学系统及显示方法
CN112684529A (zh) * 2020-12-28 2021-04-20 上海慧希电子科技有限公司 光学器件、系统及光学设备

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB1505873A (en) * 1976-09-03 1978-03-30 British Aircraft Corp Ltd Optical display devices
WO2008065569A1 (fr) * 2006-11-30 2008-06-05 Koninklijke Philips Electronics, N.V. Dispositif de front d'onde électronique et procédé de rendu électronique d'un front d'onde
US20090180165A1 (en) * 2005-12-22 2009-07-16 See Real Technologies S.A. Method for the Compensation of an Inhomogeneous Brightness Perception in Holographically Reconstructed Scenes
US20090219589A1 (en) * 2004-01-06 2009-09-03 Takahiro Ikeda Three-dimensional image display and three-dimensional image displaying method

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101743519B (zh) * 2007-05-16 2013-04-24 视瑞尔技术公司 全息显示装置
CN101794028B (zh) * 2010-03-04 2011-12-28 首都师范大学 光学实时三维立体显示装置及方法
WO2012147271A1 (fr) * 2011-04-27 2012-11-01 パナソニック株式会社 Dispositif d'affichage
TW201300834A (zh) * 2011-06-23 2013-01-01 Seereal Technologies Sa 顯示裝置,尤其是頭戴式顯示器或護目鏡
US20140146394A1 (en) * 2012-11-28 2014-05-29 Nigel David Tout Peripheral display for a near-eye display device

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB1505873A (en) * 1976-09-03 1978-03-30 British Aircraft Corp Ltd Optical display devices
US20090219589A1 (en) * 2004-01-06 2009-09-03 Takahiro Ikeda Three-dimensional image display and three-dimensional image displaying method
US20090180165A1 (en) * 2005-12-22 2009-07-16 See Real Technologies S.A. Method for the Compensation of an Inhomogeneous Brightness Perception in Holographically Reconstructed Scenes
WO2008065569A1 (fr) * 2006-11-30 2008-06-05 Koninklijke Philips Electronics, N.V. Dispositif de front d'onde électronique et procédé de rendu électronique d'un front d'onde

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
STEPHAN REICHELT ET AL: "Holographic 3-D Displays -Electro-holography within the Grasp of Commercialization", 1 April 2010 (2010-04-01), XP055149317, Retrieved from the Internet <URL:http://www.intechopen.com> [retrieved on 20141028] *

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11543773B2 (en) 2016-02-22 2023-01-03 Real View Imaging Ltd. Wide field of view hybrid holographic display
US10788791B2 (en) 2016-02-22 2020-09-29 Real View Imaging Ltd. Method and system for displaying holographic images within a real object
US10795316B2 (en) 2016-02-22 2020-10-06 Real View Imaging Ltd. Wide field of view hybrid holographic display
US11663937B2 (en) 2016-02-22 2023-05-30 Real View Imaging Ltd. Pupil tracking in an image display system
US10877437B2 (en) 2016-02-22 2020-12-29 Real View Imaging Ltd. Zero order blocking and diverging for holographic imaging
US11754971B2 (en) 2016-02-22 2023-09-12 Real View Imaging Ltd. Method and system for displaying holographic images within a real object
WO2017150633A1 (fr) * 2016-03-04 2017-09-08 Sharp Kabushiki Kaisha Visiocasque à unité d'éclairage de panneau directionnel
WO2017150636A1 (fr) * 2016-03-04 2017-09-08 Sharp Kabushiki Kaisha Visiocasque utilisant un modulateur spatial de lumière pour générer une image holographique
US9964768B2 (en) 2016-03-04 2018-05-08 Sharp Kabushiki Kaisha Head mounted display using spatial light modulator to generate a holographic image
CN106445167A (zh) * 2016-10-20 2017-02-22 网易(杭州)网络有限公司 单眼视界自适配调整方法及装置、头戴式可视设备
US11960083B2 (en) * 2016-11-15 2024-04-16 Creal Sa Near-eye sequential light-field projector with correct monocular depth cues
AU2017360746B2 (en) * 2016-11-15 2022-07-28 Creal Sa Near-eye sequential light-field projector with correct monocular depth cues
CN108152955A (zh) * 2016-12-06 2018-06-12 艾菲瑞斯特有限公司 用于近眼显示器的图像引导光学器件
WO2018108800A1 (fr) * 2016-12-15 2018-06-21 tooz technologies GmbH Lunettes intelligentes, verre de lunettes pour des lunettes intelligentes et procédé permettant de générer une image sur la rétine
CN110121671A (zh) * 2016-12-15 2019-08-13 图茨技术股份有限公司 数据眼镜、用于数据眼镜的眼镜片和用于在视网膜上生成图像的方法
US11630306B2 (en) 2016-12-15 2023-04-18 tooz technologies GmbH Smartglasses, lens for smartglasses and method for generating an image on the retina
CN115145023B (zh) * 2016-12-31 2024-02-09 鲁姆斯有限公司 用于导出人眼睛的注视方向的设备
CN115145023A (zh) * 2016-12-31 2022-10-04 鲁姆斯有限公司 用于导出人眼睛的注视方向的设备
US10558049B2 (en) 2017-01-19 2020-02-11 Facebook Technologies, Llc Focal surface display
US10330936B2 (en) 2017-01-19 2019-06-25 Facebook Technologies, Llc Focal surface display
EP3351999A1 (fr) * 2017-01-19 2018-07-25 Oculus VR, LLC Dispositif d'affichage à surface focale
CN110402410A (zh) * 2017-02-28 2019-11-01 赛伊视觉公司 使用空间光调制器的近眼显示装置
CN110402410B (zh) * 2017-02-28 2022-08-30 赛伊视觉公司 使用空间光调制器的近眼显示装置
CN106842880B (zh) * 2017-03-27 2018-09-28 深圳市美誉镜界光电科技有限公司 全息图像生成方法、处理器及全息图像显示装置、设备
WO2018176996A1 (fr) * 2017-03-27 2018-10-04 深圳市美誉镜界光电科技有限公司 Procédé de génération d'image holographique, processeur, appareil et dispositifs d'affichage d'image holographique
CN106842880A (zh) * 2017-03-27 2017-06-13 深圳市美誉镜界光电科技有限公司 全息图像生成方法、处理器及全息图像显示装置、设备
JP2020537767A (ja) * 2017-10-18 2020-12-24 シーリアル テクノロジーズ ソシエテ アノニムSeereal Technologies S.A. 大きな視野を生成するための表示装置および方法
US11604352B2 (en) 2020-07-29 2023-03-14 Meta Platforms Technologies LLC Waveguide-based projector
US11953689B2 (en) 2020-09-30 2024-04-09 Hes Ip Holdings, Llc Virtual image display system for virtual reality and augmented reality devices

Also Published As

Publication number Publication date
CN107438796B (zh) 2020-07-03
CN107438796A (zh) 2017-12-05
EP3237957A1 (fr) 2017-11-01

Similar Documents

Publication Publication Date Title
US10571696B2 (en) Near-to-eye display device
US10241328B2 (en) Near-to-eye display device with variable resolution
CN107438796B (zh) 头戴式显示装置、近眼显示装置和方法
WO2016105282A1 (fr) Dispositif d&#39;affichage près de l&#39;œil à modulateur spatial de lumière et dispositif de suivi de pupille
JP7329105B2 (ja) ディスプレイシステムのための深度ベース中心窩化レンダリング
US11710469B2 (en) Depth based foveated rendering for display systems
CN107300769B (zh) 虚拟和增强现实系统与方法
WO2016105283A1 (fr) Dispositif d&#39;affichage près de l&#39;œil avec sources de lumière mobile
US20170188021A1 (en) Optical engine for creating wide-field of view fovea-based display
WO2016105284A1 (fr) Appareil pour générer un éclairage à faisceau cohérent
EP3513254B1 (fr) Affichage holographique à champ de vision large
US20210080719A1 (en) Low-obliquity beam scanner with polarization-selective grating
EP3531214B1 (fr) Dispositif d&#39;affichage holographique et son procédé de commande
TW202235963A (zh) 異質分層的體積布拉格光柵波導架構

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14850109

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

REEP Request for entry into the european phase

Ref document number: 2014850109

Country of ref document: EP