WO2014030158A1 - Visual aid projector - Google Patents

Visual aid projector Download PDF

Info

Publication number
WO2014030158A1
WO2014030158A1 PCT/IL2013/050702 IL2013050702W WO2014030158A1 WO 2014030158 A1 WO2014030158 A1 WO 2014030158A1 IL 2013050702 W IL2013050702 W IL 2013050702W WO 2014030158 A1 WO2014030158 A1 WO 2014030158A1
Authority
WO
WIPO (PCT)
Prior art keywords
light
retina
light source
video
intensity
Prior art date
Application number
PCT/IL2013/050702
Other languages
English (en)
French (fr)
Other versions
WO2014030158A4 (en
Inventor
Haim Chayet
Boris Greenberg
Lior BEN-HUR
Original Assignee
Ic Inside Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to CA2882747A priority Critical patent/CA2882747A1/en
Priority to KR1020157007083A priority patent/KR20150048789A/ko
Priority to JP2015528004A priority patent/JP2015526187A/ja
Priority to CN201380055324.9A priority patent/CN104812342B/zh
Priority to EP13830523.0A priority patent/EP2887913A4/en
Priority to RU2015108858A priority patent/RU2015108858A/ru
Priority to AU2013307201A priority patent/AU2013307201B2/en
Application filed by Ic Inside Ltd filed Critical Ic Inside Ltd
Priority to BR112015003899A priority patent/BR112015003899A2/pt
Priority to US14/423,668 priority patent/US10058454B2/en
Publication of WO2014030158A1 publication Critical patent/WO2014030158A1/en
Publication of WO2014030158A4 publication Critical patent/WO2014030158A4/en
Priority to IL237384A priority patent/IL237384A0/en
Priority to IN1864DEN2015 priority patent/IN2015DN01864A/en

Links

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61FFILTERS IMPLANTABLE INTO BLOOD VESSELS; PROSTHESES; DEVICES PROVIDING PATENCY TO, OR PREVENTING COLLAPSING OF, TUBULAR STRUCTURES OF THE BODY, e.g. STENTS; ORTHOPAEDIC, NURSING OR CONTRACEPTIVE DEVICES; FOMENTATION; TREATMENT OR PROTECTION OF EYES OR EARS; BANDAGES, DRESSINGS OR ABSORBENT PADS; FIRST-AID KITS
    • A61F9/00Methods or devices for treatment of the eyes; Devices for putting in contact-lenses; Devices to correct squinting; Apparatus to guide the blind; Protective devices for the eyes, carried on the body or in the hand
    • A61F9/08Devices or methods enabling eye-patients to replace direct visual perception by another kind of perception
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B26/00Optical devices or arrangements for the control of light using movable or deformable optical elements
    • G02B26/08Optical devices or arrangements for the control of light using movable or deformable optical elements for controlling the direction of light
    • G02B26/0816Optical devices or arrangements for the control of light using movable or deformable optical elements for controlling the direction of light by means of one or more reflecting elements
    • G02B26/0833Optical devices or arrangements for the control of light using movable or deformable optical elements for controlling the direction of light by means of one or more reflecting elements the reflecting element being a micromechanical device, e.g. a MEMS mirror, DMD
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B26/00Optical devices or arrangements for the control of light using movable or deformable optical elements
    • G02B26/08Optical devices or arrangements for the control of light using movable or deformable optical elements for controlling the direction of light
    • G02B26/10Scanning systems
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B26/00Optical devices or arrangements for the control of light using movable or deformable optical elements
    • G02B26/08Optical devices or arrangements for the control of light using movable or deformable optical elements for controlling the direction of light
    • G02B26/10Scanning systems
    • G02B26/101Scanning systems with both horizontal and vertical deflecting means, e.g. raster or XY scanners
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/0093Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/10Beam splitting or combining systems
    • G02B27/1006Beam splitting or combining systems for splitting or combining different wavelengths
    • G02B27/102Beam splitting or combining systems for splitting or combining different wavelengths for generating a colour image from monochromatic image signal sources
    • G02B27/104Beam splitting or combining systems for splitting or combining different wavelengths for generating a colour image from monochromatic image signal sources for use with scanning systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N3/00Scanning details of television systems; Combination thereof with generation of supply voltages
    • H04N3/02Scanning details of television systems; Combination thereof with generation of supply voltages by optical-mechanical means only
    • H04N3/08Scanning details of television systems; Combination thereof with generation of supply voltages by optical-mechanical means only having a moving reflector
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/12Picture reproducers
    • H04N9/31Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
    • H04N9/3179Video signal processing therefor
    • H04N9/3188Scale or resolution adjustment
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/0118Head-up displays characterised by optical features comprising devices for improving the contrast of the display / brillance control visibility
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/0138Head-up displays characterised by optical features comprising image capture systems, e.g. camera
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/014Head-up displays characterised by optical features comprising information/image processing systems
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/0141Head-up displays characterised by optical features characterised by the informative content of the display
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • G02B2027/0178Eyeglass type

Definitions

  • the present invention relates to an apparatus, a system and a method for aiding the vision of visually impaired individuals, and in particular, to such an apparatus, system and method which assists visually impaired individuals having a retina with reduced functionality.
  • Visually impaired individuals frequently suffer from reduced vision due to reduced functionality of their retina. Such reduced functionality may be due to damage, and/or to various disease processes which gradually reduce the function of the retina and hence lead to impaired vision over time.
  • One example of such a disease process relates to AMD, or age- related macular degeneration, the leading cause of blindness among adults in the US.
  • the macula is the part of the retina which provides sharply focused images and hence is particularly required for such activities as reading and driving.
  • AMD poses another challenge, which is that since its effects on the retina are uneven, mainly affecting the macula, such that assistive or corrective devices need to be able to relate to uneven retinal function, which may further vary between individuals.
  • assistive or corrective devices need to be able to relate to uneven retinal function, which may further vary between individuals.
  • the most effective devices have involved implants to the eye, but these devices are highly experimental, very invasive and clearly undesirable (from a therapeutic standpoint) in many ways.
  • Some taught devices do not require any implantable element, as described for example in US Patent No. 7023621 to Dietrich; however this device assumes an ideal retina, with entirely homogenous functioning over the entire retina. For a normal, healthy retina, such an assumption, while incorrect, may still yield a reasonably effective device; however, for a retina with reduced function, or otherwise uneven function, such an assumption is not correct and greatly reduces the effectiveness of the device.
  • the background art does not teach or suggest an apparatus, system or method for aiding the vision of visually impaired individuals having a retina with reduced functionality.
  • reduced functionality it is meant that the retina functions at a lower or lesser level than for a normal healthy retina, and/or that the retina has uneven function at different locations, so that for example and without limitation, one or more parts of the retina at different locations may function at a higher level, while one or more other parts of the retina, at other locations, may function at a lower level.
  • the present invention is of an apparatus, system or method for aiding the vision of visually impaired individuals having a retina with reduced functionality, which overcomes the drawbacks of the background art by overcoming such reduced and/or uneven retinal function.
  • the system of the present invention comprises a mapping unit, for mapping the retinal function at various locations of the retina; and a projection unit for projecting light onto the retina as a video projector, in which the location and/or intensity of the light at various parts of the retina is determined at least in part by the mapped retinal function.
  • both units may be implemented in a single apparatus; the mapping software and analysis functions may also optionally be implemented in that apparatus or alternatively may be implemented at a remote location. Communication between such a remote location and the apparatus may optionally be wireless or wired, as necessary.
  • the level of retinal function at each of a plurality of portions of the retina determines whether that portion of the retina is stimulated through the video projection process or alternatively whether that portion of the retina is avoided during the video projection process; for the latter situation, the video data is projected onto the "remapped" retina, in which excessively low function portions of the retina are not stimulated but instead are avoided, as described in greater detail below.
  • this division between “stimulation” and “avoidance” may be implemented flexibly, according to the overall light level available, the type of vision desired (reading, driving, night vision) and so forth.
  • the apparatus in order to compensate for movement of the eye, further comprises an eye tracker.
  • Figure 1 shows a virtual retinal display (V D) according to at least some embodiments of the present invention that projects a direct video image onto the user's retina;
  • the VRD 100 comprises light source(s) 120 such as for example a laser, or more particularly a laser diode;
  • Figure 3 shows an exemplary, illustrative, optional embodiment of the projection optics 140 as a schematic block diagram
  • FIG. 4 shows operation of video processor 170
  • Figure 5 is an exemplary segmentation of retinal FOV where each angular segment is represented with a cell in the RST;
  • Figure 6 is an example of the first 10 rows by 10 columns of an illustrative RRST, for a retina which demonstrates reduced sensitivity within the center of this region;
  • Figure 7 relates to an exemplary, illustrative method for receiving an incoming image and for projecting that image appropriately onto the retina of the subject through the retinal projection apparatus;
  • Figure 8 shows exemplary analog video signal derived from the camera of a raster scan; the middle plot shows the retinal position of the light signal during the scan; the lower plot shows the video signal along the retinal position;
  • Figure 9 is a video signal resulting from the camera, here shown as a digital signal for clarity; the middle plot indicates the relative sensitivity scores of the retina according the position of the retina;
  • Figure 10 in the upper diagram (10a), shows an image of letters ABC represented with the retinal projector as raster scan of row 3 and 4; the second diagram of Figure 10 (10b) represents how the projected image appears in field of vision with stocoma at the central portion; the third diagram of Figure 10 (10c) shows the letters ABC when the raster image was projected at rows 5 to 6;
  • Figure 11 relates to an exemplary, illustrative method according to at least some embodiments of the present invention for receiving an incoming image and for projecting that image appropriately onto the retina of the subject through the retinal projection apparatus, for example as described herein, in which one or more portions of the retina are essentially nonfunctional;
  • Fig 12 shows the schematics of an exemplary, illustrative scan over the field of vision where the indicated path, in the shape of an "S-curve" with a plurality of inflection points, indicates the trajectory of the scan.
  • Photons from a light source are modulated with video information by an intensity modulator.
  • the modulated light is scanned in a first direction and in a second direction generally perpendicular to the first direction by a scanners to create a raster of photons that is projected directly onto the retina of the user by projection optics to produce the perception an image without any intermediate image outside of the eye that is viewed or perceived by the user.
  • the light source can be a monochromatic light or beams of red, green and yellow or blue light may be RGB video signals to scan colored photons directly onto the user's eye.
  • V D virtual retinal display
  • the VRD 100 utilizes a light beam 110 emitted by a light source(s) 120, an optical scanner 130 that scans the light beam 110 in a raster mode horizontally and vertically , and is further relayed by projection optics 140 through the pupil 151 of user's eye 155 directly on the retina 152.
  • a video processor 170 accepts an external video signal 171 from an external video source 175 and processes signal 171 in accordance with aspects of mapping as described in greater detail below, and optionally and preferably also according to the instantaneous position of pupil 151 as optionally and more preferably detected by an eye tracking system 160.
  • the video processor 170 then generates a synchronization signal 172 to the scanner 130, and a modulation signal 173 according which light beam 110 is intensity modulated.
  • the intensity modulated beam of light 110 as emitted from scanner 130 through optics 140 raster scans the retina 152 and so produces the perception of an image without any intermediate image outside of the eye 155.
  • VRD 100 Since the image is formed directly on the retina, without any intermediate displays or projection screens, the device can be reduced in size and weight, which in turn makes VRD 100 suitable for mounting onto head 150 of the user. However in any case it should be noted that VRD 100 as described herein is not intended for implantation to the eye 155 or attachment to the eye 155 and instead remains outside of the eye 155; only the intensity modulated beam of light enters the eye 155. As seen in Fig 1 VRD 100 preferably features a head-mounted part 101 which includes the light source(s) 120, the scanner 130, projection optics 140 and optionally and preferably eye tracking system 160.
  • a power source such as battery (not shown) may optionally be located in a remote module 102 which can also be portable but need not be located in proximity to the eye 155.
  • a power source such as battery (not shown)
  • the power source of remote module 102 is connected to head-mounted part 101 with a wire 103 or other power connector or cable.
  • the VRD 100 comprises light source(s) 120 such as for example a laser, or more particularly a laser diode.
  • the VRD 100 may contain one or several light sources 120.
  • light sources 120 preferably feature at least three lasers which emit at red, green and blue (designated 201, 202, and 203
  • the light sources 120 do not need to be coherent or at a narrow spectral width; light emitting diodes (LED) may also optionally be employed.
  • LED light emitting diodes
  • the individual light sources 120 are preferably further provided with collimation optics 205 as shown.
  • the light beam 110 emitted has low divergence.
  • Light emitted from a laser source can become collimated to a beam with a very small divergence angle.
  • collimation of incoherent light as LED is more difficult.
  • the LED can be collimated by using an aspheric lens with numerical aperture greater than or equal to numerical aperture of the beam, which could for example optionally be incorporated into collimation optics 205 as shown.
  • the beam will have residual divergence.
  • VRD 100 comprises further optical elements to project the light into the pupil 151, such as projection optics 140 for example, a LED source can still provide sufficient intensity and low divergence to be utilized as light source 120.
  • the light sources 120 are optionally and preferably further provided with a light attenuator 206 such that the light emitted from light source 120 is well within the eye-safe intensity level.
  • Light attenuator 206 may optionally comprise silica glass or ceramics for example.
  • each light source 120 is optionally and preferably provided with a separate light attenuator 206, as shown in the exemplary configuration of Figure 2.
  • Projection of color images provides a particular challenge, because the intensity perception of the eye 155 is different at each color (wavelength); therefore each color needs to be attenuated separately and to a differing degree to comply with eye-safety requirements at the specific spectral range.
  • the required safety levels are typically given in various regulations and directives e.g., the US standard ANSI Z136 or the European standard EN 207.
  • the beams from individual light sources are optionally and preferably combined by a beam combiner 207 based on dichroic filters, or spectrally dispersive components, to form light beam 110.
  • a beam combiner 207 based on dichroic filters, or spectrally dispersive components, to form light beam 110.
  • the beam of red light 221 emitted from the red light source 201 is first bent by mirror 224 and subsequently combined with beam of green light 222 by dichroic filter 225.
  • the dichroic filter 225 transmits the red light 221 and reflects the green light 222.
  • dichroic filter 226 reflects the blue light 223, while the red and the green beams 221, 222 are transmitted.
  • a beam of light 110 that contains all RGB components is emitted from the beam combiner 207.
  • an additional light attenuator 211 is situated to receive light beam 110.
  • the light passes a polarization control unit 212 that can contain polarizers, waveplates such as ⁇ /2 or ⁇ /4, or other optical elements for controlling the polarization.
  • polarization control unit 212 can contain polarizers, waveplates such as ⁇ /2 or ⁇ /4, or other optical elements for controlling the polarization.
  • additional attenuation can be provided by the polarization elements of polarization control unit 212.
  • a non-limiting example of a device that could be used to implement polarization control unit 212 is a variable attenuator based on polarization control, which may for example and without limitation be purchased from Newport Inc. (Irvine, CA, USA), (http://assets.newport.com/webDocuments- EN/images/Variable Attenuators.pdf).
  • the beam 110 can be further deflected, here shown as being bent by mirror 227.
  • the light beam 110 passes through an intensity monitor unit ("light tap") 215.
  • a small fraction of light 213 is split from the beam 110 by a beam splitter 214, typically 1% to 5% of the main beam 110.
  • Light fraction 213 is then fed to intensity monitor unit 215 and used to monitor the light intensity.
  • the signal from intensity monitor 215 is used as a feedback signal 216 to control the light intensity and to disable (shut off) the light sources 120, which provides additional level of safety to the operation of VRD 100.
  • feedback signal 216 is optionally and preferably communicated to control logic electronics 217
  • each light source 120 could have its own control logic electronics 217).
  • Control logic electronics 217 compare the value of the power of light source 120, such as a laser for example, to a predetermined value (which corresponds to eye-safe power value). If the monitored light power is exceeds the predetermined value, then control logic electronics 217 would shut down the respective light source 120 or light sources 120.
  • the light source 120 can be disabled in a variety of ways, for example by interrupting the supply of power to light source
  • control logic 217 is preferably implemented in the driver wherein light emission is disabled by turning off the current to the laser.
  • Control logic electronics 217 is an example of a modulator, albeit in the form of a cut-off device.
  • light beam 110 is also optionally intensity modulated as follows.
  • the video processor 170 generates a modulation signal 173 according which the beam of light 110 is intensity modulated.
  • the intensity modulation is achieved by modulating the current driving the light source 120, which may for example be a laser diode.
  • the light source 120 is directly modulated the emitted beam of light 110 emerges from the source
  • the modulation signal 173 is fed as a drive signal to a transducer of the acousto-optic modulator.
  • the modulation signal is preferably provided to all sources 120, according to the RGB video information for RGB light sources.
  • the signals are communicated to the external modulator drivers and each light source 120 is provided with a modulator (not shown).
  • the VRD 100 may include a safety feature where the laser is disabled by the intensity monitor signal 216.
  • modulators can be utilized.
  • the dynamic range of the modulation is defined by the original image data requirements and it is limited by the actual modulation ability of the chosen light modulating device.
  • the light beam 110 incident on the scanner 130 is intensity modulated. With color projection each individual beam is intensity modulated according the RGB video information before being combined to light beam 110.
  • the beam scanner 130 is used to steer or deflect the light beam 110 horizontally and vertically in a two dimensional raster pattern.
  • the deflected beam 111 is further focused on the retina 152 by the projection optics 140, and optical system of the eye 155.
  • a particular point on the retina 152 where the light focuses (such a point may also be referred to herein as a "pixel") corresponds to a particular horizontal and vertical angle spanned by the scanner 130.
  • the focused beam 112 scans the retina 152 as the scanner 130 causes focused beam to scan horizontally and vertically.
  • the intensity of the focused beam corresponds to the light intensity of the modulated beam 110. In such a manner a raster display is drawn directly onto the retina 152 which the user perceives as an image.
  • beam scanner 130 features a fast steering mirror for creation of the row raster, described herein as horizontal scanner 231, and a slow addressable axis spanning the rows, described herein as vertical scanner 232.
  • MEMS micro-electro ⁇ mechanical systems
  • Scanner requirements for horizontal scanner 231 and vertical scanner 232 are defined by at least the following factors: frame rate, resolution, pixel count, and optical system limitations.
  • Fast steering axis ("horizontal scanner") 231 frequency is defined as [M of rows in a frame] x [N of frames], in which each frame is an image projected onto the retina as the previously described raster display which is composed of a number of rows M.
  • Slow steering axis (“vertical scanner”) 232 movement bandwidth is determined by the required frame rate. It is preferable to use an addressable scanner with position feedback for this purpose in order to be able to compensate for various aberrations of the surfaces of the optical path, including with regard to any aberrations of the various lenses and mirrors in the optical path.
  • Each of horizontal scanner 231 and vertical scanner 232 is preferably driven by video processor 170 according to synchronization signal 172 for the scanning voltage, but may also include a slow varying voltage that depends on the feedback signal 263 received from the eye tracking system 160 as will be explained further.
  • vector scan where the light is steered between specified points, which may also be implemented according to various embodiments of the present invention.
  • the mirrors that is, horizontal scanner 231 and vertical scanner 232
  • Vector scan mode requires addressable scanners that operate in a close loop with feedback on their position and velocity for both axes.
  • the projection optics 140 acts in conjugation with the refractive optical system of the eye 155 to convey the light beam through the pupil 151 and focus on the retina 152.
  • the projection optics 140 is designed such that the beam of light 111 deflected by the scanner 130 at angles spanning the entire field of view (FOV) essentially crosses the optical axis at the center of pupil 151 while avoiding vignetting.
  • FOV field of view
  • Vignetting is a visual artifact which occurs when the peripheral parts of an image appear darker, due to a reduction of the image's brightness or saturation at the periphery compared to the image center.
  • Fig 3 shows an exemplary, illustrative, optional embodiment of the projection optics 140 as a schematic block diagram, in which some components have been omitted for the sake of clarity.
  • the specific embodiment shown features an arrangement sometimes referred to as a "4-f scanner design" where the scan mirror 305 is conjugated with the pupil 151.
  • Lens 301 is positioned essentially such that its back focal plane coincides with the pupil 151. Focusing of the beam 112 on the retina 152 is achieved by the refractive mechanism of the ophthalmic system of eye 155, particularly the cornea 306 and the lens 307.
  • the front focal plane of lens 301 coincides with back focal plane of lens 302.
  • a third lens 303 may be required so as to provide the fine tuning divergence of the collimated beam 110.
  • projecting optics for projections optics 140 are optionally possible including but not limited to more optical elements depending on details of the optical components such as light sources and scanner, and to account for refraction errors of the ophthalmic system (e.g., astigmatisms; not shown).
  • an additional lens (not shown) is provided to overcome for ophthalmic conditions such as myopia, or hyperopia, hence a certain beam divergence, or convergence, need to be introduced, while on the other hand a perfectly collimated beam is preferably adapted for individuals that require no refracting correction.
  • the projection optics 140 can be designed such that no light other than projected light beam 112 passes to the user's eye 155, or alternatively the projection optics 140 can be designed such that the user views the real world separately from the projected image.
  • the virtual image projected by the VRD 100 is preferably superimposed on the image of the real-world.
  • the VRD 100 optionally and preferably further comprises an eye tracking system 160 that may contain an IR illumination 261 and IR sensitive camera 262 which captures IR reflections from the outer surface of eye 155 and in this way detects the position of the pupil 151 as it moves.
  • the IR light path 264 can be combined with the video signal light by a beam combiner 265.
  • the instantaneous position of the pupil 151 provides a feedback signal to offset the scanner 130 deflection in the direction of the pupil 151, and is further processed in the video processor 170 to derive the video information to be projected onto retina 152.
  • the operation of video processor 170 will now be explained also in conjunction with Figure 4.
  • the video processor 170 optionally and preferably receives two main external input signals: an external video signal 171 from an external video source 175, namely an electronic device which generates video data such as a camera, a computer display, a mobile device or any other device that can generate video data (not shown).
  • a second input to the video processor 170 is preferably a signal communicated by the eye tracking system 160 of the instantaneous position of the pupil 151.
  • Other inputs to the video processor 170 not shown for clarity may optionally include feedback from the scanner and the instantaneous intensity communicated by intensity monitor unit 215.
  • the video processor 170 preferably has two main output signals: it provides the signal of modulation of light beam intensity (modulation signal 173), and provides signal 172 to the scanner 130 synchronized with the modulation signal 173.
  • the signal 172 to the scanner 130 is composed of total of four signals, two for each scanning axes.
  • the signal to each scanning direction is composed of high frequency signal that triggers the raster scan, and a slow varying voltage which is an offset to the scanning mirror 130 to track the instantaneous position of the pupil 151.
  • the video processor 170 In a light tracking system light, typically infrared, is reflected from the eye 155 and sensed by a video camera or some other optical sensor 262. The information is then analyzed and the position of the pupil 151 is determined. In response to the detected position of the pupil 151, the scanning mirror 130 is moved, both vertically and horizontally, such that the exit pupil of the VRD 100 is aligned with the pupil 151. Accordingly, the video processor 170 generates a high frequency signal 411 for triggering the raster scan in a horizontal direction, and slow varying signal 412 for horizontal offset which are fed through an amplifier stage 415 to horizontal scan 231 of the scanning mirror 130. Similarly, the video processor generates a high frequency signal 413 for triggering the raster scan in a vertical direction, and slow varying signal 414 for vertical offset which are fed through an amplifier stage 416 to vertical scan 232 of the scanning mirror 130.
  • a high frequency signal 411 for triggering the raster scan in a horizontal direction
  • the input of the pupil position also serves to process the projected video frame from the external video frame as follows.
  • the entire retinal field of view (FOV) is decided by the largest angle of light rays that enter the pupil and reach the retina. Rays projected at higher angles will not reach the retina and hence will not be perceived.
  • the virtual image projected by the VRD can be tuned to match exactly the frame size of the incoming external video, or alternatively the VRD can be tuned to project only part of the external video frames.
  • the external video signal that enters the video processor fills the entire retinal FOV the virtual retinal image is said to be at unity magnification (xl).
  • the VRD projects only part of the external video frame across the entire FOV of the retina the VRD operates at larger magnifications.
  • magnification in this example refers to the portion of the external video frame that is projected onto the retina, such that such a frame can be said to be magnified when only part of the frame is projected over the retinal FOV. It should also be noted that the retinal FOV does not necessarily refer to the entire retina, as described in greater detail below, since part of the retina may be damaged or inoperative.
  • the situation is similar to an ordinary process of visual perception when an extended (panoramic) object is viewed.
  • the user moves her pupil across the object in such a manner that different parts of the object are perceived at a given position of the pupil.
  • the instantaneous position of the pupil therefore defines a partial frame of the object that the user perceives.
  • the VRD operates at magnification higher than unity the instantaneous position of the pupil defines the part of the external video frame that can be projected to the retina.
  • the video processor stores the external video frames in a buffer, and according the instantaneous position of the pupil, processes a sub-frame to be projected by the VRD to accommodate available FOV.
  • the central part of the frame corresponds to the center of the pupil where the pupil aperture defines the "window" frame.
  • VRD at large magnification
  • AMD age related macular degeneracy
  • the parts of the image that fall on those parts are not perceived.
  • the image is magnified details of the magnified image are projected over the undamaged parts in the retina which can compensate for the damaged ones.
  • the shortcoming in magnification is that peripheral parts of the external video frame are not imaged.
  • the VRD of the present invention is provided with tunable magnification controller 420 that can be adjusted by the user to provide input to video processor 170 regarding the extent of desired magnification.
  • tunable magnification controller 420 can be adjusted by the user to provide input to video processor 170 regarding the extent of desired magnification.
  • magnification controller 420 can be adjusted by the user to provide input to video processor 170 regarding the extent of desired magnification.
  • magnification controller 420 can be adjusted by the user to provide input to video processor 170 regarding the extent of desired magnification.
  • the frame processed from the external incoming video is further processed in the video processor with respect to retinal mapping of each individual as explained below.
  • a retinal sensitivity map, or map of levels of retinal function at different locations (pixels), is preferably provided for the best operation of the below described method according to at least some embodiments of the present invention.
  • Art known devices failed to include such a map, which means that they would not be effective for individuals having damaged retinas or retinas with reduced and/or uneven functioning.
  • the map is preferably obtained by a mapping unit, which may optionally and preferably be implemented with the device of Figures 1 and 2 according to at least some embodiments of the present invention.
  • mapping unit may optionally be implemented with the projection unit (shown as the apparatus of Figures 2 above in a non-limiting example), including the mapping software and analysis functions; alternatively these latter functions may optionally be performed separately and then the results fed to the processing unit of the projection apparatus. It may optionally be performed offline by a separate computer or alternatively through a combined device.
  • Electroretinogram (ERG), pattern ERG (PERG), visual evoked potential (VEP) and other optometric and ophthalmic tests described in the art specify the visual field of an individual wherein these test provide information of an loss of vision or a reduction in sensitivity (threshold) of selected areas of the human retina.
  • ERG Electroretinogram
  • PERG pattern ERG
  • VEP visual evoked potential
  • other optometric and ophthalmic tests described in the art specify the visual field of an individual wherein these test provide information of an loss of vision or a reduction in sensitivity (threshold) of selected areas of the human retina.
  • These tests are well known in the art and may optionally be implemented with various types of apparatus, also as known in the art.
  • US Patent No. 5233373 describes an apparatus and a method of use thereof for implementing a PERG test on a retina of a patient, which is hereby incorporated by reference as if fully set forth herein, which includes a method of testing different locations of the retina in
  • the retinal sensitivity map obtained in ophthalmic evaluation is decoded into a retinal sensitivity table (RST).
  • RST retinal sensitivity table
  • a typical adult human retina is about 72% of a sphere of 22 mm in diameter.
  • the entire retinal field of view (FOV) is decided by the largest angle of light rays that enter the pupil and reach the retina.
  • Fig 5 Shown in Fig 5 is an exemplary segmentation of retinal FOV where each angular segment is represented with a cell in the RST; the row number and column number corresponds to an angular sector of the entire retinal field of view area element, and the value in the cell represents the sensitivity of the area as follows: the highest level of sensitivity has the score 1, the second level of sensitivity has the score 2, and so forth, where complete loss of vision is represented with the highest number.
  • the sensitivity scale is decided by the resolving capability of the applied diagnostic test; optionally also the number and location of the cells are determined according to the capabilities of the applied diagnostic test.
  • the retinal map is transformed into the RST according to one or more conversion factors; for example, if the resolution of the diagnostic test is lower than expected for conversion to the RST, then the conversion process needs to accommodate this lowered resolution in order for all cells of the RST to be populated with data.
  • the placement of the eye, and specifically of the pupil, during the diagnostic test from which the retinal map is obtained also needs to be considered with regard to the transformation to the RST, so that any movements of the pupil away from the placement during the test may be suitably compensated.
  • the diagnostic test is applied more than once, so that the retinal map is obtained a plurality of times with the pupil in a plurality of different locations, in order to assist with the transformation process and also with calibration of the RST according to pupil movements.
  • the light signal is scanned in horizontal and vertical directions and projected onto the user eye the light signal is scanned over the user retina, horizontally and vertically. Since the RST represents a map of the retina where each cell corresponds to a retinal sector or pixel, it is important that the RST is arranged according to the light scan trajectory.
  • RST represents the absolute sensitivity of different sectors of the retina, which is non-uniform (non-homogenous) even for a normal retina, i.e., of a healthy individual with unimpaired vision. Certain portions of the retina are more sensitive to the light than other portions even in such an individual. For individuals with diseased or damaged retinas, the lack of uniformity may be much more pronounced, even to the point of having non-functioning parts of the retina.
  • a relative RST (RRST) is constructed, where sensitivity values or the RST obtained for an individual are normalized relative to standard RST, obtained for individuals with normal vision. Such a standard RST may optionally be obtained for a particular population or alternatively is obtained for a number of individuals with healthy, normal vision, and is then used for calibration of RSTs obtained from individuals that are to receive the retinal projector apparatus as described herein.
  • Fig 6 is an example of the first 10 rows by 10 columns of an illustrative RRST, for a retina which demonstrates reduced sensitivity within the center of this region, such that sectors corresponding to row 6, column 5, and row 6 column 6 (written as (6,5) and (6,6) for short) correspond to lower functionality and hence greater loss of vision, while the peripheral pixels of this region still have normal sensitivity as compared to healthy individuals.
  • Figure 7 relates to an exemplary, illustrative method for receiving an incoming image and for projecting that image appropriately onto the retina of the subject through the retinal projection apparatus, for example as described herein.
  • the method may be generalized for a plurality of images by repeating the process described herein for one image.
  • the external digital image is converted to respective video signals for a plurality of pixels.
  • these video signals for the plurality of pixels are adjusted at least in intensity according to the RRST, to form transformed video signals.
  • the method is adjusted if there are some parts of the retina that are so damaged or reduced in function that they are practically inoperative, at least in terms of receiving projected light, as described in greater detail below with regard to Figures 10.
  • all (or nearly all) of the pixels of the retina receive projected light.
  • RRST RRST
  • the upper plot in Figure 8 shows exemplary analog video signal derived from the camera of a raster scan; the middle plot shows the retinal position of the light signal during the scan.
  • the video signal corresponds to the value Vi, at the same time the retinal position is Pi.
  • the video signal has the value V 2 , while the retinal position at that instance is P 2 .
  • the lower plot shows the video signal along the retinal position.
  • the upper plot is a video signal resulting from the camera, here shown as a digital signal for clarity.
  • the middle plot indicates the relative sensitivity scores of the retina according the position of the retina.
  • the light source (a laser, LED or other light source) is modulated with signal which is the multiplication of video data (signals) from the camera with the value of the RRST given and corresponding to the video data.
  • the function of the RRST can be also understood as follows.
  • the user creates the image of the observed scene on the retina.
  • To every point in the observed scene (object) corresponds a small area element on the retina.
  • an image is acquired by a digital camera. Every pixel in the camera image corresponds to particular point in observed object. Therefore every pixel in the camera image is associated with an area element in the user retina.
  • the camera image is intensified in those pixels associated with the retinal elements with a reduced sensitivity.
  • the RRST contains the information where the image should be intensified, and the amplification factor. Mathematically, if the image of the camera is represented by a matrix M, the RRST is represented by the matrix R, then the intensity corrected matrix A results in
  • the video processor before conveying the signal to intensity modulator preferably processes the projected video frames by multiplication with the RRST as explained.
  • the RSST is determined according to the relative position of the user pupil.
  • the VRD operates at magnification larger than xl only a portion is available to the user at a specific location of the pupil. The user moves his pupil to observe the entire panoramic view. Since the RSST is affixed to the pupil the mapping of RSST needs to be processed only on the projected image which is smaller than the buffered image external video image.
  • stage 3 the position of the eye, and hence the relative location and angle of the pupil is determined by the eye tracking mechanism.
  • stage 4 the video signals are adjusted according to the position of the eye: more specifically, the eye tracker provides the feedback to offset to the scanners mirrors so that the exit pupil of the Retinal Display (i.e. - the light projector) is approximately aligned with the entrance pupil of the eye.
  • the instantaneous position of the pupil is also feedback to the video processor, such that only the video signal associated to the visible window as determined by the instantaneous position of the pupil is transferred to logic unit and multiplied according to the RSST and its associated sync signals are fed to the scanner mirrors.
  • stage 5 light is projected onto the retina according to the transformed video signals.
  • Stages 3-5 are optionally and preferably repeated for each row or column of pixels for the transformed video signals, as necessary in order to project the entire image onto the retina.
  • an exemplary, illustrative method for projecting light onto a retina in which at least a portion of the retina is effectively non-functional.
  • effectively non-functional it is meant that the specific portion of the retina cannot respond with sufficient functionality to be able to effectively receive projected light.
  • Such effectively non-functional areas may include for example scotomas.
  • FIG. 10 shows an image of letters ABC represented with the retinal projector as raster scan of row 3 and 4.
  • the second diagram of Figure 10 (10b) represents how the projected image appears in field of vision with stocoma at the central portion.
  • the third diagram of Figure 10 (10c) shows the letters ABC when the raster image was projected at rows 5 to 6. The degree of severity of stocoma and its position within the retina is deduced in the RRST. If the region where significant part of field of vision is degenerated to the extent of complete vision loss then according to the present invention the respective video information is translated to the raster scanned rows corresponding to healthy sectors.
  • the circumvention of degenerated sector in the retina can be in non-continuous signals such that the projected image has certain "black" rows corresponding to rows where light is not projected.
  • the projected image can appear as in the last graph (lOd) of Fig 10 in which the image was broken and re-projected at rows ahead.
  • Figure 11 relates to an exemplary, illustrative method according to at least some embodiments of the present invention for receiving an incoming image and for projecting that image appropriately onto the retina of the subject through the retinal projection apparatus, for example as described herein, in which one or more portions of the retina are essentially non-functional.
  • the method may be generalized for a plurality of images by repeating the process described herein for one image.
  • stage 1 may optionally be performed as for Figure 7.
  • stage 2 due to one or more essentially non-functioning parts of the retina, at least a portion of these video signals for the plurality of pixels are adjusted both in intensity and also in location according to the RRST, to form transformed video signals. At least some video signals may optionally be adjusted only in intensity, but for the sake of clarity, the description herein focusses on those signals which need to be adjusted both in intensity and location.
  • stage 3 the position of the eye, and hence the relative location and angle of the pupil according to a previously determined position during the previously described diagnostic test, is determined.
  • stage 4 the transformed video signals are adjusted according to the position of the eye and more specifically, any differences between the current location and angle of the pupil, and the previously determined location and angle of the pupil during diagnostic tests. For example, depending upon movement of the pupil a "high" signal at a pixel may in fact be transformed to a "low” signal to be projected at a pixel, effectively shifting and transforming the map according to the current pupil position, but may also determine whether a signal is sent at all, as described in greater detail below.
  • adjusting the location of the video signals is performed as follows. Translating a portion of the image from degenerated to healthy portion of the retina is accomplished by introducing a controlled delay of the video signal relative to the raster scanner.
  • Fig 12 shows the schematics of an exemplary, illustrative scan over the field of vision where the indicated path, in the shape of an "S-curve" with a plurality of inflection points, indicates the trajectory of the scan. The frame starts at time indicated by t 0 . With a bi-directional raster scan t indicates the start time of row number 1, t 2 indicates the start time of the second row, and similarly for subsequent lines, where r represent indicate the begin time of n th -row scan.
  • the laser is intensity modulated with respect to the video information of the camera.
  • the video information associated with row# 1 starts at ti and ends at ti+T.
  • the time t 2 - (ti+T) to beginning of raw # 2 is "dead" time required for the scanner to re-position vertically to assume the raster scan of subsequent row.
  • the plots on the left hand side in FIG 12 show exemplary video signal according which the laser is intensity modulated. For each raw there corresponds a video signal.
  • a typical image may include hundreds of raster rows, in the example only the first four rows are shown.
  • Plots on the right hand side of Fig 12 illustrate an example when the video signal corresponding originally to rows #3 and rows #4 is delayed and instead appears shifted to row #5, while during the scan of rows #3 and #4 the laser is turned off.
  • the original video signal that was corresponding to subsequent rows has also has been shifted by one row.
  • the video information of the last row will be omitted from the video frame.
  • the last row, or even a number of rows if the shift is more than one row correspond to the peripheral of the field of vision and very often are less critical for vision related activities such as for example reading.
  • stage 5 Evaluating the best strategy for re-projecting and more precisely which rows are skipped depends largely on the severity of retinal degeneration at its central regions information which is stored in the RSST.
  • stage 5 light is projected onto the retina according to the transformed video signals.
  • Stages 3-5 are optionally and preferably repeated for each row or column of pixels for the transformed video signals, as necessary in order to project the entire image onto the retina.

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Optics & Photonics (AREA)
  • Health & Medical Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Biomedical Technology (AREA)
  • Ophthalmology & Optometry (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Vascular Medicine (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Animal Behavior & Ethology (AREA)
  • General Health & Medical Sciences (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Eye Examination Apparatus (AREA)
  • Chemical And Physical Treatments For Wood And The Like (AREA)
  • Illuminated Signs And Luminous Advertising (AREA)
PCT/IL2013/050702 2012-08-24 2013-08-19 Visual aid projector WO2014030158A1 (en)

Priority Applications (11)

Application Number Priority Date Filing Date Title
BR112015003899A BR112015003899A2 (pt) 2012-08-24 2013-08-19 projetor para auxílio visual
JP2015528004A JP2015526187A (ja) 2012-08-24 2013-08-19 視覚補助プロジェクタ
CN201380055324.9A CN104812342B (zh) 2012-08-24 2013-08-19 视觉帮助投影仪
EP13830523.0A EP2887913A4 (en) 2012-08-24 2013-08-19 VISUAL AID PROJECTOR
RU2015108858A RU2015108858A (ru) 2012-08-24 2013-08-19 Проектор для коррекции зрения
CA2882747A CA2882747A1 (en) 2012-08-24 2013-08-19 Visual aid projector
KR1020157007083A KR20150048789A (ko) 2012-08-24 2013-08-19 시각 보조 프로젝터
AU2013307201A AU2013307201B2 (en) 2012-08-24 2013-08-19 Visual aid projector
US14/423,668 US10058454B2 (en) 2012-08-24 2013-08-19 Visual aid projector for aiding the vision of visually impaired individuals
IL237384A IL237384A0 (en) 2012-08-24 2015-02-23 Sight aid projector
IN1864DEN2015 IN2015DN01864A (GUID-C5D7CC26-194C-43D0-91A1-9AE8C70A9BFF.html) 2012-08-24 2015-03-07

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201261693247P 2012-08-24 2012-08-24
US61/693,247 2012-08-24

Publications (2)

Publication Number Publication Date
WO2014030158A1 true WO2014030158A1 (en) 2014-02-27
WO2014030158A4 WO2014030158A4 (en) 2014-04-17

Family

ID=50149514

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IL2013/050702 WO2014030158A1 (en) 2012-08-24 2013-08-19 Visual aid projector

Country Status (11)

Country Link
US (1) US10058454B2 (GUID-C5D7CC26-194C-43D0-91A1-9AE8C70A9BFF.html)
EP (1) EP2887913A4 (GUID-C5D7CC26-194C-43D0-91A1-9AE8C70A9BFF.html)
JP (1) JP2015526187A (GUID-C5D7CC26-194C-43D0-91A1-9AE8C70A9BFF.html)
KR (1) KR20150048789A (GUID-C5D7CC26-194C-43D0-91A1-9AE8C70A9BFF.html)
CN (1) CN104812342B (GUID-C5D7CC26-194C-43D0-91A1-9AE8C70A9BFF.html)
AU (1) AU2013307201B2 (GUID-C5D7CC26-194C-43D0-91A1-9AE8C70A9BFF.html)
BR (1) BR112015003899A2 (GUID-C5D7CC26-194C-43D0-91A1-9AE8C70A9BFF.html)
CA (1) CA2882747A1 (GUID-C5D7CC26-194C-43D0-91A1-9AE8C70A9BFF.html)
IN (1) IN2015DN01864A (GUID-C5D7CC26-194C-43D0-91A1-9AE8C70A9BFF.html)
RU (1) RU2015108858A (GUID-C5D7CC26-194C-43D0-91A1-9AE8C70A9BFF.html)
WO (1) WO2014030158A1 (GUID-C5D7CC26-194C-43D0-91A1-9AE8C70A9BFF.html)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015132775A1 (en) * 2014-03-03 2015-09-11 Eyeway Vision Ltd. Eye projection system
WO2015183500A1 (en) * 2014-05-28 2015-12-03 Sightex LLC Relocated virtual retinal image system
JP2017012746A (ja) * 2015-06-30 2017-01-19 トムソン ライセンシングThomson Licensing 視線追跡デバイスおよび前記視線追跡デバイスを埋め込む頭部装着型デバイス
WO2018057373A1 (en) * 2016-09-22 2018-03-29 Microsoft Technology Licensing, Llc Display engines for use with optical waveguides
CN108319015A (zh) * 2017-04-21 2018-07-24 北京耐德佳显示技术有限公司 视网膜投影式近眼显示装置
JP2019049724A (ja) * 2018-10-01 2019-03-28 アイウェイ ビジョン エルティーディー. 目用投影システム
EP3693061A1 (en) * 2019-02-05 2020-08-12 Gensight Biologics Method for controlling an optogenetic device using filtering and associated devices
US11054639B2 (en) 2014-03-03 2021-07-06 Eyeway Vision Ltd. Eye projection system
US12066624B2 (en) 2020-01-06 2024-08-20 Eyeway Vision Ltd. Eye tracking device and method thereof

Families Citing this family (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI546062B (zh) * 2014-06-09 2016-08-21 國立交通大學 人工視網膜系統
US9752761B2 (en) 2014-07-16 2017-09-05 Telebrands Corp. Landscape light
USD773707S1 (en) 2014-10-30 2016-12-06 Telebrands Corp. Landscape light
KR102630754B1 (ko) 2015-03-16 2024-01-26 매직 립, 인코포레이티드 증강 현실 펄스 옥시미트리
US9952434B2 (en) * 2015-04-22 2018-04-24 The Florida International University Board Of Trustees Device for assisting visually impaired patients
USD816890S1 (en) 2015-05-11 2018-05-01 Telebrands Corp. Light projector
USD766483S1 (en) 2015-05-11 2016-09-13 Telebrands Corp. Light projector
USD824066S1 (en) 2015-05-11 2018-07-24 Telebrands Corp. Light projector
USD778478S1 (en) 2015-05-11 2017-02-07 Telebrands Corp. Light projector
US9879847B2 (en) 2015-12-03 2018-01-30 Telebrands Corp. Decorative lighting apparatus having two laser light sources
JP6923552B2 (ja) 2016-04-08 2021-08-18 マジック リープ, インコーポレイテッドMagic Leap,Inc. 可変焦点レンズ要素を用いた拡張現実システムおよび方法
JP6255522B2 (ja) * 2016-06-09 2017-12-27 株式会社Qdレーザ 視野視力検査システム、視野視力検査装置、視野視力検査方法、視野視力検査プログラム及びサーバ装置
JP6964239B2 (ja) * 2016-08-31 2021-11-10 パナソニックIpマネジメント株式会社 表示装置
CN109791278B (zh) * 2016-09-30 2022-05-03 三美电机株式会社 光扫描装置以及视网膜扫描型头戴式显示器
US11160688B2 (en) * 2016-11-10 2021-11-02 Samsung Electronics Co., Ltd. Visual aid display device and method of operating the same
EP3586176B1 (en) 2017-02-23 2024-01-10 Magic Leap, Inc. Variable-focus virtual image devices based on polarization conversion
JP6891589B2 (ja) * 2017-03-28 2021-06-18 セイコーエプソン株式会社 画像表示装置およびヘッドマウントディスプレイ
WO2018209118A2 (en) * 2017-05-10 2018-11-15 The Johns Hopkins University Providing sensory stimulations via photoacoustic, piezo-based, thermal, and/or electrical effects
CA3062740A1 (en) 2017-05-29 2018-12-06 Eyeway Vision Ltd Image projection system
IL252585A0 (en) * 2017-05-29 2017-08-31 Eyeway Vision Ltd System and method for projection to the eye with focus control
US11127212B1 (en) * 2017-08-24 2021-09-21 Sean Asher Wilens Method of projecting virtual reality imagery for augmenting real world objects and surfaces
US10386645B2 (en) 2017-09-27 2019-08-20 University Of Miami Digital therapeutic corrective spectacles
US10742944B1 (en) 2017-09-27 2020-08-11 University Of Miami Vision defect determination for facilitating modifications for vision defects related to double vision or dynamic aberrations
US10389989B2 (en) 2017-09-27 2019-08-20 University Of Miami Vision defect determination and enhancement using a prediction model
US10409071B2 (en) 2017-09-27 2019-09-10 University Of Miami Visual enhancement for dynamic vision defects
US10531795B1 (en) 2017-09-27 2020-01-14 University Of Miami Vision defect determination via a dynamic eye-characteristic-based fixation point
US10175490B1 (en) * 2017-12-20 2019-01-08 Aperture In Motion, LLC Light control devices and methods for regional variation of visual information and sampling
US10313645B1 (en) * 2018-01-19 2019-06-04 Microsoft Technology Licensing, Llc Variable emission period for scanned-beam display
FR3089785B1 (fr) * 2018-12-17 2020-11-20 Pierre Briand Dispositif médical d’aide à la perception d’environnement pour des utilisateurs aveugles ou malvoyants
WO2020192941A1 (en) * 2019-03-28 2020-10-01 Pixium Vision Sa System for projecting a pattern of interest onto a retinal area of a human eye
EP3951478A4 (en) 2019-03-28 2022-12-21 QD Laser, Inc. Image relay device and image projection system
GB2595707B (en) * 2020-06-04 2023-02-15 Sony Interactive Entertainment Inc Gaze tracking apparatus and systems
NL2026709B1 (en) * 2020-10-20 2022-06-16 Rabbit Eyes B V Display device
WO2024160342A1 (en) * 2023-01-30 2024-08-08 Pixium Vision Sa A system for selectively projecting an object of interest onto a predefined retinal area

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5396303A (en) * 1990-08-17 1995-03-07 American Cyanamid Company System and method for projecting an image (such as an ERG pattern) onto the retina
US20020109819A1 (en) * 2001-02-15 2002-08-15 Tveye Inc. Method and apparatus for low bandwidth transmission of data utilizing of the human eye anatomy
WO2010062481A1 (en) * 2008-11-02 2010-06-03 David Chaum Near to eye display system and appliance

Family Cites Families (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5233373A (en) 1990-08-17 1993-08-03 Peters Daniel R System and method for projecting an image (such as an ERG pattern) significantly anterior to the posterior retina
US5653751A (en) 1994-12-07 1997-08-05 Samiy; Nassrollah Systems and methods for projecting an image onto a retina
US6396461B1 (en) * 1998-08-05 2002-05-28 Microvision, Inc. Personal display with vision tracking
US6813085B2 (en) * 2000-06-26 2004-11-02 Angus Duncan Richards Virtual reality display device
JP3709921B2 (ja) * 2001-03-26 2005-10-26 英二 志水 眼疾患者への映像付与方法及び映像付与装置
DE10127367A1 (de) 2001-06-06 2002-12-12 Klaus Dietrich Vorrichtung und Verfahren zur Laserprojektion hochauflösender Bilder auf die Netzhaut des Auges, überlagert mit dem Bildinhalt des Gesichtsfeldes
US20040136570A1 (en) * 2002-04-30 2004-07-15 Shimon Ullman Method and apparatus for image enhancement for the visually impaired
US7001427B2 (en) * 2002-12-17 2006-02-21 Visioncare Ophthalmic Technologies, Inc. Intraocular implants
WO2005072667A1 (ja) * 2004-01-29 2005-08-11 Konica Minolta Photo Imaging, Inc. 視覚補助表示装置
KR20060016545A (ko) * 2004-08-18 2006-02-22 삼성전자주식회사 빔 스캐닝 광학 시스템
EP2143273A4 (en) * 2007-04-02 2012-08-08 Esight Corp DEVICE AND METHOD FOR SUPPRESSION
US8554327B2 (en) * 2010-11-24 2013-10-08 Second Sight Medical Products, Inc. Method and apparatus for predicting and controlling the percepts induced by a visual prosthesis
WO2010009447A2 (en) * 2008-07-18 2010-01-21 Doheny Eye Institute Optical coherence tomography - based ophthalmic testing methods, devices and systems
US20100149073A1 (en) * 2008-11-02 2010-06-17 David Chaum Near to Eye Display System and Appliance
CN201269949Y (zh) * 2008-09-09 2009-07-08 东南大学 直接在视网膜上实现三维立体成像的装置
US8130262B2 (en) * 2009-01-15 2012-03-06 International Business Machines Corporation Apparatus and method for enhancing field of vision of the visually impaired
WO2011012148A1 (en) * 2009-07-31 2011-02-03 Carl Zeiss Smt Gmbh Optical beam deflecting element and method of adjustment
JP2012078532A (ja) * 2010-09-30 2012-04-19 Brother Ind Ltd 画像表示装置

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5396303A (en) * 1990-08-17 1995-03-07 American Cyanamid Company System and method for projecting an image (such as an ERG pattern) onto the retina
US20020109819A1 (en) * 2001-02-15 2002-08-15 Tveye Inc. Method and apparatus for low bandwidth transmission of data utilizing of the human eye anatomy
WO2010062481A1 (en) * 2008-11-02 2010-06-03 David Chaum Near to eye display system and appliance

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP2887913A4 *

Cited By (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI710797B (zh) * 2014-03-03 2020-11-21 以色列商愛威願景有限公司 眼睛投影系統
US10539789B2 (en) 2014-03-03 2020-01-21 Eyeway Vision Ltd. Eye projection system
IL271410B2 (en) * 2014-03-03 2023-02-01 Eyeway Vision Ltd Eye projection system
IL271410B (en) * 2014-03-03 2022-10-01 Eyeway Vision Ltd Eye projection system
US11054639B2 (en) 2014-03-03 2021-07-06 Eyeway Vision Ltd. Eye projection system
WO2015132775A1 (en) * 2014-03-03 2015-09-11 Eyeway Vision Ltd. Eye projection system
US10042161B2 (en) 2014-03-03 2018-08-07 Eyeway Vision Ltd. Eye projection system
US20180341107A1 (en) * 2014-03-03 2018-11-29 Eyeway Vision Ltd. Eye projection system
CN106164743B (zh) * 2014-03-03 2020-10-27 埃韦视觉有限公司 眼睛投影系统
TWI666467B (zh) * 2014-03-03 2019-07-21 以色列商愛威願景有限公司 眼睛投影系統
AU2014385317B2 (en) * 2014-03-03 2019-11-07 Eyeway Vision Ltd. Eye projection system
WO2015183500A1 (en) * 2014-05-28 2015-12-03 Sightex LLC Relocated virtual retinal image system
JP2017012746A (ja) * 2015-06-30 2017-01-19 トムソン ライセンシングThomson Licensing 視線追跡デバイスおよび前記視線追跡デバイスを埋め込む頭部装着型デバイス
US9959818B2 (en) 2016-09-22 2018-05-01 Microsoft Technology Licensing, Llc Display engines for use with optical waveguides
WO2018057373A1 (en) * 2016-09-22 2018-03-29 Microsoft Technology Licensing, Llc Display engines for use with optical waveguides
CN108319015A (zh) * 2017-04-21 2018-07-24 北京耐德佳显示技术有限公司 视网膜投影式近眼显示装置
JP2019049724A (ja) * 2018-10-01 2019-03-28 アイウェイ ビジョン エルティーディー. 目用投影システム
EP3693061A1 (en) * 2019-02-05 2020-08-12 Gensight Biologics Method for controlling an optogenetic device using filtering and associated devices
WO2020161117A1 (en) 2019-02-05 2020-08-13 Gensight Biologics Method for controlling an optogenetic device using filtering and associated devices
US11986668B2 (en) 2019-02-05 2024-05-21 Gensight Biologics Method for controlling an optogenetic device using filtering and associated devices
US12066624B2 (en) 2020-01-06 2024-08-20 Eyeway Vision Ltd. Eye tracking device and method thereof

Also Published As

Publication number Publication date
CA2882747A1 (en) 2014-02-27
CN104812342A (zh) 2015-07-29
KR20150048789A (ko) 2015-05-07
EP2887913A1 (en) 2015-07-01
CN104812342B (zh) 2018-05-11
WO2014030158A4 (en) 2014-04-17
AU2013307201A1 (en) 2015-03-12
AU2013307201B2 (en) 2017-02-23
RU2015108858A (ru) 2016-10-20
BR112015003899A2 (pt) 2017-07-04
JP2015526187A (ja) 2015-09-10
US10058454B2 (en) 2018-08-28
EP2887913A4 (en) 2016-04-13
IN2015DN01864A (GUID-C5D7CC26-194C-43D0-91A1-9AE8C70A9BFF.html) 2015-07-31
US20150238362A1 (en) 2015-08-27

Similar Documents

Publication Publication Date Title
AU2013307201B2 (en) Visual aid projector
JP7709198B2 (ja) 画像投影システム
US20040227989A1 (en) Microscopy system for eye surgery and method of illumination
JP2019537149A (ja) 視線追跡装置および頭部装着型表示装置
HK1256558A1 (zh) 眼睛投影系统和方法
WO2019177728A1 (en) Single chip superluminous light emitting diode array for waveguide displays
US10452138B1 (en) Scanning retinal imaging system for characterization of eye trackers
US20180285642A1 (en) Head Mounted Display
WO2016072272A1 (ja) 視覚検査装置およびヘッドマウント型表示装置
WO2018025125A1 (en) Method of retinal sensitivity mapping using near-eye display
US10485419B2 (en) Optical imaging apparatus and method for controlling the same
WO2009101960A1 (ja) 画像表示装置
US11571122B2 (en) Ophthalmologic image processing method and fundus imaging apparatus
KR20160022110A (ko) 근적외선 광원을 이용한 안과 수술용 근적외선 입체 영상 시스템
US20210059520A1 (en) Ophthalmic systems and methods for direct retinal stimuli with local retinal angle of incidence control
JP7133163B2 (ja) 網膜走査型画像投影装置、網膜走査型画像投影方法、網膜走査型画像投影システム
US12142244B2 (en) Image display device and image display method
JP2021153786A (ja) 画像処理装置、画像処理方法及びプログラム
US20240115345A1 (en) Illumination system for surgical microscope with tunable coaxial and oblique beams
KR101534842B1 (ko) 양안용 시각 모의 실험 장치

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13830523

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2882747

Country of ref document: CA

ENP Entry into the national phase

Ref document number: 2015528004

Country of ref document: JP

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: 237384

Country of ref document: IL

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 14423668

Country of ref document: US

ENP Entry into the national phase

Ref document number: 2013307201

Country of ref document: AU

Date of ref document: 20130819

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: 2013830523

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 20157007083

Country of ref document: KR

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 2015108858

Country of ref document: RU

Kind code of ref document: A

REG Reference to national code

Ref country code: BR

Ref legal event code: B01A

Ref document number: 112015003899

Country of ref document: BR

ENP Entry into the national phase

Ref document number: 112015003899

Country of ref document: BR

Kind code of ref document: A2

Effective date: 20150224