WO2009120543A1 - Optical coherence tomography device, method, and system - Google Patents

Optical coherence tomography device, method, and system Download PDF

Info

Publication number
WO2009120543A1
WO2009120543A1 PCT/US2009/037448 US2009037448W WO2009120543A1 WO 2009120543 A1 WO2009120543 A1 WO 2009120543A1 US 2009037448 W US2009037448 W US 2009037448W WO 2009120543 A1 WO2009120543 A1 WO 2009120543A1
Authority
WO
WIPO (PCT)
Prior art keywords
coherence tomography
user
optical
optical coherence
eye
Prior art date
Application number
PCT/US2009/037448
Other languages
French (fr)
Inventor
Alexander C. Walsh
Paul G. Updike
Srinivas R. Sadda
Original Assignee
Doheny Eye Institute
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US12/111,894 external-priority patent/US8348429B2/en
Application filed by Doheny Eye Institute filed Critical Doheny Eye Institute
Publication of WO2009120543A1 publication Critical patent/WO2009120543A1/en

Links

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B3/00Apparatus for testing the eyes; Instruments for examining the eyes
    • A61B3/10Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions
    • A61B3/102Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for optical coherence tomography [OCT]

Definitions

  • Embodiments of the invention relate to the field of optical coherence tomography and, in particular, to devices, systems, methods of utilizing such optical coherence tomography data to perform precision measurements on eye tissue for the detection of eye diseases.
  • OCT optical coherence tomography
  • doctors generally desire a non-invasive, in vivo imaging technique for obtaining sub-surface, cross-sectional and/or three-dimensional images of translucent and/or opaque materials at a resolution equivalent to low-power microscopes. Accordingly, in the coming years, it is projected that there will be 20 million OCT scans performed per year on patients. Most of these will probably occur in the field of ophthalmology.
  • doctors or other medical professionals administer the OCT scans in the doctors' medical office or medical facilities.
  • optical coherence tomography which generally refers to an interferometric, non-invasive optical tomographic imaging technique, that can be used to detect and analyze, for example, eye tissue, and/or disease features, including but not limited to cystoid retinal degeneration, outer retinal edema, subretinal fluid, subretinal tissue, macular holes, drusen, or the like.
  • an optical coherence tomography instrument comprises an eyepiece for receiving at least one eye of a user; a light source that outputs light that is directed through the eyepiece into the user's eye; an interferometer configured to produce optical interference using light reflected from the user's eye; an optical detector disposed so as to detect said optical interference; electronics coupled to the detector and configured to perform a risk assessment analysis based on optical coherence tomography measurements obtained using said interferometer; and an output device electrically coupled to the electronics, and the output device may be configured to output the risk assessment to the user through the output device.
  • the optical coherence tomography instruments, devices, systems, and methods disclosed herein can be self-administered, and the eyepiece can be a monocular system or a binocular system.
  • a method for providing a self-administered optical coherence tomography test of a user's eye comprises automatically evaluating a plurality of fields of view of the optical coherence tomography instrument, each field of view corresponding to a different state of at least one optical component of said optical coherence tomography instrument; automatically adjusting said at least one optical component at least partly based on said evaluation; and with said at least one optical component adjusted, performing the optical coherence tomography test, wherein the optical coherence tomography instrument is configured to scan the at least one user eye.
  • the method step of automatically evaluating a plurality of fields of view of the optical coherence tomography instrument comprises rotating a galvanometer to direct a probe beam along different trajectories into the user eye.
  • the trajectories can intersect at a common location.
  • the common location comprises at least one of a location in or near the pupil of the user eye, a location in a plane of the iris of the user eye, a location within the lens of the user eye, and a location posterior to the pupil of the user eye.
  • the adjusting said at least one optical component comprises positioning the optical coherence tomography instrument along an anterior- posterior axis of the user.
  • the method of Claim 1 wherein said adjusting said at least one optical component comprises positioning said at least one optical component along an optical axis of said optical coherence tomography instrument.
  • the optical coherence tomography instrument has a working distance and said adjusting said at least one optical component alters the working distance of said optical coherence tomography instrument.
  • the at least one optical component can comprise at least one lens.
  • the evaluating the fields of view comprises collecting A-scan data.
  • the evaluating the fields of view can comprise collecting B-scan data.
  • the fields of view can comprise angular fields of view.
  • the method can further comprise providing the user with an output of the optical coherence tomography test from an output device.
  • the method can further comprise receiving the user eye via an eyepiece of the optical coherence tomography instrument; directing light from a source through the eyepiece into the user eye; producing optical interference using light reflected from the user eye; and detecting optical interference.
  • the method can further comprise automatically performing a diagnosis or risk assessment based on said detected optical interference, wherein the output can comprise the diagnosis or risk assessment.
  • the method can further comprise identifying one or more diseases or abnormal findings that the user may manifest indications of based on said detected optical interference, wherein the output can comprise a notification said identified diseases or abnormal findings.
  • a method for providing a self-administered optical coherence tomography test of a user's eye comprises determining a longitudinal position of the user relative to an optical coherence tomography instrument; automatically adjusting at least one optical component of said optical coherence tomography instrument at least partly based on said determined position thereby increasing the field of view of the optical coherence tomography instrument; and with said at least one optical component adjusted, performing the optical coherence tomography test, wherein the optical coherence tomography instrument is configured to scan the at least one user eye.
  • the determining a position of the user comprises detecting a portion of the user based on a sensor measurement.
  • the determining a position of the user can comprise determining the position of the user's eye.
  • the determining a position of the user can comprise radiating a signal to the user and detecting the signal returned from the user.
  • the signal can comprise light or ultrasound.
  • the signal can be reflected off of at least one of a cornea, an iris, a retina, a vitreous, an anterior chamber, and a tear film interface of the user.
  • the signal can be reflected off of at least one of an orbital rim, a nasal bridge, a cheekbone (maxilla), a frontal bone, an eyelid and a skin surface.
  • the determining a position of the user can comprise determining the position of the user ' s eye cornea or tear film interface.
  • the determining a position of the user can comprise determining the position of the user's eye pupil or iris.
  • the determining a position of the user can comprise determining the position of the user's retina.
  • the determining a position of the user can comprise analyzing optical coherence tomography data obtained by the optical coherence tomography instrument.
  • the method can further comprise automatically evaluating a plurality of fields of view of the optical coherence tomography instrument, each field of view corresponding to a different state of at least one optical component of said optical coherence tomography instrument, said automatically adjusting said at least one optical component being at least partly based on said evaluation.
  • the optical coherence tomography instrument can comprise rotating a galvanometer that directs a probe beam along different trajectories into the user eye, said trajectories intersecting at a common location, hi an embodiment, the automatically adjusting the at least one optical component positions the common location in a pupil of the user eye.
  • the adjusting the at least one optical component comprises positioning the optical coherence tomography instrument along an anterior- posterior axis of the user.
  • the adjusting the at least one optical component can comprise positioning the at least one optical component along an optical axis of said optical coherence tomography instrument.
  • the optical coherence tomography instrument can have a working distance and the adjusting of the at least one optical component can alter the working distance of the optical coherence tomography instrument.
  • the at least one optical component can comprise at least one lens.
  • the method can further comprise providing the user with an output of the optica) coherence tomography test from an output device.
  • the optical coherence tomography instrument can further comprise a rotational unit configured to rotate the trajectory of the probe beam into the user ' s eye.
  • the the rotational unit can comprise a galvanometer.
  • the rotational unit can cause the probe beam to propagate along a plurality of different trajectories that intersect at a common location and said at least one adjustable optical component is adjustable so as to place said common location in or near the pupil of the user eye.
  • the position of the optical coherence tomography instrument can be adjustable along an anterior-posterior axis of the user so as to adjust the position of the at least one adjustable optical component so as to adjust the field of view.
  • the at least one adjustable optical component can have a position that is adjustable along an optical axis of said optical coherence tomography instrument so as to adjust the field of view.
  • the optical coherence tomography instrument has a working distance and adjustment of said at least one optical adjustable component alters the work distance of said optical coherence tomography instrument.
  • the at least one optical component can comprise at least one lens.
  • the processor can be configured to compare a plurality of sets of A-scans, each set being associated with a different state of the adjustable optical component.
  • the processor can be configured to compare the field of view of different A-scan sets.
  • the processor can be configured to dete ⁇ nine a setting of the adjustable optical component selected based on monitored field of view data associated with each of the number of settings of the adjustable optical component.
  • the processor can be configured to determine a position of the adjustable optical component that increases the portion of the user ' s eye that can be imaged by the optical coherence tomography instrument.
  • the optical coherence tomography can further comprise an output device configured to output a result of the analysis to the user.
  • the optical tomography instrument can further comprise a sensor for determining a position of a structure of the user's eye, said processor being configured to control the position of the adjustable optical component based at least in part on the determined position of structure of the eye.
  • the structure can comprise the pupil, the iris, the cornea, the anterior chamber, the lens, the vitreous or the retina.
  • the sensor can comprise an optical sensor or an ultrasound sensor.
  • the optical coherence tomography instrument can further comprise an ultrasound emitter.
  • the optical coherence tomography instrument can further comprise a light emitter that directs an optical signal to said eye to be detected by said sensor.
  • the structure can comprise a pupil, iris or lens, or the like.
  • the structure can comprise a cornea or anterior chamber.
  • the structure can comprise a retina or the vitreous.
  • the at least one adjustable optical component can comprise at least one lens.
  • the at least one adjustable optical component can comprise an eyepiece.
  • the optical coherence tomography can further comprise an output device configured to output a result of the analysis to the user.
  • the processor can be configured to position the chin rest with respect to the eyepiece based at least partly on a no ⁇ native position based on across a patient population.
  • the processor can be configured to monitor a field of view of the optical coherence tomography instrument and position the chin rest with respect to the eyepiece based at least partly on said monitored field of view.
  • the processor can be configured to compare a plurality of sets of A-scans, each data set being associated with a different position of the chin rest relative to the eyepiece.
  • the processor can be configured to position the chin rest relative to the eyepiece at least partly based on a comparison of a plurality of fields of view from each of the plurality of A-scan data sets.
  • the optical coherence tomography instrument can further comprise a sensor to sense the user or part of the user.
  • the sensor can be configured to sense a position of the user ' s eye or structure of the user ' s eye.
  • the controller can be configured to position the chin rest relative to the eyepiece based at least partly on data collected by the sensor.
  • the optical coherence tomography instrument can further comprise a movable horizontal stage supporting at least one component of the optical coherence tomography instrument sensor, the stage being configured to adjust a medial-lateral position of the light being directed through the eyepiece into the user's eye.
  • an optical coherence tomography instrument comprises an eyepiece for receiving at least one eye of a user; a light source that outputs light that is directed through the eyepiece into the user ' s eye: an interferometer configured to produce optical interference using light reflected from the user ' s eye; an optical detector disposed so as to detect said optical interference; electronics coupled to the detector and configured to perform a risk assessment analysis based on optical coherence tomography measurements obtained using said interferometer; and an output device electrically coupled to the electronics.
  • the light from the light source can be controlled using a low performance (e.g., resolution) galvanometer and a high performance (e.g., resolution) galvanometer.
  • Figure ] is a schematic diagram of one embodiment of the optical coherence tomography system described herein.
  • Figure 2 is a schematic diagram of one embodiment of an interferometer arranged to perform measurements of an eye.
  • Figure 3A is a schematic diagram of one embodiment of an OCT system comprising a main body configured to conveniently interfere with a person's eyes, the main body being in communication with various systems as described herein.
  • Figure 3B is a perspective view schematically illustrating an embodiment of the main body shown in Figure 3A.
  • Figure 4 is schematic diagram of one embodiment of a spectrometer used to analyze data from an interferometer used for OCT.
  • Figure 5 is a schematic diagram of the main body of an OCT system comprising a single display for presenting a display target to a patient.
  • Figures 6A - 6C are schematic diagrams illustrating the use of optical coherence tomography to scan retinal tissue to generate A-scans and B-scans.
  • Figures 7A - 7F are schematic diagrams illustrating embodiments for adjusting and/or calibrating interpupillary distance.
  • Figure 8 is a block diagram schematically illustrating one embodiment of the computer system of the optical coherence tomography system described herein.
  • Figure 9 is illustrates a process flow diagram of one embodiment of performing precision measurements on retinal tissue for the detection of pathognomonic disease features.
  • Figures I OA- 1 OD illustrate possible embodiments of disposing the main body of an optical coherence tomography device with respect to a user.
  • Figures UA-I lB illustrate possible embodiments of output reports generated by the optical coherence tomography device.
  • Figure 12 is a block diagram schematically illustrating another embodiment of the computer system for an optical coherence tomography system described herein.
  • Figure 13 is a block diagram schematically illustrating components in one embodiment of the computer system for an optical coherence tomography system described herein.
  • Figure 14A is a diagram schematically illustrating one embodiment for determining a risk assessment.
  • Figure 14B is a schematic illustration of a plot of risk of retinal disease versus retinal thickness for detennining a risk assessment in another embodiment.
  • Figure 15 is an illustration of RPE detection and RPE polynomial fit curvature, and the difference there between.
  • Figure 16 is an illustration of retinal tissue segmented into inner and outer retinal tissue regions.
  • Figures 17A-C show B-scans obtained when the OCT system is positioned too far anterior, at a position that provides increased field of view, or too far posterior with respect to the eye.
  • Figures 18A-C show light beam trajectories when the OCT system is positioned too far anterior, at a position that provides increased field of view, or too far posterior wherein the intersection of the trajectoires is behind the pupil, at a pupil plane or in front of the pupil.
  • Figure ⁇ 9 shows one embodiment of a suitable working distance between an optical coherence tomography system and a patient's retina.
  • optical coherence tomography' and "OCT ' ' generally refer to an interferometric technique for imaging samples, in some cases, with micrometer lateral resolution.
  • This non-invasive optical tomographic imaging technique is used in ophthalmology to provide cross-sectional images of the eye, and more particularly the posterior of the eye, though it can also be used to image other samples or tissues in areas of the user ' s body.
  • OCT employs an interferometer.
  • a light source for example, a broadband light source
  • a beamsplitter for example, a beamsplitter
  • a portion of the light from the sample arm is reflected by the sample.
  • Light is also reflected from a mirror in the reference arm.
  • Light from the test arm and the reference arm is recombined, for example by the beamsplitter.
  • optical interference occurs, which affects the intensity of the recombined light.
  • the intensity of the combined reflected light varies depending on the sample properties. Thus, variations for the intensity of the reflectance measured are indications of the physical features of the sample being tested.
  • the length of the reference arm can be varied (for example, by moving one or more reference mirrors).
  • the reflectance observed as the reference arm distance changes indicates sample properties at different depths of the sample.
  • the length of the sample arm is varied instead of or in addition to the variation of the reference arm length.
  • the distance of the reference arm can be fixed, and the reflectance can then be measured at different frequencies.
  • the frequency of light emitted from a light source can be scanned across a range of frequencies or a dispersive element, such as a grating, and a detector array may be used to separate and detect different wavelengths.
  • Fourier analysis can convert the frequency- dependent reflectance properties to distance-dependent reflectance properties, thereby indicating sample properties at different sample depths.
  • OCT can show additional information or data than nonmydriatic color fundus imaging.
  • A-scan describes the light reflectivity associated with different sample depths.
  • B-scan refers to the use of cross-sectional views of tissues formed by assembly of a plurality of A-scans.
  • A-scans and B-scans can be used, for example, for differentiating normal and abnormal eye tissue or for measuring thicknesses of tissue layers in the eyes.
  • an A-scan can generally include data from the cornea to the retina.
  • a B-scan can include cross-sectional data from a medial border to a lateral border of the eye and from the comea to the retina.
  • a B-scan can include cross-sectional data from a superior border to an inferior border of the eye and from the cornea to the retina.
  • a 3D-OCT can be formed by combining a plurality of B-scans.
  • the terms "user” or ''patient” may be used interchangeably, and the foregoing terms comprise without limitation human beings, whether or not under the care of a physician, and other mammals.
  • eye scan is broad interchangeable terms that generally refer to the measurement of any part or substantially all of the eye, including but not limited to the cornea, the retina, the eye lens, the iris, optic nerve, or any other tissue or nerve related to the eye.
  • risk assessment and “diagnosis,” may be used in the specification interchangeably although the terms have different meanings.
  • the term “risk assessment” generally refers to a probability, number, score, grade, estimate, etc. of the likelihood of the existence of one or more illnesses, diseases, ailments, or the like.
  • the term “diagnosis” generally refers to a determination by examination and/or tests the nature and circumstances of an illness, ailment, or diseased condition.
  • Various methods, systems, and devices may be used to generate and utilize optical coherence tomography image data to perform precision measurements on retinal tissue for the detection of disease features, and generating a risk assessment and/or diagnosis based on data obtained by optical coherence tomography imaging techniques. These methods, systems and devices may employ, in some embodiments, a statistical analysis of the detected disease features obtained by optical coherence tomography imaging techniques. Such methods, systems, and devices can be used to screen for diseases.
  • computer system 104 is electrically coupled to an output device 102, a communications medium 108, and a user card reader system 1 12.
  • the communications medium 108 can enable the computer system 104 to communicate with other remote systems 1 10.
  • the computer system 104 may be electrically coupled to main body 106, which the user 1 14 positions near or onto the user's eyes.
  • the main body 106 is a binocular system ⁇ for example, has a two oculars or optical paths for the eyes providing one view for one eye and another view for another eye, or the like) configured to scan two eyes without repositioning the oculars with respect to the head of the patient, thereby reducing the time to scan a patient.
  • the eyes are scanned simultaneously using a scanner (for example, galvanometer), which provides interlaces of measurements from both eyes.
  • a scanner for example, galvanometer
  • the binocular system or a two ocular system having two respective optical paths to the two eyes can be configured to scan the eyes in series, meaning one eye first, and then the second eye.
  • serial scanning of the eyes comprises scanning a first portion of the first eye, a first portion of the second eye, a second portion the first eye, and so on.
  • the main body 106 can comprise a monocular system or one ocular system or optical path to the eye for performing eye scans.
  • the user 114 can engage handle 1 18 and position (for example, up, down, or sideways) the main body 106 that is at least partially supported and connected to a zero gravity arm 1 16, and accordingly the system 100 has no chin rest.
  • this configuration can introduce positioning error due to movement of the mandible.
  • the distance between the outermost lens (the lens closest to the user) and the user ' s eye can range between 10 mm and 30 mm, or 5 mm and 25 mm, or 5 mm and 10 mm.
  • the close proximity of the lens system to the user ' s eyes increases compactness of the system, reduces position variability when the patient places his eyes (for example, orbital rims) against the man body, and increases the viewing angle of the OCT apparatus when imaging through an undilated pupil.
  • the main body 106 can also comprise eyecups 120 (for example, disposable eyecups) that are configured to contact the user ' s eye socket to substantially block out ambient light and/or to at least partially support the main body 106 on the eye socket of the user 114.
  • the eyecups 120 have central openings (for example, apertures) to allow passage of light from the light source in the instrument to the eyes.
  • the eyecups 120 can be constructed of paper, cardboard, plastic, silicon, metal, latex, or a combination thereof.
  • the eyecups 120 can be tubular, conical, or cup-shaped flexible or semi-rigid structures with openings on either end. Other materials, shapes and designs are possible.
  • the eyecups 120 are constructed of latex that conforms around eyepiece portions of the main body 106.
  • the eyecups 120 are detachable from the main body 106 after the eye scan has been completed, and new eyecups 120 can be attached for a new user to ensure hygiene and/or to protect against the spread of disease.
  • the eyecups 120 can be clear, translucent or opaque, although opaque eyecups offer the advantage of blocking ambient light for measurement in lit environments.
  • the main body 106 may comprise one or more eyepieces, an interferometer, one or more target displays, a detector and/or an alignment system.
  • the optical coherence tomography system may comprise a time domain optical coherence tomography system and/or a spectral domain optical coherence tomography system.
  • the main body 106 comprises a spectrometer, (for example, a grating) and a detector array.
  • the main body may, in some embodiments, comprise signal processing component (for example, electronics) for performing, for example, Fourier transforms. Other types of optical coherence tomography systems may be employed.
  • FIG. 2 shows a diagram of an example optical coherence tomography system.
  • Light 150 is output from a light source 155.
  • the light source 155 may comprise a broadband light source, such as a superluminescent diode or a white light source. (Alternatively, light emitted from the light source 155 may vary in frequency as a function of time.)
  • the light 150 may comprise collimated light. In one embodiment, light 150 from the light source 155 is collimated with a collimating lens. The light is split at beamsplitter 160.
  • Beamsplitters may comprise without limitation a polarization-based beamsplitter, a temporally based beamsplitter and/or a 50/50 beamsplitter or other devices and configurations.
  • a portion of the light travels along a sample arm, directed towards a sample, such as an eye 165 of a user 1 14.
  • Another portion of the light 150 travels along a reference arm, directed towards a reference mirror 170.
  • the light reflected by the sample and the reference mirror 170 are combined at the beamsplitter 160 and sensed either by a one- dimensional photodetector or a two-dimensional detector array such as a charge-coupled device (CCD) or complementary metal-oxide-semiconductor (CMOS).
  • CCD charge-coupled device
  • CMOS complementary metal-oxide-semiconductor
  • a two-dimensional array may be included in a full field OCT instrument, which may gather information more quickly than a version that uses a one dimensional photodetector array instead.
  • the length of the reference arm (which may be determined in part by the position of the reference mirror 170) may be varying in time.
  • interference between the light reflected by the sample and the light reflected by the reference mirror/s occurs will depend on the length of the reference arm (as compared to the length of the test arm) and the frequency of the light emitted by the light source.
  • High contrast light interference occurs between light travelling similar optical distances (for example, (differences less than a coherence length).
  • the coherence length is determined by the bandwidth of the light source. Broadband light sources correspond to smaller coherence lengths.
  • the intensity of the output light may be analyzed as a function of time.
  • the light signal detected results from light rays scattered from the sample that interfere constructively with light reflected by the reference mirror/s. Increased interference occurs, however, when the lengths of the sample and reference arms are approximately similar (for example, within about one coherence length in some cases).
  • the light from the reference arm therefore, will interfere with light reflected from a narrow range of depths within the sample. As the reference (or sample) arms are translated, this narrow range of depths can be moved through the thickness of the sample while the intensity of reflected light is monitored to obtain information about the sample.
  • Samples that scatter light will scatter light back that interferes with the reference arm and thereby produce an interference signal.
  • Using a light source having a short coherence length can provide increased to high resolution (for example, 0.1-10 microns), as the shorter coherence length yields a smaller range of depths that is probed at a single instant in time.
  • the reference and sample arms are fixed.
  • Light from a broadband light source comprising a plurality of wavelengths is reflected from the sample and interfered with light reflected by the reference mirror/s.
  • the optical spectrum of the reflected signal can be obtained.
  • the light may be input to a spectrometer or a spectrograph comprising, for example, a grating and a detector array, that detects the intensity of light at different frequencies.
  • Fourier analysis performed, for example, by a processor may convert data corresponding to a plurality of frequencies to that corresponding to a plurality of positions within the sample.
  • data from a plurality of sample depths can be simultaneously collected without the need for scanning of the reference arm (or sample) arms. Additional details related to frequency domain optical coherence tomography are described in Vakhtin et al., (Vakhtin AB, Kane DJ, Wood WR and Peterson KA. "Common-path interferometer for frequency-domain optical coherence tomography, " Applied Optics. 42(34), 6953-6958 (2003)).
  • FIG. 3A shows one configuration of main body 106 comprising an optical coherence tomography system and an alignment system.
  • the main body 106 can include two eyepieces 203, each eyepiece configured to receive an eye from a user 114. In other embodiments, the main body 106 includes only one eyepiece 203.
  • Figure 3A shows one representative embodiment of an optical coherence tomography system.
  • Light from a light source 240 may propagate along a path that is modulated, for example, vertically and/or horizontally by one or more beam deflectors 280.
  • a galvanometer may be used for this purpose.
  • the galvanometer 280 can control the horizontal and/or vertical location of a light beam from the light source 240, thereby allowing a plurality of A-scans (and thus one or more B-scan and/or a 3D-OCT) to be formed.
  • the light from the light source 240 is split at beamsplitter 245.
  • beamsplitter 245 is replaced by a high frequency switch that uses, for example, a galvanometer, that directs about 100% of the light towards mirror 250a for about 1 A of a cycle and then directs about 100% of the light towards mirror 250b for the remainder of the cycle.
  • the light source 240 may include a broadband light source, such as a superluminescent light-emitting diode.
  • Light split at the beamsplitter 245 is then split again at beamsplitter 285a or 285b to form a reference arm and a sample arm.
  • a first portion of the light split at beamsplitter 285a or 285b is reflected by reference mirrors 273a or 273b, reference mirrors 270a or 270b. and reference mirrors 265a or 265b.
  • a second portion of the light split at beamsplitter 285a or 285b is reflected by mirror 250a or 250b. by mirror 255a or 255b and by mirror 260a or 260b.
  • Mirrors 255a or 255b and mirrors 250a and 250b are connected to a Z-offset adjustment stage 290b. By moving the position of the adjustment stage 290a or 290b, a different portion of the eye can be imaged.
  • the adjustment stage 290a or 290b can adjust the difference between the optical length from the light source 240 to a portion of the sample and the optical length from the light source 240 and the reference mirror 270a or 270b and/or reference mirror 273a or 273b.
  • This difference can be made small, for example, less than a coherence length, thereby promoting for optical interference to occur.
  • the positions of one or more reference mirrors are movable in addition to or instead of the adjustment stage being movable.
  • the length of the reference arm and/or of the sample arm may be adjustable.
  • the position of the adjustment stages 290a and/or 290b may be based on the signals from the device, as described in more detail below.
  • the light reflected by mirror 260a or 260b is combined with light from display 215a or 215b at beamsplitter 230a or 230b.
  • the displays 215a and 215b may comprise one or more light sources, such as in an emissive display like an array of matrix LEDs. Other types of displays can be used.
  • the display can display targets of varying shapes and configurations, including a bar and/or one or more dots.
  • a portion of the optical path from the light source 240 to the eye may be coaxial with a portion of the path from the displays 215a and 215b to the eye. These portions may extend though the eyepiece. Accordingly, a light beam from the light source 240 is coaxial with a light beam from the displays 215a and 215b such that the eyes can be positioned and aligned with respect to the eyepieces using the displays.
  • the user 1 14 may use images from the displays in order to adjust interpupilJary distance.
  • proper alignment of two images presented by the displays may indicate that the interpupillary distance is appropriately adjusted.
  • one or more adjustment controls 235 may be used to adjust the distance between the display targets 215a and 215b and/or between the eyepieces 203.
  • the adjustment controls 235 may be provided on the sides of the main body 106 or elsewhere.
  • the adjustment control 204 may comprise a handle on the main body 106, as shown in Figure 3B. In this embodiment, rotation of the adjustment control 204 may increase or decrease the interpupillary distance.
  • the combined light ⁇ that is reflected by mirror 260a or 260b and that comes from display 215a or 215b) is focused by adjustable powered optics (for example, lens) 210 possibly in conjunction with optical element 205.
  • the adjustable optics 210 may comprise a zoom lens or lens system that may be have, for example, a focal length and/or power that is adjustable.
  • the adjustable optics 210 may comprise or be part of an auto-focus system or may be manually adjusted.
  • the adjustable optics 210 may provide optical correction for those in need of such correction (for example, a user whose glasses are removed during testing)
  • the position of the powered optics 210 may be based on the signals obtained from the device, as described m more detail below
  • the focused light then travels through eyepiece windows or lens 205, positioned at a proximal end of the eyepiece 203, towards the eye of a user 114. hi the case where a lens 205 is included, this lens 205 may contnbute to focusing of the light into the eye.
  • This light directed into the eye may be scattered by tissue or features therein A portion of this scattered light may be directed back into the eyepiece Lens 205 may thus receive light 207 reflected from the user s eye.
  • which travels through the powered optics 210 reflects off of the beamsplitter 230a or 230b towards beamsplitter 220a or 220b, which reflects the light towards mirrors 295a or 295b
  • the sample arm includes the optical path between beamsplitter 285a or 285b and beamsplitter 295a or 295b that includes mirrors 250a or 250b and 255a or 255b and the sample or eye.
  • the light is then reflected by mirror 225a or 225
  • the instrument may be configured differently.
  • a common reference path may be used for each eye
  • the reference ami includes one or more movable mirrors to adjust the optical path length diffeience between the reference and sample arms.
  • Components may be added, removed, or repositioned in other embodiments Other techniques, may be used
  • polarizers and polarizing beamspitters may be used to control the propagation of light through the optical path in the optical system
  • Other variations are possible.
  • Other designs may be used [0075]
  • an A-scan may be formed in the time domain.
  • the Z-offset adjustment stage and corresponding mirror 255a or 255b and mirror 255a or 255b may change positions in time.
  • reference mirrors 270a and 270b and reference mirrors 273a and 273b or other mirrors in the reference or sample arms may be translated. The combined light associated with various mirror positions may be analyzed to determine characteristics of an eye as a function of depth.
  • an A-scan may be formed in the spectral domain.
  • the frequencies of the combined light may be analyzed to determine characteristics of an eye as a function of depth.
  • one or more galvanometers 280 can control the horizontal and/or vertical location of the A-scan.
  • a plurality of A-scans can be obtained to form a B-scan and/or a 3D-OCT.
  • Light output from the structure 275 can be input into a data acquisition device 202, which may comprise, for example, a spectrometer or a light meter.
  • a grating may be in the main body 106.
  • the data acquisition device 202 is coupled to a computer system 104, which may present output based on scans to the user 1 14.
  • the output device may include a monitor screen, in which output results are displayed.
  • the output device may include a printer, which prints output results.
  • the output device may be configured to store data on a portable medium, such as a compact disc or USB drive, or a custom portable data storage device.
  • the computer system 104 analyzes data received by the data acquisition device 202 in order to determine whether one or more of the adjustment stages 290a and/or 290b and/or one or more movable components and/or the powered optics 210 should be adjusted.
  • an A-scan is analyzed to determine a position (for example, a coarse position) of the retina such that data on the retina may be obtained by the instrument.
  • each A-scan comprises a plurality of light intensity values, each associated with a different depth into the sample. The A-scan may be obtained, in some embodiments, by translating the Z adjustment stage 290a or 290b.
  • the A- scan comprises values of reflected signal for obtained for different location of Z adjustment stage.
  • the retina reflects more light than other parts of the eye, and thus, it is possible to determine a position of the adjustment stage 290a or 290b that effectively images the retina by assessing what depths provide an increase in reflected intensity.
  • the Z adjustment stage may be translated and the intensity values may be monitored.
  • An extended peak in intensity for a number of Z adjustment stage positions may correspond to the retina.
  • a variety of different approaches and values may be monitored to determine the location of the retina. For example, multiple A-scans maybe obtained at different depths and the integrated intensity of each scan may be obtained and compared to determine which depth provided a peak integrated intensity.
  • intensity values within an A- scan can be compared to other values within the A-scan and/or to a threshold.
  • the intensity value corresponding to the preferred location may be greater than a preset or relative threshold and/or may be different from the rest of the intensity values, (for example, by more than a specified number of standard deviations).
  • a wide variety of approaches may be employed.
  • subsequent image analysis may be performed to account for vibration or movement of the user's head, eyes or retinas relative to the light source 240.
  • a feedback system such as a closed loop feedback system may be employed in effort to provide a more stabilized signal in the presence of such motion.
  • the optical coherence tomography signal may be monitored and feedback provided to, for example, one or more translation stages to compensate for such vibration or movement.
  • subsequent image analysis may be based on initial image and/or detect changes in image characteristics. For example, the image analysis may determine that the brightest pixel within an A-scan has moved 3 pixels from a previous scan. The adjustment stage 290a or 290b may thus be moved based on this analysis. Other approaches may be used.
  • optical coherence tomography signals are used to adjust the powered optics 210 to provide for increased or improved focus, for example, when a patient needs refractive correction.
  • Many users/patients for example, may wear glasses and may be tested while not wearing any glasses.
  • the powered optics 210 may be adjusted based on reflected signal to determine what added correction enhances signal quality or is otherwise an improvement.
  • a plurality of A-scans is analyzed in order to determine a position for the powered optics 2] 0.
  • a plurality of A- scans is analyzed in order to determine a position for the powered optics 210. In some embodiments, this determination occurs after the position of the adjustment stage 290a or 290b has been determined.
  • One or more A-scans, one or more B-scans or a 3D-OCT may be obtained for each of a plurality of positions of the powered optics 210. These scans may be analyzed to assess, for example, image quality. The position of the powered optics 210 may be chosen based on these image quality measures.
  • the image quality measure may include a noise measure.
  • the noise measure may be estimated based on the distribution of different intensity levels of reflected light within the scans. For example, lower signals may be associated with noise. Conversely, the highest signals may be associated with a saturated signal.
  • a noise measure may be compared to a saturation measure as in signal to noise ratios or variants thereof.
  • the lowest reflectivity measured (referred to as a low measure or low value) may also be considered.
  • the positions of the adjustment stages 290a and/or 290b and/or the powered optics 210 is determined based upon a signal-to-noise measure, a signal strength measure, a noise measure, a saturation measure, and a low measure. Different combinations of these parameters may also be used. Values obtained by integrating parameters over a number of positions or scans, etc., may also be used. Other parameters as well as other image quality assessments may also be used.
  • a noise value is estimated to be a reflected light value for which approximately 75% of the measured reflected light is below and approximately 25% of the measured reflected light is above.
  • the saturation value is estimated to be a reflected light value for which approximately 99% of the measured reflected light is below and approximately 1 % of the measured reflected light is above.
  • a middle value is defined as the mean value of the noise value and the saturation value.
  • An intensity ratio is defined as the difference between the saturation value and the low value divided by the low value multiplied by 100.
  • a tissue signal ratio is defined as the number of reflected light values between the middle value and the saturation value divided by the number of reflected light values between the noise value and the saturation value.
  • a quality value is defined as the intensity ratio multiplied by the tissue signal ratio.
  • a plurality of positions are tested.
  • the powered optics may be continuously moved in defined increments towards the eyes for each scan or set of scans.
  • the plurality of positions may depend on previously determined image quality measures. For example, if a first movement of the powered optics 210 towards the eye improved an image quality measure but a subsequent second movement towards the eye decreased an image quality measure, the third movement may be away from the eye. Accordingly, optical power settings may be obtained that improve and/or maintains an improved signal. This optical power setting may correspond to optical correction and increase focus of the light beam in the eye, for example, on the retina, in some embodiments.
  • various embodiments employ an arrangement wherein a pair of oculars is employed. Accordingly, such adjustments, may be applied to each of the eyes as a user may have eyes of different size and the retina may located at different depths and thus a pair of z adjust stages may be used in some embodiments. Similarly, a user may have different prescription optical correction for the different eyes.
  • measurements and/or adjustments may be performed and completed on one eye and subsequently performed and completed the other eye. Alternatively, the measurements and/or adjustments may be performed simultaneously or interlaced. A wide variety of other variations are possible.
  • FIG. 4 shows a diagram of a spectrometer 400 that can be used as a data acquisition device 202 for a frequency domain OCT system.
  • Light 405 input into the spectrometer 400 is collected by collecting lens 410.
  • the collected light then projects through a slit 415, after which it is collimated by the collimating lens 420.
  • the collimated light is separated into various spectral components by a grating 425.
  • the grating 425 may have optical power to focus the spectral distribution onto an image plane.
  • other separation components such as a prism may be used to separate the light.
  • the separated light is then directed onto a detector array by focusing lens 430, such that spectral components of each frequency from various light rays are measured.
  • a wide variety of OCT designs are possible. For example, frequency can be varied with time.
  • the reference and sample arms can overlap.
  • a reference arm is distinct from a sample arm, while in other embodiments, the reference arm and sample arm are shared. See, for example, Vakhtin AB, Kane DJ, Wood WR and Peterson KA. "Common-path interferometer for frequency-domain optical coherence tomography," Applied Optics. 42(34), 6953-6958 (2003).
  • the OCT arrangements should not be limited to those described herein. Other variations are possible.
  • the main body 106 includes only a single display target 215.
  • Light from the display target 215 is split at an x-prism 505.
  • other optical devices that split the source light into a plurality of light rays may be used. This split light is reflected at mirror 51 Oa or 510b and directed towards the user 1 14.
  • the user may be directed to fixate on a display target 215 while one or more galvanometers 280 move light from the light source 240 to image an area of tissue.
  • the display targets 215 are moved within the user ' s field of vision while an area of tissue is imaged.
  • a display target 215 may be moved horizontally (for example, in the medial-lateral direction), such that a patient is directed to look from left to right or from right to left.
  • a vertical scanner for example, galvanometer
  • Figure 6 shows an eye, which is directed to move in the horizontal direction 605.
  • the scanned trajectory 610 covers a large portion of the eye 600. Scanning in the vertical and horizontal directions can produce a 3D-OCT. In some embodiments, continuous and/or regularly patterned A-scans are combined to form a full scan for example, B-scan or 3D-OCT. In other embodiments, discrete and/or random A-scans are combined to form the full scan.
  • Systems configured such that users 114 are directed to move their eyes throughout a scan may include fewer scanners than comparable systems configured such that users 114 keep their eyes fixated at a stationary target. For example, instead of a system comprising both a vertical and a horizontal scanner, the user 114 may move his eyes in the horizontal direction, thereby eliminating the need for a horizontal scanner.
  • two scanners e.g., a vertical and a horizontal scanner
  • the design, capabilities and/or specifications for these scanners need not be the same.
  • one of the scanners may be faster and/or higher resolution than the other.
  • a vertical scanner may be used that scans more rapidly than a horizontal scanner, or vice versa.
  • Scanners such as galvanometers having different speeds may be used in some example embodiments to scan continuously in the vertical direction and only occasionally increment along the horizontal (or vice versa).
  • one of the scanners may be Vi to 1/500 as fast as the other scanner although values outside this range are possible.
  • the 3D-OCT image may not contain as many pixels in one direction (e.g. horizontal) as the other direction (e.g.. vertical). hi some embodiments, for example, the 3D-OCT image may be 600 x 512 although other sizes are possible.
  • one scanner or galvanometer may have a reduced resolution compared to the other scanner. In instances wherein the specifications for one scanner include slower scan rates or less resolution than the other scanner, possibly a less expensive scanner or galvanometer may be used. Accordingly, the two scanners or galvanometers need not be the same type or grade. A relatively high performance (higher cost) and a relatively lower performance (lower cost) scanner may be used.
  • Figure 6B shows an example of an A scan.
  • the A scan comprises the signal strength (indicated by the brightness) as a function of depth for one horizontal and vertical position.
  • an A-scan comprises a plurality of intensity values corresponding to different anterior-posterior positions.
  • a plurality of A scans form a B scan.
  • Figure 6C shows a B-scan, in which the largest portion of the bright signal corresponds to retinal tissue and the elevated region under the retina corresponds to diseased tissue within the eye.
  • Figure 7A there is illustrated an enlarged view depicting an embodiment of the main body 106 that is configured with a handle 1 18 for adjusting the eyepieces to conform to the user's interpupillary distance.
  • the main body 106 comprises a left eyepiece 712 and a right eyepiece 714 wherein each is connected to the other by interpupillary distance adjustment device 718.
  • the interpupillary distance adjustment device 718 is coupled to the handle 1 18, wherein the handle 1 18 is configured to allow the user to engage the handle 118 to adjust the distance between the left and right eyepieces 712, 714 to match or substantially conform to the interpupillary distance between the eyes of the user.
  • the user can rotate, turn, or twist the handle 118 to adjust the distance between the left and right eyepieces 712, 714 so as to match or substantially conform to the interpupillary distance between the eyes of the user.
  • the handle 1 18 can be configured to move side to side to allow the user to adjust the distance between the left and right eyepieces 712, 714.
  • the handle 1 18 can be configured to move forward and backward to allow the user to adjust the distance between the left and right eyepieces 712, 714.
  • the handle 1 18 can be configured to move up and down to allow the user to adjust the distance between the left and right eyepieces 712, 714.
  • the motor 714 can be adjusted and/or controlled by a motor activated by the user.
  • the motor can be configured to be controlled by computer system 104 to semi- automatically position the left and right eyepieces 712. 714 to match the interpupillary distance between the eyes of the user.
  • eye tracking devices may be included with a system described herein.
  • a combination of the foregoing are utilized to adjust the distance between the left and right eyepieces 712. 714 to match or substantially conform to the user's interpupillary distance.
  • a user 1 14 may adjust interpupillary distance based on the user ' s viewing of one or more fixation targets on one or more displays 215.
  • the displays 215 and the fixation targets may be configured such that the user views two aligned images, which may form a single, complete image when the interpupillary distance is appropriate for the user 114.
  • the user 1 14 may adjust (for example, rotate) an adjustment control 204 to change the interpupillary distance based on the fixation target images, as shown in Figure 7A.
  • Figures 7B-7F illustrate one embodiment of fixation targets as seen by the viewer under a plurality of conditions; however, other fixation targets are possible, including but not limited to a box configuration.
  • Figure 7B shows a U-shaped fixation target 715a on the display 215a for the left eye.
  • Figure 7C shows an upside-down U-shaped fixation target 715b on the display 215b for the right eye.
  • the bottom and top images 715a and 715b are aligned, as shown in Figure 7D to form a complete H-shaped fixation target 715.
  • the fixation target 715a on the display 215a for the left eye appear shifted to the right and the fixation target on the display 215b for the right eye appear shifted to the left and the user sees the image shown in Figure 7E.
  • the fixation target 715a on the display 215a for the left eye appear shifted to the left and the fixation target on the display 215b for the right eye appear shifted to the right and the user sees the image shown in Figure 7F .
  • the interpupillary distance may be adjusted based on these images.
  • the alignment image 715 is in the shape of an "H.”
  • the fixation targets on the left and right displays overlap to form an "H”.
  • Other alignment images 715 may be provided.
  • the computer system 104 can comprise a scan control and analysis module 824 configured to control the scanning operations performed by the main body 106.
  • the computer system 104 can also comprise a fixation marker control system 822 configured to display a fixation marker visible by the user from main body 106.
  • the fixation marker is displayed as an ''X, " a dot, a box, or the like.
  • the fixation marker can be configured to move horizontally, vertically, diagonally, circularly, or a combination thereof.
  • the fixation marker can be repositioned quickly to relocate the beam location on the retina as the eye repositions itself.
  • the computer system 104 can also comprise a focus adjust module 820 for automatically adjusting the focusing lenses in the main body 106 as further discussed herein.
  • the computer system 104 can also comprise a Z positioning module 818 for automatically adjusting the Z offset as herein discussed.
  • the computer system 104 comprises in the illustrative embodiment a disease risk assessment / diagnosis module 808 for storing and accessing information, data, and algorithms for determining, assessing the risk or likelihood of disease, and/or generating a diagnosis based on the data and/or measurements obtained from scanning the eyes of the user.
  • the scan control and analysis module 824 is configured to compare the data received from the main body 106 to the data stored in the disease risk assessment / diagnosis module 808 in order to generate a risk assessment and/or diagnosis of disease in the eyes of the user as further illustrated.
  • the computer system 104 can also comprise an image/scans database configured to store images and/or scans generated by the main body 106 for a plurality of users, and to store a unique identifier associated with each image and/or scan.
  • the scan control and analysis module 824 uses historical images and/or scans of a specific user to compare with current images and/or scans of the same user to detect changes in the eyes of the user.
  • the scan control and analysis module 824 uses the detected changes to help generate a risk assessment and/or diagnosis of disease in the eyes of the user.
  • the computer system 104 can comprise a user / patient database 802 for storing and accessing patient information, for example, user name, date of birth, mailing address, residence address, office address, unique identifier, age. affiliated doctor, telephone number, email address, social security number, ethnicity, gender, dietary history and related info ⁇ nation, lifestyle and/or exercise history information, use of corrective lens, family health history, medical and/or ophthalmic history, prior procedures, or other similar user information.
  • patient information for example, user name, date of birth, mailing address, residence address, office address, unique identifier, age. affiliated doctor, telephone number, email address, social security number, ethnicity, gender, dietary history and related info ⁇ nation, lifestyle and/or exercise history information, use of corrective lens, family health history, medical and/or ophthalmic history, prior procedures, or other similar user information.
  • the computer system 104 can also comprise a physician referral database for storing and accessing physician information, for example, physician name, physician training and/or expertise/specialty, physician office address, physician telephone number and/or email address, physician scheduling availability, physician rating or quality, physician office hours, or other physician information.
  • physician information for example, physician name, physician training and/or expertise/specialty, physician office address, physician telephone number and/or email address, physician scheduling availability, physician rating or quality, physician office hours, or other physician information.
  • the computer system 104 can also comprise a user interface module 805 (which can comprise without limitation commonly available input/output (I/O) devices and interfaces as described herein) configured to communicate, instruct, and/or interact with the user through audible verbal commands, a voice recognition interface, a key pad, toggles, a joystick handle, switches, buttons, a visual display, touch screen display, etc. or a combination thereof.
  • the user interface module 805 is configured to instruct and/or guide the user in utilizing and/or positioning the main body 106 of the optical coherence tomography system 100.
  • the computer system 104 can also comprise a reporting / output module 806 configured to generate, output, display, and/or print a report (for example, Figures 1OA and 10B) comprising the risk assessment and/or diagnosis generated by the disease risk assessment / diagnosis module 808.
  • the report comprises at least one recommended physician to contact regarding the risk assessment.
  • the computer system 104 can also comprise an authentication module 816 for interfacing with user card reader system 1 12. wherein a user can insert a user identification card into the user card reader system 1 12.
  • the authentication module 816 is configured to authenticate the user by reading the data from the identification card and compare and/or store the information with the data stored in the user / patient database 802.
  • the authentication module 816 is configured to read or obtain the user ' s insurance information from the user ' s identification card through the user card reader system 1 12.
  • the authentication module 816 can be configured to compare the user ' s insurance information with the data stored in the insurance acceptance database 828 to determine whether the user's insurance is accepted or whether the user ' s insurance company will pay for scanning the user ' s eyes. In other embodiments, the authentication module communicates with the billing module 810 to send a message and/or invoice to the user ' s insurance company and/or device manufacturer to request payment for performing a scan of the patient ' s eyes.
  • the card can activate one or more functions of the machine allowing the user, for example, to have a test performed or receive output from the machine.
  • the billing module 810 is configured to communicate with the user interface module 805 to request payment from the user to pay for all or some (for example, co-pay) of the cost for performing the scan.
  • the billing module 810 is configured to communicate with the user card reader system 112 to obtain card information from the user's credit card, debit card, gift card, or draw down credit stored on the user's identification card.
  • the billing module 810 is configured to receive payment from the user by communicating and/or controlling an interface device for receiving paper money, coins, tokens, or the like.
  • the billing module 810 is configured to receive payment from the user by communicating with the user's mobile device through Bluetooth® or other communications protocols/channels in order to obtain credit card information, billing address, or to charge the users mobile network service account (for example, the cellular carrier network).
  • the users mobile network service account for example, the cellular carrier network.
  • the user card may be used by insurers to track which users have used the system.
  • the system can print (on the face of the card) or store (in a chip or magnetic stripe) the scan results, risk assessment, and/or report directly onto or into the card that the patient inserts into the system (wherein the card is returned to the user).
  • the system can be configured to store multiple scan results, risk assessments, and/or reports, and/or clear prior scan results, risk assessments, and/or reports before storing new information on the magnetic stripe.
  • the calculation of the risk assessment is performed by the system (for example, scanning analysis module 824).
  • the calculated risk assessment is transmitted a centralized server system (for example, remote systems 1 10) in another location that provides the results via a web page to physicians, users, patients, or the like.
  • the centralized server system (for example, remote system 1 10) allows the user, patients, or doctors to enter their card code to see the results which are saved in the centralized database.
  • the computer system 104 can comprise a network interface 812 and a firewall 814 for communicating with other remote systems U O through a communications medium 108.
  • Other remote systems 1 10 can comprise without limitation a system for checking the status/accuracy of the optical coherence tomography system 100; a system for updating the disease risk assessment / diagnosis database 808, the insurance acceptance database 828, the physician referral database 804, and/or the scan control and analysis module 824.
  • the computer system 104 can be configured to communicate with a remote system 110 to conduct a primary and/or secondary risk assessment based on the data from scanning the user ' s eyes with the main body 106.
  • the remote system 1 10 can be configured to remotely perform (on an immediate, delayed, and/or batch basis) a risk assessment and/or diagnosis and transmit through a network or communications medium the risk assessment, diagnosis, and/or report to the computer system 104 for output to the user using output device 102.
  • the output device 102 is configured to display the risk assessment, diagnosis, and/or report as a webpage that can be printed, emailed, transmitted, and/or saved by the computer system 104.
  • the remote system 1 10 can also be configured to transmit through a network or communications medium the risk assessment, diagnosis, and/or report to the user ' s (or doctor) cellular phone, computer, email account, fax, or the like.
  • FIG. 10103 With reference to Figure 9, there is shown an illustrated method of using the optical coherence tomography system 100 to self-administer an OCT scan of the user ' s eyes and obtain a ⁇ sk assessment or diagnosis of various diseases and ailments.
  • the process begins at block 901 wherein the user approaches the optical coherence tomography system 100 and activates the system, by for example pushing a button or typing in a activation code or anonymous identification number, hi other embodiments, the user interface module 805 instructs users at block 901 to first insert an identification card or anonymous coded screening card in user card reader system 112 to activate the system.
  • the system can also be activated at block 901 when users insert their user identification card in user card reader system 112
  • Other means of activating the system are possible as well as, including without limitation, a motion sensor, a weight sensor, a radio frequency identification (RFID) device, or other actuator to detect the presence of the user.
  • the optical tomography system 100 can be activated when the billing module 810 detects that the user has inserted paper money, coins, tokens, or the like into an interface device configured to receive such payment.
  • the billing module 810 can also be configured to activate the optical tomography system 100 when the billing module 810 communicates with a user's mobile device in order to obtain the user ' s credit card information, billing address, or the like, or to charge the user ' s mobile network service account (for example, the cellular earner network)
  • the user interface module 805 is configured to direct the user to attach disposable eyecups onto the main body 106, and then position the main body 106 with the disposable eyecups near the eyes of the user and/or support the disposable eyecups against the user ' s eye socket.
  • the user interface module 805 instructs the user to engage handle 1 18 to adjust the distance between the left and right eyepieces 612, 614 to match or substantially conform to the interpupillary distance of the user as described with respect to Figures 6A-6F.
  • the user inputs into or indicates to the user interface module 805 to begin the scan.
  • the scan control and analysis module 824 substantially restricts movement or locks the position of the zero gravity arm and/or the distance between the left and right tubes 612, 614 to begin the scan.
  • the Z module 818 automatically adjusts the z-offset in the main body 106 at block 906 such that the OCT measurement will be obtained, for example, from tissue in the retina.
  • the Z module 818 may identify and/or estimate a position of part of the sample (for example, part of an eye of a user 1 14) and adjust the location of one or more optical components based on the position.
  • the Z module 818 may comprise a motor, such as a piezoelectric motor, to translate the reference mirror/s longitudinally such that the optical path length from the beam splitter to the retina is about equal to (within a coherence length of) the optical path length in the reference arm.
  • Block 910 illustrates an optional test performed by the computer system 104 to determine the visual function and/or acuity of the user's eye. Such visual function and/or acuity tests will be appreciated by those skilled in the art. In one embodiment, the visual acuity test works with or is combined with the fixation marker control system 722, and can test both eyes simultaneously or one eye at time.
  • fixation marker will initially appear small and then gradually increase in size until the user indicates through the user interface module 705 that the fixation marker is visible. Based on the size at which the user can clearly see the fixation marker, fixation marker control system 722 can estimate or determine or assess the visual acuity of the users eyes (for example, 20/20, 20/40, or the like).
  • the user interface module 805 instructs the user to follow the movement of the fixation marker that is visible to the user from the main body 106.
  • the fixation marker control 822 is configured to display a fixation marker that moves horizontally.
  • the horizontal movement of the fixation marker allows the scan control and analysis module 824 to scan the eye vertically as the eye moves horizontally, thus possibly obtaining a two-dimensional, volume, or raster scan of the eye tissue at issue.
  • the scan control and analysis module 824 and/or the fixation marker control may cause the fixation marker or the beam to jump or move around to obtain measurements at different lateral locations on the eye.
  • the scan control and analysis module 824 could be configured to detect at block 913 whether there has been a shift in the position of the main body 106 relative to the user.
  • the scan control and analysis module 824 can detect (in real-time, substantially real-time, or with a delay) whether a shift has occurred based on what the values the module 824 expects to receive during the scanning process. For example, as the scan control and analysis module 824 scans the retina, the module 824 expects to detect a change in signal as the scanning process approaches the optic nerve (for example, based on the location of the fixation target and/or state of the scanner(s)).
  • the expected values or the expected change in values can also be determined or generated using a nomogram. If the system does not detect an expected signal change consistent with a detection of the optic nerve and/or receives no signal change, then the module 824 can be configured to interpret such data as the user is not tracking properly. Other features, for example, the fovea, or the like, can be used to determine whether the expected signal is observed. If improper tracking occurs enough (based on, for example, a threshold), the system 100 may request that the user fixate again (using fixation marker control 822) for another scan.
  • the system can be configured to complete the scan, perform data analysis, and during the analysis the system can be configured to detect whether a shift occurred during the scan. If a substantial shift is detected, then the user may be instructed (through visual, audible, or verbal instructions using the user interface module 805) to sit forward again so another scan can be performed. If the system detects a shift 2 or 3 or more times, the system can be configured to refer the user to a general eye doctor.
  • the scan control and analysis module 824 can be configured to produce a confidence value that indicates how likely the nomograms wil! be to apply to this patient. For example, if the patient had borderline fixation, the confidence value might be lower than a patient whose fixation appeared to be good.
  • the system can be configured to perform rapid cross-correlations between adjacent A-scans or B-scans to make sure the eye is moving somewhat.
  • the foregoing can be advantageous for ANSI laser safety standards so as to avoid having users stare at the same location with laser energy bombarding the user's retina.
  • the system is configured with a laser time-out feature if the system detects no eye moment (for example, cross-correlations above a certain threshold).
  • signal data may be analyzed prior to performing an FFT.
  • Other technologies can be used to deterrmne that the user has some eye movement.
  • the scan control and analysis module 824 completes the scan of the user's eyes, stores the image and/or scan data in the images/scans database 826, and analyzes the A-scan data at block 915 to generate/determine a nsk assessment and/or diagnosis at block 916 by accessing the data and/or algorithms stored m the disease nsk assessment/diagnosis database 808.
  • groups of A-scans, partial or full B scans, or partial or full 3D-OCT data can be analyzed.
  • nomogram generally refers to predictive tools, algorithms, and/or data sets. Nomograms in general can provide predictions for a user based on the comparison of characteristics of the user with the nomogram.
  • the nomograms are de ⁇ ved, generated, calculated, or computed from a number, for example, hundreds, thousands, or millions of users/patients who exhibited the same condition (normal or diseased).
  • nomograms compare the risk of having a disease based on physical characteristics. Accordingly, in some cases, nomograms can provide individualized predictions that are relative to risk groupings of patient populations who share similar disease characteristics.
  • nomograms can be used to provide the risk estimation or risk assessment on a 0-100% scale.
  • nomograms used herein can provide an expected value, for example, at a certain position in the eye there is an expected eye thickness value of 100 microns.
  • nomograms have been developed and validated in large patient populations and are highly generalizable, and therefore, nomograms can provide the objective, evidence-based, individualized risk estimation or assessment. Accordingly, nomograms can be used as described herein to empower patients and allow them to better understand their disease. Further, nomograms as used herein can assist physicians with clinical decision-making and to provide consistent, standardized and reliable predictions.
  • an eye health assessment or eye health grade report is generated for the user by accessing the disease risk assessment / diagnosis database 808.
  • the physician referral database 804 is accessed to generate a recommendation of when the user should visit a physician (for example, within one to two weeks).
  • the physician referral database 804 is also accessed to generate, compile a listing of physicians suitable for treating the patient.
  • the physician referral list can be randomly generated or selected based on referral fee payments paid by physicians, insurance companies, or based on location of the physician relative to the user ' s present location or office/home address, or based on the type of detected disease, or based on the severity of the detected disease, based on the location or proximity of the system relative the location of the physician, or based on a combination thereof.
  • the report is displayed to the user by using reporting / output module 806 and output device 102.
  • the report data is stored in the user/patient database 802 for future analysis or comparative analysis with future scans.
  • the main body 106 is not supported by the user 1 14.
  • the main body 106 may be supported by a free-standing structure, as shown in Figure 1OA.
  • the user 1 14 may look into the eyepiece(s).
  • the user 1 14 may be seated on a seating apparatus, which may include a height-adjusting mechanism.
  • the main body 106 may supported by a height- adjustable support.
  • a strap 1005 is connected to the main body 106.
  • the strap may function to fully or partly support the main body 106, as shown in Figure 1OB.
  • the strap 905 may be excluded in some embodiments.
  • the main body 106 may be hand held by the user.
  • the main body 106 may be supported on eyewear frames, hi some embodiments, all of the optics are contained within the main body 106 that is directly or indirectly supported by the user 114.
  • the main body 106 in Figure 1OB may include an optical coherence tomography system, an alignment system, and a data acquisition device.
  • the data acquisition device may wirelessly transmit data to a network or computer system or may use a cable to transfer control signals.
  • Figure 1OC is similar to that of Figure 1 and is supported by a separate support structure (for example, an zero gravity arm).
  • a strap, belt, or other fastener assists in the alignment of the main body 106 with one or both eyes of the user 1 14.
  • the user wears an object 1010 connected to the eyepiece.
  • the wearable object 1010 may include a head-mounted object, a hat or an object to be positioned on a user's head.
  • the main body 106 is supported on an eyewear frame worn by the user like glasses.
  • the wearable object 1010 may fully or partly support the main body 106 and/or may assist in aligning the main body 106 with one or both eyes of the user 114.
  • the eye health grades report can comprise without limitation a numeric and/or letter grade for each eye of the user for various eye health categories, including but not limited to macular health, optic nerve health, eye clarity, or the like.
  • the eye health grades report can also comprise at least one recommendation to see or consult a physician within a certain period of time, and can provide at least one possible physician to contact. Data for generating the recommendation info ⁇ nation and the list of referral physicians are stored in the physician referral database 804.
  • the eye health assessment report can comprise a graphical representation for each eye of the user for various eye health categories. The report can be presented to the user on an electronic display, printed on paper, printed onto a card that the user inserted into the machine, electronically stored on the users identification card, emailed to the user, or a combination thereof.
  • the billing module 810 can be configured to communicate with billing / insurance reporting payment systems 1201 through communications medium 108 in order to request or process an insurance claim for conducting a scan of the user's eyes. Based on communications with billing / insurance reporting and payment system 1201, the billing module 810 can also be configured to determine the amount payable or covered by the user ' s insurance company and/or calculate or determine the co-pay amount to be charge the consumer.
  • the user can interact with the user interface module 805 to schedule an appointment with the one of the recommended physicians and/or schedule a reminder to be sent to the user to consult with a physician.
  • the computer system 104 or a remote system 1 10 can be configured to send the user the reminder via email, text message. regular mail, automated telephone message, or the like.
  • the systems, computer clients and/or servers described above take the form of a computing system 1300 shown in Fig. 13, which is a block diagram of one embodiment of a computing system (which can be a fixed system or mobile device) that is in communication with one or more computing systems 1310 and/or one or more data sources 1315 via one or more networks 1310.
  • the computing system 1300 may be used to implement one or more of the systems and methods described herein.
  • the computing system 1300 may be configured to process image files. While Fig. 13 illustrates one embodiment of a computing system 1300. it is recognized that the functionality provided for in the components and modules of computing system 1300 may be combined into fewer components and modules or further separated into additional components and modules.
  • the system 1300 comprises an image processing and analysis module 1306 that carries out the functions, methods, and/or processes described herein.
  • the image processing and analysis module 1306 may be executed on the computing system 1300 by a central processing unit 1304 discussed further below.
  • the processes, systems, and methods illustrated above may be embodied in part or in whole in software that is running on a computing device.
  • the functionality provided for in the components and modules of the computing device may comprise one or more components and/or modules.
  • the computing device may comprise multiple central processing units (CPUs) and a mass storage device, such as may be implemented in an array of servers.
  • module refers to logic embodied in hardware or firmware, or to a collection of software instructions, possibly having entry and exit points, written in a programming language, such as, for example, Java, C or C++, or the like.
  • a software module may be compiled and linked into an executable program, installed in a dynamic link library, or may be written in an interpreted programming language such as, for example, BASIC, Perl, Lua, or Python. It will be appreciated that software modules may be callable from other modules or from themselves, and/or may be invoked in response to detected events or interrupts.
  • Software instructions may be embedded in firmware, such as an EPROM.
  • hardware modules may be comprised of connected logic units, such as gates and flip-flops, and/or may be comprised of programmable units, such as programmable gate arrays or processors.
  • the modules described herein are preferably implemented as software modules, but may be represented in hardware or firmware. Generally, the modules described herein refer to logical modules that may be combined with other modules or divided into sub-modules despite their physical organization or storage.
  • the computing system 1300 also comprises a mainframe computer suitable for controlling and/or communicating with large databases, performing high volume transaction processing, and generating reports from large databases.
  • the computing system 1300 also comprises a central processing unit ("CPU' " ) 1304, which may comprise a conventional microprocessor.
  • the computing system 1300 further comprises a memory 1305, such as random access memory (“RAM “ ) for temporary storage of information and/or a read only memory (“ROM'”) for permanent storage of information, and a mass storage device 1301, such as a hard drive, diskette, or optical media storage device.
  • RAM random access memory
  • ROM' read only memory
  • mass storage device 1301 such as a hard drive, diskette, or optical media storage device.
  • the modules of the computing system 1300 are connected to the computer using a standards based bus system
  • the standards based bus system could be Peripheral Component Interconnect (PCI), MicroChannel. SCSI, Industrial Standard Architecture (ISA) and Extended ISA (EISA) architectures, for example.
  • PCI Peripheral Component Interconnect
  • ISA Industrial Standard Architecture
  • EISA Extended ISA
  • the example computing system 1300 comprises one or more commonly available input/output (I/O) devices and interfaces 1303, such as a keyboard, mouse, touchpad, and printer.
  • the I/O devices and interfaces 1303 comprise one or more display devices, such as a monitor, that allows the visual presentation of data to a user. More particularly, a display device provides for the presentation of GUIs, application software data, and multimedia presentations, for example.
  • the I/O devices and interfaces 1303 also provide a communications interface to various external devices.
  • the computing system 1300 may also comprise one or more multimedia devices 1302, such as speakers, video cards, graphics accelerators, and microphones, for example.
  • the computing system 1300 may run on a variety of computing devices, such as, for example, a server, a Windows server, a Structure Query Language server, a Unix server, a personal computer, a mainframe computer, a laptop computer, a cell phone, a personal digital assistant, a kiosk, an audio player, and so forth.
  • the computing system 1300 is generally controlled and coordinated by operating system software, such as z/OS, Windows 95, Windows 98, Windows NT, Windows 2000, Windows XP, Windows Vista, Linux, BSD, SunOS, Solaris, or other compatible operating systems, hi Macintosh systems, the operating system may be any available operating system, such as MAC OS X.
  • the computing system 1300 may be controlled by a proprietary operating system.
  • Conventional operating systems control and schedule computer processes for execution, perform memory management, provide file system, networking, and I/O services, and provide a user interface, such as a graphical user interface ('OUF " ), among other things.
  • Networks such as a graphical user interface ('OUF " ), among other things.
  • the computing system 1300 is coupled to a network 1310, such as a modem system using POTS/PSTN (plain old telephone service/public switched telephone network), ISDN, FDDI, LAN, WAN, or the Internet, for example, via a wired, wireless, or combination of wired and wireless, communication link 1315.
  • the network 1310 communicates (for example, constantly, intermittently, periodically) with various computing devices and/or other electronic devices via wired or wireless communication links.
  • the network 1310 is communicating with one or more computing systems 1317 and/or one or more data sources 1319.
  • Access to the image processing and analysis module 1306 of the computer system 1300 by remote computing systems 1317 and/or by data sources 1319 may be through a web-enabled user access point such as the computing systems' 1317 or data source's 1319 personal computer, cellular phone, laptop, or other device capable of connecting to the network 1310.
  • a web-enabled user access point such as the computing systems' 1317 or data source's 1319 personal computer, cellular phone, laptop, or other device capable of connecting to the network 1310.
  • Such a device may have a browser module implemented as a module that uses text, graphics, audio, video, and other media to present data and to allow interaction with data via the network 1310.
  • the browser module or other output module may be implemented as a combination of an all points addressable display such as a cathode-ray tube (CRT), a liquid crystal display (LCD), a plasma display, or other types and/or combinations of displays.
  • the browser module or other output module may be implemented to communicate with input devices 1303 and may also comprise software with the appropriate interfaces which allow a user to access data through the use of stylized screen elements such as, for example, menus, windows, dialog boxes, toolbars, and controls (for example, radio buttons, check boxes, sliding scales, and so forth).
  • the browser module or other output module may communicate with a set of input and output devices to receive signals from the user.
  • the input device(s) may comprise a keyboard, roller ball, pen and stylus, mouse, trackball, voice recognition system, or pre-designated switches or buttons.
  • the output device(s) may comprise a speaker, a display screen, a printer, or a voice synthesizer.
  • a touch screen may act as a hybrid input/output device, hi another embodiment, a user may interact with the system more directly such as through a system terminal connected to the score generator without communications over the Internet, a WAN, or LAN, or similar network.
  • the system 1300 may comprise a physical or logical connection established between a remote microprocessor and a mainframe host computer for the express purpose of uploading, downloading, or viewing interactive data and databases online in real time.
  • the remote microprocessor may be operated by an entity operating the computer system 1300, including the client server systems or the main server system, and/or may be operated by one or more of the data sources 1319 and/or one or more of the computing systems, hi some embodiments, terminal emulation software may be used on the microprocessor for participating in the micro-mainframe link.
  • computing systems 1317 that are internal to an entity operating the computer system 1300 may access the image processing and analysis module 1306 internally as an application or process run by the CPU 1304.
  • a user access point comprises a personal computer, a laptop computer, a cellular phone, a GPS system, a Blackberry® device, a portable computing device, a server, a computer workstation, a local area network of individual computers, an interactive kiosk, a personal digital assistant, an interactive wireless communications device, a handheld computer, an embedded computing device, or the like.
  • the network 1310 may communicate with other data sources or other computing devices.
  • the computing system 1300 may also comprise one or more internal and/or external data sources.
  • one or more of the data repositories and the data sources may be implemented using a relational database, such as DB2, Sybase, Oracle, CodeBase and Microsoft® SQL Server as well as other types of databases such as, for example, a flat file database, an entity- relationship database, and object-oriented database, and/or a record-based database.
  • the scan control and analysis module 824 is configured to determine the thickness of the retina based on the A-scan data derived from the main body 106. This data may include but is not limited to A- scan data from different A-scans.
  • the scan control and analysis module 824 can also be configured to access data and algorithms in the disease risk assessment / diagnosis database 808 to calculate the risk assessment of retinal disease based on the measured thickness of the retina as illustrated by the function curve in Figure 14A.
  • the reporting / output module 806 can be configured to normalize the calculated risk assessment value into an eye health letter or numerical grade or score.
  • the reporting / output module 806 can also be configured to access data and algorithms in the physician referral database 804 to calculate a recommended time to see a physician based on the calculated risk assessment value.
  • the system is configured to generate scan data for portions of the eye scanned to determine thickness of the retina at any one point, and compare such data to histograms and/or nomograms (for example, nomograms that show expected thickness at said location likelihood of or disease for a given thickness) to derive a risk assessment.
  • the system can also be configured to generate an average thickness for the entire retina that is scanned, and compare such data to histograms and/or nomograms to derive a risk assessment.
  • the term "histogram” as used herein generally refers to an algorithm, curve, or data or other representation of a frequency distribution for a particular variable, for example, retinal thickness.
  • the variable is divided into ranges, interval classes, and/or points on a graph (along the X-axis) for which the frequency of occurrence is represented by a rectangular column or location of points; the height of the column and/or point along the Y-axis is proportional to or otherwise indicative of the frequency of observations within the range or interval.
  • "Histograms,” as referred to herein can comprise measured data obtained, for example, from scanning the eyes of a user, or can comprise data obtained from a population of people.
  • Histograms of the former case can be analyzed to determine the mean, minimum, or maximum values, and analyze changes in slope or detect shapes or curvatures of the histogram curve. Histograms of the latter case can be used to determine the frequency of observation of a measured value in a surveyed sample.
  • an average thickness value is derived from the scan data
  • the foregoing can also be applied to minimum retinal thickness or any other value, measurement, and/or detectable condition in the eye. For example, with minimum retinal thickness, if the user has a macular hole, there will only be a small area of zero thickness, and possibly not enough to significantly reduce the average thickness, but definitely an abnormality that may be detected.
  • the system may be configured to create histograms of measured thickness and/or measured intensity values and/or slopes or derivatives of intensity values and/or variables to identify abnormalities. For example, changes or substantial changes in slope (calculated as the derivative of adjacent intensity values) may indicate hyporeflective or hyperreflective structures that may not affect mean or average intensity values, but may be indicative of disease or conditions.
  • the system can determine if the distribution of retinal thicknesses across the measured portion of the retina matches that of the normal population. Deviation from such a "normal " histogram would result in lower health grades/ higher risk assessments.
  • the methods or processes described herein can be used to determine or generate a risk assessment of maculopathy based, for example, on abnormal thickening of the retina or fovea, the presence of hyperreflective (bright or high intensity) or hyporeflective (dark or low intensity) structures in the outer half of the retina, the presence of hyporeflective (dark) structures in the inner half of the retina, the presence of irregularities in the contour of the retinal pigment epithelium that depart from the normal curvature of the eye, or of the presence of hypertransmiss ⁇ on of light through the retinal pigment epithelium when compared to a database of normal values stored in the disease risk assessment / diagnosis database 708.
  • scan data is compared to data found in normal people to identify similarities or differences from a nomogram and/or histogram.
  • scan data is compared to data found in people with diseases to identify similarities or differences from nomograms and/or histograms.
  • the pathognomonic disease features could be indicated by similarity to nomograms, for example, images, histograms, or other data, etc. from diseased patients.
  • "normal'" data are created for retinal thickness in each region of the retina (optic nerve, fovea, temporal retina) and compare to measured, detected, scanned, or encountered values to these "normal " data (for example, histograms) to determine relative risks of retinal disease or other diseases.
  • the same can be performed for nerve fiber layer (NFL) thickness to detect glaucoma.
  • the detection or generation of a risk assessment for glaucoma is performed or generated by analyzing collinear A-scan data to see if curvilinear thinning indicates the presence of glaucoma because glaucoma tends to thin the NFL in curvilinear bundles.
  • the NFL radiates out from the optic nerve in a curvilinear fashion like iron filings around a magnet. Measuring and analyzing a sequence of A-scan data that follow such a curvilinear path may be useful to identify such thinning that is characteristic of glaucoma. The analysis could be centered on and/or around the optic nerve or centered on and/or around the fovea or elsewhere. In another embodiment, the detection and/or generation of a risk assessment for glaucoma is performed or generated by analyzing the inner surface of the optic nerve to determine the optic disc cup volume.
  • the system can also be configured to detect and/or generate a risk assessment for optical clarity wherein the system integrates A-scan data in the Z direction and compares some or all the A-scan data to a nomogram value or values, or, for example, a histogram.
  • a nomogram value or values or, for example, a histogram.
  • darker A-scans will probably indicate the presence of media opacities, for example, cataracts, that decrease optical clarity ⁇ therefore, increase the subject's risk of having an optical clarity problem, for example, cataracts).
  • the system can also be configured to detect or generate risk assessments for retinal pigment epithelium (RPE) features that depart from the norma] curvature of the eye (drusen, retinal pigment epithelial detachments).
  • RPE retinal pigment epithelium
  • Such RPE features can be detected by fitting the detected RPE layer to a polynomial curve that mimics the expected curvature for the eye, and using a computer algorithm to analyze, compare, or examine the difference between these curves.
  • the system can be configured to subtract the polynomial curve that mimics the expected curvature of the RPE layer 1502 from the detected RPE layer curve 1504, and analyze and/or compare the resulting difference/value 1506 with the values (for example, in a histogram or nomogram) from normal and/or diseased eyes to generate a diagnosis or risk assessment.
  • the foregoing method and process is similar to a measure of tortuosity in that a bumpy RPE detection will generally have more deviations from a polynomial curve than smooth RPE detections, which are common in young, healthy people.
  • RPE detection can also be used to detect increased transmission through the RPE which is essentially synonymous with RPE degeneration or atrophy.
  • the system is configured to analyze the tissue layer beyond or beneath the RPE layer. Using imaging segmentation techniques, the RPE layer can be segmented.
  • the system is configured to add up all of the intensity values beneath the RPE detection. When atrophy is present, there are generally many high values beneath the RPE line, which makes the integral value high and would increase the patient's risk of having a serious macular condition, such as geographic atrophy.
  • the system can also be used to detect or generate risk factors for abnormal intensities within the retina.
  • the system is configured to divide the retina into an inner 1602 and outer 1604 half based on the midpoint between the internal limiting membrane (ILM) detection 1606 and the RPE detection lines 1608.
  • a blur filter for example, a Gaussian blur, radial blur, or the like
  • a first derivative of the intensity values can be calculated to determine the slope of the curve to differentiate the areas where there are large changes from dark to bright or vice versa across lateral dimensions of the tissue.
  • intensities or derivatives within the retina can be compared to, for example, normal histograms, wherein inner retinal hypointensity can be an indicator of cystoid macular edema; or wherein outer retinal hypointensity can be an indicator of cystoid macular edema, subretinal fluid, or diffuse macular edema; or wherein outer retinal hyperintensity can be an indication of diabetes (which may be the cause of diabetic retinopathy, or damage to the retina due to, for example, complications of diabetes mellitus), or age-related macular degeneration.
  • inner retinal hypointensity can be an indicator of cystoid macular edema
  • outer retinal hypointensity can be an indicator of cystoid macular edema, subretinal fluid, or diffuse macular edema
  • outer retinal hyperintensity can be an indication of diabetes (which may be the cause of diabetic retinopathy, or damage to the retina due to, for
  • Data from normal patients can used to compile histograms of intensity and/or slope (derivative) data to indicate expected values for normal people.
  • Data from people with various diseases can also be placed into histograms of intensity and/or derivative (slope) values to indicate expected values for those people with diseases, hi certain embodiments, a relative risk will then be developed for each entry on the histogram such that this risk can be applied to unknown cases. For example, in some instances, people with 10% of their outer retinal intensity values equal to 0 have an 85% chance of having a retinal problem. Accordingly, such users may receive a health grade of 15. In another example, people with any inner retinal points less than 10 have a 100% chance of disease, and therefore such users may receive a health grade of 5.
  • the foregoing method or process can also be used to determine or generate a risk assessment of glaucoma based on patterns of thinning of the macular and/or peripapillary nerve fiber layer or enlarged cupping of the optic nerve head as compared to a database of normal and abnormal values stored in the disease risk assessment / diagnosis database 708.
  • a histogram of expected intensity values above the inner retinal surface (in the vitreous) can be used to detect or develop a risk assessment for uveitis.
  • the presence of large, bright specks (for example, high intensity areas) in the vitreous cavity would indicate possible uveitis and would likely indicate a need for referral.
  • the foregoing method and process can also be used to determine or generate a risk of eye disease based on the intensity levels of the image signal as compared to a database of normal and abnormal values stored in the disease risk assessment / diagnosis database 708.
  • the foregoing method and process can also be used to determine or generate a risk assessment of uveitis based on hyperreflective features in the vitreous cavity as compared to normal and abnormal hyperreflective features stored in the disease risk assessment / diagnosis database 708.
  • the foregoing method and process can also be used to determine or generate a risk assessment of anterior eye disease based on detection of pathognomonic disease features, such as cystoid retinal degeneration, outer retinal edema, subretinal fluid, subretinal tissue, macular holes, drusen, retinal pigment epithelial detachments, and/or retinal pigment epithelial atrophy, wherein the detected features are compared with such pathognomonic disease features stored in the disease risk assessment / diagnosis database 708.
  • pathognomonic disease features such as cystoid retinal degeneration, outer retinal edema, subretinal fluid, subretinal tissue, macular holes, drusen, retinal pigment epithelial detachments, and/or retinal pigment epithelial atrophy
  • the system is configured to perform template matching wherein the system detects, compares, and/or matches characteristics from A-scans generated from scanning a user, also known as unknown A-scans, with a database of patterns known to be associated with disease features, such as subretinal fluid, or the like.
  • the optical coherence tomography system 100 is configured to allow the user to self-administer an OCT scan of the user ' s eyes without dilation of the eyes, and obtain a risk assessment or diagnosis of various diseases and ailments without the engaging or involving a doctor and/or technician to align the user's eyes with the system, administer the OCT scan and/or interpret the data from the scan to generate or determine a risk assessment or diagnosis.
  • the optical coherence tomography system 100 can perform a screening in less than two minutes, between 2-3 minutes, or 2-5 minutes.
  • the use of the binocular system allows the user to self-align the optical coherence tomography system 100.
  • the optical coherence system 100 with a binocular system is faster since it scans both eyes without repositioning and can allow the optical coherence tomography system 100 to scan a person's bad eye because the person's bad eye will follow the person's good eye as the latter tracks the fixation marker. Accordingly, the optical coherence tomography system 100 reduces the expense of conducting an OCT scan, thereby making OCT scanning more accessible to more people and/or users, and saving millions of people from losing their eye sight due to eye diseases or ailments that are preventable through earlier detection.
  • the optical coherence tomography system 100 is configured to have a small-foot print and/or to be portable, such that the optical coherence tomography system 100 can be installed or placed in drug stores, retail malls or stores, medical imaging facilities, grocery stores, libraries, and/or mobile vehicles, buses, or vans, a general practitioner ' s or other doctor's office, such that the optical coherence tomography system 100 can be used by people who do not have access to a doctor.
  • Additional features may be added to the optical coherence tomography system 100. hi some instances, the additional features may enhance performance of the system 100.
  • Figures 17A-C show B-scans obtained when the OCT system is positioned too far anterior, at a position that provides increased field of view, or too far posterior with respect to the eye. As shown, when the OCT system is too far anterior or too far posterior with respect to the eye the field of view (here the size or width of the B-scan) is reduced.
  • Figures 18A-C further show how a field of view of the system 100 can be affected by the location of the OCT system with respect to the eye.
  • Figures 18A-C each show two probe beams 2005a and 2005b emitted from an optical coherence tomography system along different trajectories, as shown, by, for example, rotating a galvanometer 280 to probe different portions of the retina.
  • rotation of the galvanometer 280 may cause light to be emitted along different trajectories as described above.
  • the trajectories may intersect with each other at rotation point 2010. Movement, e.g., rotation, of the galvanometer 280, may cause the trajectory of the probe beam 2005 to rotate about the rotation point 2010.
  • a plurality of beams 2005 will be emitted by the system, such that the eye tissue can be sufficiently imaged.
  • numerous other beams are emitted between beams 2005a and 2005b.
  • the beams are shown to intersect with each other at a rotation point or common point 2010. In some embodiments, the location of this point may coincide with a focus of the beams.
  • Each of the beams 2005a and 2005b and the beams therebetween can cause structures of the eye to reflect light, such that A-scan data can be beams associated with each beam.
  • Figures 18A-C show a region 2015 that can be imaged by the plurality of beams. Thus, the emitted light may sweep across a swath of points of the retina.
  • the position of the rotation point 2010 may influence the lateral dimension (e.g., length or width) of this region 2015.
  • the region 2015 may be described as a field of view and may be correlated with the amount of data within a B-scan or set of A-scans that is above a threshold intensity.
  • the rotation point 2010 is located behind/ posterior to the pupil 2030.
  • Light beams 2020a and 2020b incident at high incident angles will therefore be unable to enter the eye, as the will be blocked by the iris 2025.
  • the angle of incidence and therefore the region 2015 of the eye that can be imaged are limited in this situation.
  • the rotation point 2010 is located at a pupil plane at the pupil 2030 (e.g., in the plane of the pupil). Because the light beams intersect at the rotation point 2010, no incident light will be blocked by the iris 2025. Therefore, the region 2015 of the eye that can be imaged is not limited to obstruction by the iris as shown. A larger field of view is thereby provided.
  • the rotation point 2010 is located in front of/ anterior to the pupil 2030.
  • light beams 2020a and 2020b incident at high incident angles will therefore be unable to enter the eye, as the will be blocked by the iris 2025.
  • the angle of incidence and therefore the region 2015 of the eye that can be imaged are limited in this situation. Accordingly, the regions 2015 probed in Figures 18A and 18C are shown reduced in comparison to the region 2015 probed in Figure 18B.
  • FIG. 17A-C examples of how B-scans can be affected by the position of one or more movable components are shown.
  • Figure 17A When the rotation point 2010 is too far anterior (Figure 17A) or too far posterior (Figure 17C), less tissue is imaged than if the rotation point 2010 is positioned at a more optimal position ( Figure 17B). In each case, light from the center of the eye is reflected back towards the OCT system. However, when the rotation point 2010 is at a non-optimal location, light from the more extreme positions of the eye is not reflected back towards the OCT system. It is theorized that this light is instead scattered by the iris before it ever enters the eye, as illustrated in Figures 18A-C.
  • a risk assessment or diagnosis may then (e.g., automatically) be performed by the OCT system using an improved field of view, the improved field of view being obtained when the movable components are in a first position, and the improved field of view being larger than a field of view obtained from when the movable components are in a different second position.
  • OCT system 100 is configured to adjust an anterior-posterior distance of the OCT system with respect to the eye or a working distance of the OCT system.
  • Figure 19 shows at least one movable optical component (e.g., lens 205) of the optical coherence tomography system 100.
  • the position of the at least one movable optical component can at least partly determine the position of the rotation point 2010 and the working distance of the OCT system.
  • the working distance may at least partly determine the position of the rotation point 2010.
  • the working distance 3005 may be measured, for example, as the distance between the outermost lens or window of the eyepiece and the rotation point 2010 or a position of the eyecup 120 and the rotation point 2010. (Other reference locations on the OCT system 100 can be used.)
  • increasing the working distance can move the rotation point 2010 further anterior.
  • the OCT system may be moved with respect to the eye.
  • a field of view (a size of a set of A-scans, a B-scan or a region of the eye that can be imaged) is monitored as the galvanometer 280 is rotated.
  • a moveable or adjustable optical component such as one or more of the lens 205, adjustable optics 210, eyecup 120, and the eyepiece 203, may be moved to adjust a working distance, the location of the eyepiece 203 and/or OCT system 100 (in whole or part) with respect to the eye, which may at least partly control the field of view.
  • the adjustment may change a rotation point to, for example, position the rotation point in or near the plane of the pupil.
  • the adjustment may allow more light (e.g., a wider range of probe beam trajectories) to enter the eye from the system than would otherwise occur, thereby increasing a field of view.
  • the adjustment may increase the number of probe beam orientations that can enter the eye across a B-scan by reducing the light blocked by one or more structures of the eye (e.g., the iris).
  • Translation stages and other actuators or movement devices may be employed to position the eye or the optics of the OCT system in the anterior-posterior direction or otherwise adjusted to provide movement in a longitudinal direction along the optical axis of the OCT instrument.
  • the position (e.g., longitudinally along the optical axis of the OCT instrument) of the movable component may determine the anterior-posterior position of the rotation point.
  • a translation stage such as a stage configured to move laterally (e.g. horizontally) may be included.
  • a translation stage or actuator may determine the horizontal position of the rotation point.
  • the iris may block a portion of the light from entering the eye that would be used to form a medial or lateral portion of, for example, a B-scan.
  • the iris may block a medial portion of the scan if the stage is too far medial, while in others, it may block a lateral portion.
  • the field of view e.g., B-scans
  • the translation stage for the eye with the smaller field of view can be adjusted to increase the field of view of that eye.
  • a B-scan or other OCT measurement may be analyzed or a plurality of scans or measurements are compared to determine whether a lateral (e.g., horizontal) movement of the stage or actuator is advantageous.
  • the stage or actuator may be adjusted for example after a user-conducted interpupillary distance alignment process using for example a fixation target, such as that described above. Additional alignment may be performed subsequent to adjustment of the stage or actuator which may affect interpupillary distance.
  • movement of the stage or actuator may affect the position of the rotation point or the portion of the sample being imaged and additional modifications of the positions of the system components may be made to account for this effect.
  • movement of the stage may move one or more of the components of the OCT system.
  • the stage may support lens 205, adjustable optics 210, beamsplitter 230 and/or mirror 260.
  • One such stage may be included for each of the eyes.
  • the components supported by the stage are those such that stage movement does not affect the angle at which the beam is output from the device.
  • one or more movements or adjustments e.g., of a horizontal stage
  • the one or more of the movable or adjustable optical components may be positioned (e.g., systematically) in a plurality of positions, and data (e.g., optical coherence data) may be obtained at these positions.
  • the one or more movable/adjustable optical components may then be adjusted to be positioned at a desired position, the desired position being based on a comparison the optical coherence data obtained at each of the positions.
  • the image data, e.g., B-scan, may be obtained at the desired position.
  • one or more sets of A-scans or one or more B-scans are analyzed to determine a position of one or more movable/adjustable components.
  • Each of the B-scans or the sets of A-scans can be associated with a distinct position/setting of the one or more movable/adjustable components.
  • a property of the scans e.g.. an image quality measure or signal intensity value
  • the sum total of the integrated intensities across the B-scans or sets of A-scans are compared.
  • the intensities e.g. sum total of integrated intensities
  • a variable may be defined as the sum of the intensity at the approximate location of the retina across all A-scans within a B-scan or set of A-scans. This variable may then be compared across sets of A-scans or B-scans.
  • a resultant position/setting of the one or more movable/adjustable components may be defined as the position/setting with a set of A-scans or a B-scan having a value for the variable that is above a threshold or is maximum (e.g., greatest total intensity). Other values may be measured, calculated or considered and other approaches may be used to determine the desired position/setting and thereby increase the field of view.
  • a plurality e.g., a predetermined number
  • B-scans or sets of A-scans are obtained and a preferred position/setting of the one or more movable/adjustable components is determined as a position/setting associated with one of the B-scans or sets of A-scans.
  • the data is used to predict a preferred position/setting that may or may not be a position/setting associated with the collected data. For example, extrapolation or interpolation may be employed, hi some instances, the B-scan or A-scan set data is dynamically collected.
  • the one or more movable components may repeatedly be adjusted until a variable crosses a threshold. Other approaches and methods may be used.
  • rotating the probe beams at the pupil may increase or maximize the amount of tissue that may be imaged by the instrument.
  • a position of the retina is determined for a plurality of B-scans or sets of A-scans, each associated with a different position or setting of the one or more movable or adjustable components.
  • the position of the retina may then be used to predict the position of the pupil, and the movable/adjustable components may be positioned/set such that a rotation point is at or near the predicted pupil position.
  • a position of an structure of the eye such as the cornea, iris, retina, vitreous, anterior chamber, or tear film interface, or from another anatomical feature, such as the orbital rim, nasal bridge, cheekbone (maxilla), frontal bone, eyelid or skin surface, is determined and the movable/adjustable components are positioned/set based on the determined location.
  • the determined location may be used to predict a location of another structure, such as that of the pupil.
  • a desire position or setting of the one or more movable or adjustable components is not based on optical coherence tomography data obtained for a specific patient.
  • the position or setting may be selected based on normative data which may comprise, for example, a normative position or setting that is determined based on population-based measurements or data.
  • normative data may comprise, for example, a normative position or setting that is determined based on population-based measurements or data.
  • field of view measurements may be made for each of a plurality of positions or settings of the one or more movable/adjustable components.
  • a preferred position or setting is determined for each patient.
  • a normative position/setting may be, for example, a mean, median or mode of the preferred positions/setting across patients, hi a second instance, the measurements are compared to a threshold for all patients. The normative position or setting may then be determined as a position or setting for which, for example, the measurements exceeded this threshold across the most patients.
  • the movable or adjustable components are fixed at a normative position or setting. The movable/adjustable component may be fixed in the same system or in different systems at the position or setting determined based on normative data. This position/setting may be used as the selected position/setting or may be used as a starting point for measuring different fields of view for different positions/setting as described above to determine the position/setting having an increased field of view.
  • the normative position/setting is not determined based on optical coherence tomography data but is instead based on anatomical data otherwise obtained.
  • the normative position may be determined based on an average distance between a pupil and a retina, an average anterior-posterior distance between an eye socket and a pupil, an average anterior-posterior distance between the cornea and the pupil, or an average anterior-posterior distance between a chin and a pupil.
  • the normative position/setting may be separately determined for different patient groups.
  • the normative position/setting may be based on a person's age, gender or race.
  • the position/setting of the one or more movable/adjustable components may be based at least partly on sensor data.
  • a sensor may detect a position of the patient or a patient feature (e.g., an eye, a pupil, an iris, a chin, an eye socket), and this position may be used to determine the position or setting of the eyepiece 203 or OCT system 100.
  • the detected position is used to predict the position of the pupil, which is used to determine the position of the one or more movable components.
  • an optical coherence tomography system (e.g., that of Figure 1 or 3) comprises a sensor or tracker.
  • the sensor or tracker may determine a position of the user, one or two eyes of the user, and/or one or more structures (e.g., a retina, pupil, cornea, or lens) of the user's eye.
  • the sensor or tracker is positioned on or attached to main body 106, zero gravity arm 116, or even eyecup 120.
  • the sensor or tracker is a device separate from the main body 106.
  • the sensor or tracker is attached to or comprised within the system shown in Figure 3.
  • the senor emits light or ultrasound from a light source and detects light reflected back.
  • the light may be reflected back from a structure of the user's eye, such as the cornea, iris, pupil, retina, vitreous, anterior chamber, or tear film interface, or from another anatomical feature, such as the orbital rim, nasal bridge, cheekbone (maxilla), frontal bone, eyelid or skin surface.
  • the sensor may determine the position of the structure based on the time difference between the time the light (e.g., a pulse) was emitted and the time the light was detected.
  • other types of sensors or trackers may be used.
  • an optical coherence tomography instrument may determine the position of an eye structure based on interference or reflectance results.
  • the position/setting of the one or more movable/adjustable components is based on a combination of approaches.
  • the position/setting may be determined based on non-optical coherence tomography sensor data and optical coherence tomography data.
  • the position/setting may be determined based on field-of-view data and sensor data and/or a determined normative position.
  • the position may be determined based on sensor data and a determined nonnative position.
  • at least one of normative data or sensor data may be used to assist in determining a starting point for multiple OCT measurements that are subsequently employed to determine a position or setting which provides a further increased field of view.
  • an optical coherence tomography system 100 comprises a chin rest.
  • the system 100 may be configured to automatically adjust or to allow for manual adjustment between the main body (and/or the eyepiece) and the patient ' s eyes.
  • the adjustment may be fine, on the order of about 0.5, 1, 2, 3, 4, 5, 10, 20, 30 or 50 millimeters.
  • the adjustment may comprise any adjustment described herein, such as an adjustment of one or more moveable optical components to, for example, improve a field of view of the system 100.
  • the distance between the main body and/or an optical component of in the main body and the patient's eye is systematically adjusted from a first distance to a second distance.
  • the chin rest may move in certain embodiments although in other embodiments the chin rest may be fixed.
  • the distance may be based at least partly on nonnative values, such as an average offset (e.g., in the anterior-posterior direction) between a chin and a pupil or an average distance between a pupil and an eyecup.
  • the distance is detennined based at least partly on a sensor reading.
  • a sensor may detect a position of the user's eye, pupil or iris.
  • the sensor may comprise an optical tomography instrument or may comprise another optical or ultrasonic instrument.
  • the sensor may emit a light and determine the time elapsed between the emission and that at which reflected light (e.g., a pulse) is received.
  • the sensor may comprise a weight sensor to sense, for example, a location of the patient's chin.
  • a sensor may detect a position or weight of the user ' s chin.
  • the chin rest may move or the main body and/or eyepiece of the OCT system may move with respect to the chin rest and the field of view monitored as described above to determine a suitable location of the eye. Other variations are possible.
  • a position/setting of one or more moveable/adjustable optical components can be manually adjusted by the patient.
  • the patient may be instructed, for example, to adjust the position/setting based on one or more images seen by the patient.
  • the patient may be instructed to adjust the position until two or more images (e.g., working distance images) are aligned. Alignment may correspond to an appropriate distance of the eye to the OCT instrument. Other designs are also possible.
  • All of the methods and processes described above may be embodied in, and fully automated via, software code modules executed by one or more general purpose computers or processors.
  • the code modules may be stored in any type of computer-readable medium or other computer storage device. Some or all of the methods may alternatively be embodied in specialized computer hardware.

Landscapes

  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Radiology & Medical Imaging (AREA)
  • Biophysics (AREA)
  • Ophthalmology & Optometry (AREA)
  • Engineering & Computer Science (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Medical Informatics (AREA)
  • Physics & Mathematics (AREA)
  • Surgery (AREA)
  • Animal Behavior & Ethology (AREA)
  • General Health & Medical Sciences (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Eye Examination Apparatus (AREA)
  • Investigating Or Analysing Materials By Optical Means (AREA)

Abstract

In accordance with one aspect of the present invention, an optical coherence tomography instrument comprises an eyepiece for receiving at least one eye of a user is provided: a light source that outputs light that is directed through the eyepiece into the user's eye; an interferometer configured to produce optical interference using light reflected from the user's eye; an optical detector disposed so as to detect said optical interference; and electronics coupled to the detector. The electronics can be configured to automatically evaluate a plurality of fields of view of the optical coherence tomography instrument. The evaluation can comprise rotating at least one galvanometer to direct a probe beam along different trajectories into an eye.

Description

DOHENY.008VPC PATENT
OPTICAL COHERENCE TOMOGRAPHY DEVICE, METHOD, AND SYSTEM
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This application claims the benefit of U.S. Provisional Application No. 61/082,167, titled, "OPTICAL COHERENCE TOMOGRAPHY DEVICE, METHOD, AND SYSTEM/" and filed July 18, 2008 (DOHEN Y.008PR), and claims the benefit of U.S. Provisional Application No. 61/040,086, titled, "OPTICAL COHERENCE TOMOGRAPHY DEVICE, METHOD, AND SYSTEM," and filed March 27, 2008 (DOHENY.002PR), and claims the benefit of U.S. Provisional Application No. 61/040,084, titled, "OPTICAL COHERENCE TOMOGRAPHY DEVICE, METHOD, AND SYSTEM," and filed March 27, 2008 (DOHENY.001PR), and claims priority to U.S. Application No. 12/11 1,894, titled "OPTICAL COHERENCE TOMOGRAPHY DEVICE, METHOD, AND SYSTEM," and filed April 29, 2008 (DOHENY.001PR), now pending. The foregoing applications are hereby incorporated by reference in their entirety, including without limitation, for example, the optical coherence tomography devices, methods, and systems disclosed therein.
BACKGROUND Field
|0002] Embodiments of the invention relate to the field of optical coherence tomography and, in particular, to devices, systems, methods of utilizing such optical coherence tomography data to perform precision measurements on eye tissue for the detection of eye diseases.
Description of the Related Art
[0003] Many industrial, medical, and other applications exist for optical coherence tomography (OCT), which generally refers to an interferometric, non-invasive optical tomographic imaging technique offering millimeter penetration (approximately 2-3 mm in tissue) with micrometer-scale axial and lateral resolution. For example, in medical applications, doctors generally desire a non-invasive, in vivo imaging technique for obtaining sub-surface, cross-sectional and/or three-dimensional images of translucent and/or opaque materials at a resolution equivalent to low-power microscopes. Accordingly, in the coming years, it is projected that there will be 20 million OCT scans performed per year on patients. Most of these will probably occur in the field of ophthalmology. In current optica] coherence tomography systems, doctors or other medical professionals administer the OCT scans in the doctors' medical office or medical facilities.
SUMMARY
|0004] Various embodiments of the present invention relate to the utilization of optical coherence tomography, which generally refers to an interferometric, non-invasive optical tomographic imaging technique, that can be used to detect and analyze, for example, eye tissue, and/or disease features, including but not limited to cystoid retinal degeneration, outer retinal edema, subretinal fluid, subretinal tissue, macular holes, drusen, or the like. For example, and in accordance with one aspect of the present invention, an optical coherence tomography instrument comprises an eyepiece for receiving at least one eye of a user; a light source that outputs light that is directed through the eyepiece into the user's eye; an interferometer configured to produce optical interference using light reflected from the user's eye; an optical detector disposed so as to detect said optical interference; electronics coupled to the detector and configured to perform a risk assessment analysis based on optical coherence tomography measurements obtained using said interferometer; and an output device electrically coupled to the electronics, and the output device may be configured to output the risk assessment to the user through the output device. Generally, the optical coherence tomography instruments, devices, systems, and methods disclosed herein can be self-administered, and the eyepiece can be a monocular system or a binocular system.
[0005] In an embodiment, a method for providing a self-administered optical coherence tomography test of a user's eye comprises automatically evaluating a plurality of fields of view of the optical coherence tomography instrument, each field of view corresponding to a different state of at least one optical component of said optical coherence tomography instrument; automatically adjusting said at least one optical component at least partly based on said evaluation; and with said at least one optical component adjusted, performing the optical coherence tomography test, wherein the optical coherence tomography instrument is configured to scan the at least one user eye. [0006] In an embodiment, the method step of automatically evaluating a plurality of fields of view of the optical coherence tomography instrument comprises rotating a galvanometer to direct a probe beam along different trajectories into the user eye. In the method, the trajectories can intersect at a common location. In an embodiment, the common location comprises at least one of a location in or near the pupil of the user eye, a location in a plane of the iris of the user eye, a location within the lens of the user eye, and a location posterior to the pupil of the user eye.
[0007] In an embodiment, the adjusting said at least one optical component comprises positioning the optical coherence tomography instrument along an anterior- posterior axis of the user. The method of Claim 1, wherein said adjusting said at least one optical component comprises positioning said at least one optical component along an optical axis of said optical coherence tomography instrument. In an embodiment, the optical coherence tomography instrument has a working distance and said adjusting said at least one optical component alters the working distance of said optical coherence tomography instrument. In the method, the at least one optical component can comprise at least one lens.
[0008] In an embodiment, the evaluating the fields of view comprises collecting A-scan data. The evaluating the fields of view can comprise collecting B-scan data. The fields of view can comprise angular fields of view. The method can further comprise providing the user with an output of the optical coherence tomography test from an output device. The method can further comprise receiving the user eye via an eyepiece of the optical coherence tomography instrument; directing light from a source through the eyepiece into the user eye; producing optical interference using light reflected from the user eye; and detecting optical interference. The method can further comprise automatically performing a diagnosis or risk assessment based on said detected optical interference, wherein the output can comprise the diagnosis or risk assessment. The method can further comprise identifying one or more diseases or abnormal findings that the user may manifest indications of based on said detected optical interference, wherein the output can comprise a notification said identified diseases or abnormal findings.
[0009] In an embodiment, a method for providing a self-administered optical coherence tomography test of a user's eye comprises determining a longitudinal position of the user relative to an optical coherence tomography instrument; automatically adjusting at least one optical component of said optical coherence tomography instrument at least partly based on said determined position thereby increasing the field of view of the optical coherence tomography instrument; and with said at least one optical component adjusted, performing the optical coherence tomography test, wherein the optical coherence tomography instrument is configured to scan the at least one user eye.
[0010] In an embodiment, the determining a position of the user comprises detecting a portion of the user based on a sensor measurement. The determining a position of the user can comprise determining the position of the user's eye. The determining a position of the user can comprise radiating a signal to the user and detecting the signal returned from the user. In an embodiment, the signal can comprise light or ultrasound. The signal can be reflected off of at least one of a cornea, an iris, a retina, a vitreous, an anterior chamber, and a tear film interface of the user. The signal can be reflected off of at least one of an orbital rim, a nasal bridge, a cheekbone (maxilla), a frontal bone, an eyelid and a skin surface. The determining a position of the user can comprise determining the position of the user's eye cornea or tear film interface. The determining a position of the user can comprise determining the position of the user's eye pupil or iris. The determining a position of the user can comprise determining the position of the user's retina. The determining a position of the user can comprise analyzing optical coherence tomography data obtained by the optical coherence tomography instrument.
[001 IJ In an embodiment, the method can further comprise automatically evaluating a plurality of fields of view of the optical coherence tomography instrument, each field of view corresponding to a different state of at least one optical component of said optical coherence tomography instrument, said automatically adjusting said at least one optical component being at least partly based on said evaluation. The optical coherence tomography instrument can comprise rotating a galvanometer that directs a probe beam along different trajectories into the user eye, said trajectories intersecting at a common location, hi an embodiment, the automatically adjusting the at least one optical component positions the common location in a pupil of the user eye. [0012] In an embodiment, the adjusting the at least one optical component comprises positioning the optical coherence tomography instrument along an anterior- posterior axis of the user. The adjusting the at least one optical component can comprise positioning the at least one optical component along an optical axis of said optical coherence tomography instrument. The optical coherence tomography instrument can have a working distance and the adjusting of the at least one optical component can alter the working distance of the optical coherence tomography instrument. The at least one optical component can comprise at least one lens. The method can further comprise providing the user with an output of the optica) coherence tomography test from an output device.
[0013] In an embodiment, an optical coherence tomography instrument for providing a self-administered optical coherence tomography test comprises an eyepiece for receiving at least one eye of a user; a light source that outputs a probe beam that is directed through the eyepiece into the user's eye; an interferometer configured to produce optical interference using light reflected from the user's eye; an optical detector disposed so as to detect the optical interference; a processor configured to analyze optical coherence tomography measurements obtained using the interferometer; and at least one adjustable optical component that may be adjusted to change the field of view of the optical coherence tomography instrument; wherein the processor is configured to monitor a field of view of the optical coherence tomography instrument and direct adjustment of the at least one adjustable optical component based on the monitored field of view.
|0014] The optical coherence tomography instrument can further comprise a rotational unit configured to rotate the trajectory of the probe beam into the user's eye. The the rotational unit can comprise a galvanometer. The rotational unit can cause the probe beam to propagate along a plurality of different trajectories that intersect at a common location and said at least one adjustable optical component is adjustable so as to place said common location in or near the pupil of the user eye. The position of the optical coherence tomography instrument can be adjustable along an anterior-posterior axis of the user so as to adjust the position of the at least one adjustable optical component so as to adjust the field of view. The at least one adjustable optical component can have a position that is adjustable along an optical axis of said optical coherence tomography instrument so as to adjust the field of view.
[0015] In an embodiment, the optical coherence tomography instrument has a working distance and adjustment of said at least one optical adjustable component alters the work distance of said optical coherence tomography instrument. The at least one optical component can comprise at least one lens. The processor can be configured to compare a plurality of sets of A-scans, each set being associated with a different state of the adjustable optical component. The processor can be configured to compare the field of view of different A-scan sets. The processor can be configured to deteπnine a setting of the adjustable optical component selected based on monitored field of view data associated with each of the number of settings of the adjustable optical component. The processor can be configured to determine a position of the adjustable optical component that increases the portion of the user's eye that can be imaged by the optical coherence tomography instrument.
[0016] The optical coherence tomography can further comprise an output device configured to output a result of the analysis to the user. The optical tomography instrument can further comprise a sensor for determining a position of a structure of the user's eye, said processor being configured to control the position of the adjustable optical component based at least in part on the determined position of structure of the eye. The structure can comprise the pupil, the iris, the cornea, the anterior chamber, the lens, the vitreous or the retina.
[0017] In an embodiment, an optical coherence tomography instrument for providing a self-administered optical coherence tomography test comprises an eyepiece for receiving at least one eye of a user; a sensor for determining a position of a structure of the user eye with respect to the optical coherence tomography instrument; a light source that outputs light that is directed through the eyepiece into the user's eye; an interferometer configured to produce optical interference using light reflected from the user's eye; an optical detector disposed so as to detect said optical interference; and at least one adjustable optical component that may be adjusted to change the field of view of the optical coherence tomography instrument; and a processor configured to analyze optical coherence tomography measurements obtained using said interferometer, wherein the processor is further configured to direct adjustment of the at least one adjustable optical component based at least in part on said position.
[0018] The sensor can comprise an optical sensor or an ultrasound sensor. The optical coherence tomography instrument can further comprise an ultrasound emitter. The optical coherence tomography instrument can further comprise a light emitter that directs an optical signal to said eye to be detected by said sensor. The structure can comprise a pupil, iris or lens, or the like. The structure can comprise a cornea or anterior chamber. The structure can comprise a retina or the vitreous. The at least one adjustable optical component can comprise at least one lens. The at least one adjustable optical component can comprise an eyepiece. The optical coherence tomography can further comprise an output device configured to output a result of the analysis to the user.
[00191 In an embodiment, an optical coherence tomography instrument for providing a self-administered optical coherence tomography test comprises an eyepiece for receiving at least one eye of a user; a chin rest for receiving a chin of the user, the chin rest movable with respect to the eyepiece to alter the longitudinal distance therebetween; a light source that outputs light that is directed through the eyepiece into the user's eye; an interferometer configured to produce optical interference using light reflected from the user's eye; an optical detector disposed so as to detect the optical interference; a processor configured to analyze optical coherence tomography measurements obtained using the interferometer, wherein the processor is further configured to control the position of the chin rest relative to the eyepiece so as to increase the field of view of the optical coherence tomography instrument.
[0020] The processor can be configured to position the chin rest with respect to the eyepiece based at least partly on a noπnative position based on across a patient population. The processor can be configured to monitor a field of view of the optical coherence tomography instrument and position the chin rest with respect to the eyepiece based at least partly on said monitored field of view. The processor can be configured to compare a plurality of sets of A-scans, each data set being associated with a different position of the chin rest relative to the eyepiece. The processor can be configured to position the chin rest relative to the eyepiece at least partly based on a comparison of a plurality of fields of view from each of the plurality of A-scan data sets.
[0021] The optical coherence tomography instrument can further comprise a sensor to sense the user or part of the user. The sensor can be configured to sense a position of the user's eye or structure of the user's eye. The controller can be configured to position the chin rest relative to the eyepiece based at least partly on data collected by the sensor. The optical coherence tomography instrument can further comprise a movable horizontal stage supporting at least one component of the optical coherence tomography instrument sensor, the stage being configured to adjust a medial-lateral position of the light being directed through the eyepiece into the user's eye.
[0022] In an embodiment, an optical coherence tomography instrument comprises an eyepiece for receiving at least one eye of a user; a light source that outputs light that is directed through the eyepiece into the user's eye: an interferometer configured to produce optical interference using light reflected from the user's eye; an optical detector disposed so as to detect said optical interference; electronics coupled to the detector and configured to perform a risk assessment analysis based on optical coherence tomography measurements obtained using said interferometer; and an output device electrically coupled to the electronics. In an embodiment, the light from the light source can be controlled using a low performance (e.g., resolution) galvanometer and a high performance (e.g., resolution) galvanometer.
[0023] For purposes of this summary, certain aspects, advantages, and novel features of the invention are described herein. It is to be understood that not necessarily all such aspects, advantages, and features may be employed and/or achieved in accordance with any particular embodiment of the invention. Thus, for example, those skilled in the art will recognize that the invention may be embodied or carried out in a manner that achieves one advantage or group of advantages as taught herein without necessarily achieving other advantages as may be taught or suggested herein.
BRIEF DESCRIPTION OF THE DRAWINGS
[0024] The foregoing and other features, aspects and advantages of the present invention are described in detail below with reference to the drawings of various embodiments, which are intended to illustrate and not to limit the invention. The drawings comprise the following figures in which:
[0025] Figure ] is a schematic diagram of one embodiment of the optical coherence tomography system described herein.
[0026] Figure 2 is a schematic diagram of one embodiment of an interferometer arranged to perform measurements of an eye.
[0027] Figure 3A is a schematic diagram of one embodiment of an OCT system comprising a main body configured to conveniently interfere with a person's eyes, the main body being in communication with various systems as described herein.
[0028] Figure 3B is a perspective view schematically illustrating an embodiment of the main body shown in Figure 3A.
[0029] Figure 4 is schematic diagram of one embodiment of a spectrometer used to analyze data from an interferometer used for OCT.
[0030] Figure 5 is a schematic diagram of the main body of an OCT system comprising a single display for presenting a display target to a patient.
[0031] Figures 6A - 6C are schematic diagrams illustrating the use of optical coherence tomography to scan retinal tissue to generate A-scans and B-scans.
[0032] Figures 7A - 7F are schematic diagrams illustrating embodiments for adjusting and/or calibrating interpupillary distance.
[0033] Figure 8 is a block diagram schematically illustrating one embodiment of the computer system of the optical coherence tomography system described herein.
[0034] Figure 9 is illustrates a process flow diagram of one embodiment of performing precision measurements on retinal tissue for the detection of pathognomonic disease features.
[0035] Figures I OA- 1 OD illustrate possible embodiments of disposing the main body of an optical coherence tomography device with respect to a user.
[0036] Figures UA-I lB illustrate possible embodiments of output reports generated by the optical coherence tomography device. [0037] Figure 12 is a block diagram schematically illustrating another embodiment of the computer system for an optical coherence tomography system described herein.
[0038] Figure 13 is a block diagram schematically illustrating components in one embodiment of the computer system for an optical coherence tomography system described herein.
[0039] Figure 14A is a diagram schematically illustrating one embodiment for determining a risk assessment.
[0040] Figure 14B is a schematic illustration of a plot of risk of retinal disease versus retinal thickness for detennining a risk assessment in another embodiment.
|0041] Figure 15 is an illustration of RPE detection and RPE polynomial fit curvature, and the difference there between.
[0042] Figure 16 is an illustration of retinal tissue segmented into inner and outer retinal tissue regions.
[0043] Figures 17A-C show B-scans obtained when the OCT system is positioned too far anterior, at a position that provides increased field of view, or too far posterior with respect to the eye.
[0044] Figures 18A-C show light beam trajectories when the OCT system is positioned too far anterior, at a position that provides increased field of view, or too far posterior wherein the intersection of the trajectoires is behind the pupil, at a pupil plane or in front of the pupil.
[0045] Figure \ 9 shows one embodiment of a suitable working distance between an optical coherence tomography system and a patient's retina.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT |0046] Embodiments of the invention will now be described with reference to the accompanying figures, wherein like numerals refer to like elements throughout. The terminology used in the description presented herein is not intended to be interpreted in any limited or restrictive manner, simply because it is being utilized in conjunction with a detailed description of certain specific embodiments of the invention. Furthermore, embodiments of the invention may comprise several novel features, no single one of which is solely responsible for its desirable attributes or which is essential to practicing the inventions herein described. The embodiments described herein make OCT screening more accessible to users thereby allowing for earlier detection and/or treatment of various diseases, ailments, or conditions, for example, maculopathy, glaucoma, or the like.
[0047] The terms "optical coherence tomography'" and "OCT'' generally refer to an interferometric technique for imaging samples, in some cases, with micrometer lateral resolution. This non-invasive optical tomographic imaging technique is used in ophthalmology to provide cross-sectional images of the eye, and more particularly the posterior of the eye, though it can also be used to image other samples or tissues in areas of the user's body.
(0048] Generally, OCT employs an interferometer. Light from a light source (for example, a broadband light source) is split (for example, by a beamsplitter) and travels along a sample arm (generally comprising the sample) and a reference arm (generally comprising a mirror). A portion of the light from the sample arm is reflected by the sample. Light is also reflected from a mirror in the reference arm. (Light from the test arm and the reference arm is recombined, for example by the beamsplitter.) When the distance travelled by light in the sample arm is within a coherence length of the distance travelled by light in the reference arm. optical interference occurs, which affects the intensity of the recombined light. The intensity of the combined reflected light varies depending on the sample properties. Thus, variations for the intensity of the reflectance measured are indications of the physical features of the sample being tested.
J0049] In time-domain OCT, the length of the reference arm can be varied (for example, by moving one or more reference mirrors). The reflectance observed as the reference arm distance changes indicates sample properties at different depths of the sample. (In some embodiments, the length of the sample arm is varied instead of or in addition to the variation of the reference arm length.) In frequency-domain OCT, the distance of the reference arm can be fixed, and the reflectance can then be measured at different frequencies. For example, the frequency of light emitted from a light source can be scanned across a range of frequencies or a dispersive element, such as a grating, and a detector array may be used to separate and detect different wavelengths. Fourier analysis can convert the frequency- dependent reflectance properties to distance-dependent reflectance properties, thereby indicating sample properties at different sample depths. In certain embodiments, OCT can show additional information or data than nonmydriatic color fundus imaging.
10050] The term "A-scan" describes the light reflectivity associated with different sample depths. The term "B-scan" as used herein refers to the use of cross-sectional views of tissues formed by assembly of a plurality of A-scans. In the case of ophthalmology, light reflected by eye tissues is converted into electrical signals and can be used to provide data regarding the structure of tissue in the eye and to display a cross-sectional view of the eye. hi the case of ophthalmology, A-scans and B-scans can be used, for example, for differentiating normal and abnormal eye tissue or for measuring thicknesses of tissue layers in the eyes.
|0051] In ophthalmic instances, an A-scan can generally include data from the cornea to the retina. In some instances, a B-scan can include cross-sectional data from a medial border to a lateral border of the eye and from the comea to the retina. In some instances, a B-scan can include cross-sectional data from a superior border to an inferior border of the eye and from the cornea to the retina. A 3D-OCT can be formed by combining a plurality of B-scans.
[0052] As used herein the terms "user" or ''patient" may be used interchangeably, and the foregoing terms comprise without limitation human beings, whether or not under the care of a physician, and other mammals.
[0053] The terms "eye scan,'" "scanning the eye," or "scan the eyes," as used herein, are broad interchangeable terms that generally refer to the measurement of any part or substantially all of the eye, including but not limited to the cornea, the retina, the eye lens, the iris, optic nerve, or any other tissue or nerve related to the eye.
[0054J The terms "risk assessment'" and "diagnosis," may be used in the specification interchangeably although the terms have different meanings. The term "risk assessment" generally refers to a probability, number, score, grade, estimate, etc. of the likelihood of the existence of one or more illnesses, diseases, ailments, or the like. The term "diagnosis" generally refers to a determination by examination and/or tests the nature and circumstances of an illness, ailment, or diseased condition. [0055] Various methods, systems, and devices may be used to generate and utilize optical coherence tomography image data to perform precision measurements on retinal tissue for the detection of disease features, and generating a risk assessment and/or diagnosis based on data obtained by optical coherence tomography imaging techniques. These methods, systems and devices may employ, in some embodiments, a statistical analysis of the detected disease features obtained by optical coherence tomography imaging techniques. Such methods, systems, and devices can be used to screen for diseases.
[0056] With reference to Figure 1, there is illustrated a block diagram depicting one embodiment of the optical coherence tomography system. In one embodiment, computer system 104 is electrically coupled to an output device 102, a communications medium 108, and a user card reader system 1 12. The communications medium 108 can enable the computer system 104 to communicate with other remote systems 1 10. The computer system 104 may be electrically coupled to main body 106, which the user 1 14 positions near or onto the user's eyes. In the illustrated example, the main body 106 is a binocular system {for example, has a two oculars or optical paths for the eyes providing one view for one eye and another view for another eye, or the like) configured to scan two eyes without repositioning the oculars with respect to the head of the patient, thereby reducing the time to scan a patient. In some embodiments, the eyes are scanned simultaneously using a scanner (for example, galvanometer), which provides interlaces of measurements from both eyes. Other embodiments are possible as well, for example, the binocular system or a two ocular system having two respective optical paths to the two eyes can be configured to scan the eyes in series, meaning one eye first, and then the second eye. In some embodiments, serial scanning of the eyes comprises scanning a first portion of the first eye, a first portion of the second eye, a second portion the first eye, and so on. Alternatively, the main body 106 can comprise a monocular system or one ocular system or optical path to the eye for performing eye scans.
[0057] Referring to Figure 1 , the user 114 can engage handle 1 18 and position (for example, up, down, or sideways) the main body 106 that is at least partially supported and connected to a zero gravity arm 1 16, and accordingly the system 100 has no chin rest. In some embodiments, this configuration can introduce positioning error due to movement of the mandible. When the main body 106 is in such a position, the distance between the outermost lens (the lens closest to the user) and the user's eye can range between 10 mm and 30 mm, or 5 mm and 25 mm, or 5 mm and 10 mm. The close proximity of the lens system to the user's eyes increases compactness of the system, reduces position variability when the patient places his eyes (for example, orbital rims) against the man body, and increases the viewing angle of the OCT apparatus when imaging through an undilated pupil.
10058] Accordingly, the main body 106 can also comprise eyecups 120 (for example, disposable eyecups) that are configured to contact the user's eye socket to substantially block out ambient light and/or to at least partially support the main body 106 on the eye socket of the user 114. The eyecups 120 have central openings (for example, apertures) to allow passage of light from the light source in the instrument to the eyes. The eyecups 120 can be constructed of paper, cardboard, plastic, silicon, metal, latex, or a combination thereof. The eyecups 120 can be tubular, conical, or cup-shaped flexible or semi-rigid structures with openings on either end. Other materials, shapes and designs are possible. In some embodiments, the eyecups 120 are constructed of latex that conforms around eyepiece portions of the main body 106. The eyecups 120 are detachable from the main body 106 after the eye scan has been completed, and new eyecups 120 can be attached for a new user to ensure hygiene and/or to protect against the spread of disease. The eyecups 120 can be clear, translucent or opaque, although opaque eyecups offer the advantage of blocking ambient light for measurement in lit environments.
[0059J The main body 106 may comprise one or more eyepieces, an interferometer, one or more target displays, a detector and/or an alignment system. The optical coherence tomography system may comprise a time domain optical coherence tomography system and/or a spectral domain optical coherence tomography system. Accordingly, in some embodiments, the main body 106 comprises a spectrometer, (for example, a grating) and a detector array. The main body may, in some embodiments, comprise signal processing component (for example, electronics) for performing, for example, Fourier transforms. Other types of optical coherence tomography systems may be employed.
|0060] Figure 2 shows a diagram of an example optical coherence tomography system. Light 150 is output from a light source 155. The light source 155 may comprise a broadband light source, such as a superluminescent diode or a white light source. (Alternatively, light emitted from the light source 155 may vary in frequency as a function of time.) The light 150 may comprise collimated light. In one embodiment, light 150 from the light source 155 is collimated with a collimating lens. The light is split at beamsplitter 160. Beamsplitters, as described herein, may comprise without limitation a polarization-based beamsplitter, a temporally based beamsplitter and/or a 50/50 beamsplitter or other devices and configurations. A portion of the light travels along a sample arm, directed towards a sample, such as an eye 165 of a user 1 14. Another portion of the light 150 travels along a reference arm, directed towards a reference mirror 170. The light reflected by the sample and the reference mirror 170 are combined at the beamsplitter 160 and sensed either by a one- dimensional photodetector or a two-dimensional detector array such as a charge-coupled device (CCD) or complementary metal-oxide-semiconductor (CMOS). A two-dimensional array may be included in a full field OCT instrument, which may gather information more quickly than a version that uses a one dimensional photodetector array instead. In time- domain OCT, the length of the reference arm (which may be determined in part by the position of the reference mirror 170) may be varying in time.
[0061] Whether interference between the light reflected by the sample and the light reflected by the reference mirror/s occurs will depend on the length of the reference arm (as compared to the length of the test arm) and the frequency of the light emitted by the light source. High contrast light interference occurs between light travelling similar optical distances (for example, (differences less than a coherence length). The coherence length is determined by the bandwidth of the light source. Broadband light sources correspond to smaller coherence lengths.
|0062] In time-domain OCT. when the relative length of the reference and sample arms varies over time, the intensity of the output light may be analyzed as a function of time. The light signal detected results from light rays scattered from the sample that interfere constructively with light reflected by the reference mirror/s. Increased interference occurs, however, when the lengths of the sample and reference arms are approximately similar (for example, within about one coherence length in some cases). The light from the reference arm, therefore, will interfere with light reflected from a narrow range of depths within the sample. As the reference (or sample) arms are translated, this narrow range of depths can be moved through the thickness of the sample while the intensity of reflected light is monitored to obtain information about the sample. Samples that scatter light will scatter light back that interferes with the reference arm and thereby produce an interference signal. Using a light source having a short coherence length can provide increased to high resolution (for example, 0.1-10 microns), as the shorter coherence length yields a smaller range of depths that is probed at a single instant in time.
[0063] In various embodiments of frequency-domain optical coherence tomography, the reference and sample arms are fixed. Light from a broadband light source comprising a plurality of wavelengths is reflected from the sample and interfered with light reflected by the reference mirror/s. The optical spectrum of the reflected signal can be obtained. For example, the light may be input to a spectrometer or a spectrograph comprising, for example, a grating and a detector array, that detects the intensity of light at different frequencies.
[0064] Fourier analysis performed, for example, by a processor may convert data corresponding to a plurality of frequencies to that corresponding to a plurality of positions within the sample. Thus, data from a plurality of sample depths can be simultaneously collected without the need for scanning of the reference arm (or sample) arms. Additional details related to frequency domain optical coherence tomography are described in Vakhtin et al., (Vakhtin AB, Kane DJ, Wood WR and Peterson KA. "Common-path interferometer for frequency-domain optical coherence tomography," Applied Optics. 42(34), 6953-6958 (2003)).
[0065] Other methods of performing optical coherence tomography are possible. For example, in some embodiment of frequency domain optical coherence tomography, the frequency of Hght emitted from a light source varies in time. Thus, differences in light intensity as a function of time relate to different light frequencies. When a spectrally time- varying light source is used, a detector may detect light intensity as a function of time to obtain optical spectrum of the interference signal. The Fourier transform of the optical spectrum may be employed as described above. A wide variety of other techniques are also possible. [0066] Figure 3A shows one configuration of main body 106 comprising an optical coherence tomography system and an alignment system. Other optical coherence tomography systems and/or alignment systems may be included in place of or in addition to the systems shown in Figure 3A. As shown, the main body 106 can include two eyepieces 203, each eyepiece configured to receive an eye from a user 114. In other embodiments, the main body 106 includes only one eyepiece 203.
10067} Figure 3A shows one representative embodiment of an optical coherence tomography system. Light from a light source 240 may propagate along a path that is modulated, for example, vertically and/or horizontally by one or more beam deflectors 280. A galvanometer may be used for this purpose. The galvanometer 280 can control the horizontal and/or vertical location of a light beam from the light source 240, thereby allowing a plurality of A-scans (and thus one or more B-scan and/or a 3D-OCT) to be formed.
[0068] The light from the light source 240 is split at beamsplitter 245. In some embodiments, beamsplitter 245 is replaced by a high frequency switch that uses, for example, a galvanometer, that directs about 100% of the light towards mirror 250a for about 1A of a cycle and then directs about 100% of the light towards mirror 250b for the remainder of the cycle. The light source 240 may include a broadband light source, such as a superluminescent light-emitting diode. Light split at the beamsplitter 245 is then split again at beamsplitter 285a or 285b to form a reference arm and a sample arm. A first portion of the light split at beamsplitter 285a or 285b is reflected by reference mirrors 273a or 273b, reference mirrors 270a or 270b. and reference mirrors 265a or 265b. A second portion of the light split at beamsplitter 285a or 285b is reflected by mirror 250a or 250b. by mirror 255a or 255b and by mirror 260a or 260b. Mirrors 255a or 255b and mirrors 250a and 250b are connected to a Z-offset adjustment stage 290b. By moving the position of the adjustment stage 290a or 290b, a different portion of the eye can be imaged. Thus, the adjustment stage 290a or 290b can adjust the difference between the optical length from the light source 240 to a portion of the sample and the optical length from the light source 240 and the reference mirror 270a or 270b and/or reference mirror 273a or 273b. This difference can be made small, for example, less than a coherence length, thereby promoting for optical interference to occur. In some embodiments, the positions of one or more reference mirrors (for example, reference mirror 270a or 270b and reference mirror 273a or 273b) are movable in addition to or instead of the adjustment stage being movable. Thus, the length of the reference arm and/or of the sample arm may be adjustable. The position of the adjustment stages 290a and/or 290b may be based on the signals from the device, as described in more detail below.
[00691 The light reflected by mirror 260a or 260b is combined with light from display 215a or 215b at beamsplitter 230a or 230b. The displays 215a and 215b may comprise one or more light sources, such as in an emissive display like an array of matrix LEDs. Other types of displays can be used. The display can display targets of varying shapes and configurations, including a bar and/or one or more dots. A portion of the optical path from the light source 240 to the eye may be coaxial with a portion of the path from the displays 215a and 215b to the eye. These portions may extend though the eyepiece. Accordingly, a light beam from the light source 240 is coaxial with a light beam from the displays 215a and 215b such that the eyes can be positioned and aligned with respect to the eyepieces using the displays.
|0070] As described in greater detail below, for example, the user 1 14 may use images from the displays in order to adjust interpupilJary distance. In various embodiments, for example, proper alignment of two images presented by the displays may indicate that the interpupillary distance is appropriately adjusted. Thus, one or more adjustment controls 235 may be used to adjust the distance between the display targets 215a and 215b and/or between the eyepieces 203. The adjustment controls 235 may be provided on the sides of the main body 106 or elsewhere. In certain embodiments, the adjustment control 204 may comprise a handle on the main body 106, as shown in Figure 3B. In this embodiment, rotation of the adjustment control 204 may increase or decrease the interpupillary distance.
|0071] The combined light {that is reflected by mirror 260a or 260b and that comes from display 215a or 215b) is focused by adjustable powered optics (for example, lens) 210 possibly in conjunction with optical element 205. The adjustable optics 210 may comprise a zoom lens or lens system that may be have, for example, a focal length and/or power that is adjustable. The adjustable optics 210 may comprise or be part of an auto-focus system or may be manually adjusted. The adjustable optics 210 may provide optical correction for those in need of such correction (for example, a user whose glasses are removed during testing) The position of the powered optics 210 may be based on the signals obtained from the device, as described m more detail below The focused light then travels through eyepiece windows or lens 205, positioned at a proximal end of the eyepiece 203, towards the eye of a user 114. hi the case where a lens 205 is included, this lens 205 may contnbute to focusing of the light into the eye.
[0072] This light directed into the eye may be scattered by tissue or features therein A portion of this scattered light may be directed back into the eyepiece Lens 205 may thus receive light 207 reflected from the user s eye. which travels through the powered optics 210, reflects off of the beamsplitter 230a or 230b towards beamsplitter 220a or 220b, which reflects the light towards mirrors 295a or 295b At 295a or 295b, light reflected by the sample interferes with light in the reference arm (path between beamsplitter 285a or 285b and beamsplitter 295a or 295b that includes mirrors 273a or 273b and 270a or 270b) (Accordingly, the sample arm includes the optical path between beamsplitter 285a or 285b and beamsplitter 295a or 295b that includes mirrors 250a or 250b and 255a or 255b and the sample or eye.) The light is then reflected by mirror 225a or 225b towards switch 275 In some embodiments, the switch 275 comprises a switchable deflector that switches optical paths to the first or second eye to collect data from the respective eye to be sent to the data acquisition device 202 The switch may comprise a low-frequency switch, such that all data to be collected from one eye is obtained before the data is collected from the other eye Alternatively, the switch may comprise a high-frequency switch, which may interlace data collected from each eye
[0073] The instrument may be configured differently. For example, a common reference path may be used for each eye In some embodiments, the reference ami includes one or more movable mirrors to adjust the optical path length diffeience between the reference and sample arms. Components may be added, removed, or repositioned in other embodiments Other techniques, may be used
|0074] Although not shown, for example, polarizers and polarizing beamspitters may be used to control the propagation of light through the optical path in the optical system Other variations are possible. Other designs may be used [0075] In some embodiments, an A-scan may be formed in the time domain. In these instances, the Z-offset adjustment stage and corresponding mirror 255a or 255b and mirror 255a or 255b may change positions in time. Alternatively, reference mirrors 270a and 270b and reference mirrors 273a and 273b or other mirrors in the reference or sample arms may be translated. The combined light associated with various mirror positions may be analyzed to determine characteristics of an eye as a function of depth. In other embodiments, an A-scan may be formed in the spectral domain. In these instances, the frequencies of the combined light may be analyzed to determine characteristics of an eye as a function of depth. Additionally, one or more galvanometers 280 can control the horizontal and/or vertical location of the A-scan. Thus, a plurality of A-scans can be obtained to form a B-scan and/or a 3D-OCT.
[0076J Light output from the structure 275 can be input into a data acquisition device 202, which may comprise, for example, a spectrometer or a light meter. A grating may be in the main body 106. The data acquisition device 202 is coupled to a computer system 104, which may present output based on scans to the user 1 14. The output device may include a monitor screen, in which output results are displayed. The output device may include a printer, which prints output results. The output device may be configured to store data on a portable medium, such as a compact disc or USB drive, or a custom portable data storage device.
[0077] In some embodiments, the computer system 104 analyzes data received by the data acquisition device 202 in order to determine whether one or more of the adjustment stages 290a and/or 290b and/or one or more movable components and/or the powered optics 210 should be adjusted. In one instance, an A-scan is analyzed to determine a position (for example, a coarse position) of the retina such that data on the retina may be obtained by the instrument. In some embodiments, each A-scan comprises a plurality of light intensity values, each associated with a different depth into the sample. The A-scan may be obtained, in some embodiments, by translating the Z adjustment stage 290a or 290b. Likewise, the A- scan comprises values of reflected signal for obtained for different location of Z adjustment stage. The retina reflects more light than other parts of the eye, and thus, it is possible to determine a position of the adjustment stage 290a or 290b that effectively images the retina by assessing what depths provide an increase in reflected intensity. In some embodiments, the Z adjustment stage may be translated and the intensity values may be monitored. An extended peak in intensity for a number of Z adjustment stage positions may correspond to the retina. A variety of different approaches and values may be monitored to determine the location of the retina. For example, multiple A-scans maybe obtained at different depths and the integrated intensity of each scan may be obtained and compared to determine which depth provided a peak integrated intensity. In certain embodiments, intensity values within an A- scan can be compared to other values within the A-scan and/or to a threshold. The intensity value corresponding to the preferred location may be greater than a preset or relative threshold and/or may be different from the rest of the intensity values, (for example, by more than a specified number of standard deviations). A wide variety of approaches may be employed.
[0078] After the positions of the adjustment stages 290a and 290b have been determined, subsequent image analysis may be performed to account for vibration or movement of the user's head, eyes or retinas relative to the light source 240. A feedback system such as a closed loop feedback system may be employed in effort to provide a more stabilized signal in the presence of such motion. The optical coherence tomography signal may be monitored and feedback provided to, for example, one or more translation stages to compensate for such vibration or movement. In some embodiments, subsequent image analysis may be based on initial image and/or detect changes in image characteristics. For example, the image analysis may determine that the brightest pixel within an A-scan has moved 3 pixels from a previous scan. The adjustment stage 290a or 290b may thus be moved based on this analysis. Other approaches may be used.
|0079] In some instances, optical coherence tomography signals are used to adjust the powered optics 210 to provide for increased or improved focus, for example, when a patient needs refractive correction. Many users/patients, for example, may wear glasses and may be tested while not wearing any glasses. The powered optics 210 may be adjusted based on reflected signal to determine what added correction enhances signal quality or is otherwise an improvement. Accordingly, in some embodiments, a plurality of A-scans is analyzed in order to determine a position for the powered optics 2] 0. In some instances, a plurality of A- scans is analyzed in order to determine a position for the powered optics 210. In some embodiments, this determination occurs after the position of the adjustment stage 290a or 290b has been determined. One or more A-scans, one or more B-scans or a 3D-OCT may be obtained for each of a plurality of positions of the powered optics 210. These scans may be analyzed to assess, for example, image quality. The position of the powered optics 210 may be chosen based on these image quality measures.
[0080J The image quality measure may include a noise measure. The noise measure may be estimated based on the distribution of different intensity levels of reflected light within the scans. For example, lower signals may be associated with noise. Conversely, the highest signals may be associated with a saturated signal. A noise measure may be compared to a saturation measure as in signal to noise ratios or variants thereof. The lowest reflectivity measured (referred to as a low measure or low value) may also be considered. In some embodiments, the positions of the adjustment stages 290a and/or 290b and/or the powered optics 210 is determined based upon a signal-to-noise measure, a signal strength measure, a noise measure, a saturation measure, and a low measure. Different combinations of these parameters may also be used. Values obtained by integrating parameters over a number of positions or scans, etc., may also be used. Other parameters as well as other image quality assessments may also be used.
[0081] In one embodiment, a noise value is estimated to be a reflected light value for which approximately 75% of the measured reflected light is below and approximately 25% of the measured reflected light is above. The saturation value is estimated to be a reflected light value for which approximately 99% of the measured reflected light is below and approximately 1 % of the measured reflected light is above. A middle value is defined as the mean value of the noise value and the saturation value. An intensity ratio is defined as the difference between the saturation value and the low value divided by the low value multiplied by 100. A tissue signal ratio is defined as the number of reflected light values between the middle value and the saturation value divided by the number of reflected light values between the noise value and the saturation value. A quality value is defined as the intensity ratio multiplied by the tissue signal ratio. Additional details are described, for example, in Stein DM, Ishikawa H, Hariprasad R, Wollstein G, Noecker RJ, Fujimoto JG, Schuman JS. A new quality assessment parameter for optical coherence tomography. Br. J. Ophthalmol. 2006;90; 186-190. A variety of other approaches may be used to obtain a figure of merit to use to measure performance and adjust the instrument accordingly.
[0082] In the case of adjusting the adjustable power optics, 210, in some embodiments, a plurality of positions are tested. For example, the powered optics may be continuously moved in defined increments towards the eyes for each scan or set of scans. Alternatively, the plurality of positions may depend on previously determined image quality measures. For example, if a first movement of the powered optics 210 towards the eye improved an image quality measure but a subsequent second movement towards the eye decreased an image quality measure, the third movement may be away from the eye. Accordingly, optical power settings may be obtained that improve and/or maintains an improved signal. This optical power setting may correspond to optical correction and increase focus of the light beam in the eye, for example, on the retina, in some embodiments.
|0083] As described above, various embodiments employ an arrangement wherein a pair of oculars is employed. Accordingly, such adjustments, may be applied to each of the eyes as a user may have eyes of different size and the retina may located at different depths and thus a pair of z adjust stages may be used in some embodiments. Similarly, a user may have different prescription optical correction for the different eyes. A variety of arrangements may be employed to accommodate such needs. For example, measurements and/or adjustments may be performed and completed on one eye and subsequently performed and completed the other eye. Alternatively, the measurements and/or adjustments may be performed simultaneously or interlaced. A wide variety of other variations are possible.
[0084] Figure 4 shows a diagram of a spectrometer 400 that can be used as a data acquisition device 202 for a frequency domain OCT system. Light 405 input into the spectrometer 400 is collected by collecting lens 410. The collected light then projects through a slit 415, after which it is collimated by the collimating lens 420. The collimated light is separated into various spectral components by a grating 425. The grating 425 may have optical power to focus the spectral distribution onto an image plane. Notably, other separation components, such as a prism may be used to separate the light. The separated light is then directed onto a detector array by focusing lens 430, such that spectral components of each frequency from various light rays are measured.
[0085] A wide variety of OCT designs are possible. For example, frequency can be varied with time. The reference and sample arms can overlap. In some embodiments, a reference arm is distinct from a sample arm, while in other embodiments, the reference arm and sample arm are shared. See, for example, Vakhtin AB, Kane DJ, Wood WR and Peterson KA. "Common-path interferometer for frequency-domain optical coherence tomography," Applied Optics. 42(34), 6953-6958 (2003). The OCT arrangements should not be limited to those described herein. Other variations are possible.
[0086] In some embodiments, as shown in Figure 5, the main body 106 includes only a single display target 215. Light from the display target 215 is split at an x-prism 505. Notably, other optical devices that split the source light into a plurality of light rays may be used. This split light is reflected at mirror 51 Oa or 510b and directed towards the user 1 14.
|0087] The user may be directed to fixate on a display target 215 while one or more galvanometers 280 move light from the light source 240 to image an area of tissue. In some embodiments, the display targets 215 are moved within the user's field of vision while an area of tissue is imaged. For example, in Figure 6A7 a display target 215 may be moved horizontally (for example, in the medial-lateral direction), such that a patient is directed to look from left to right or from right to left. Meanwhile, a vertical scanner (for example, galvanometer) allows the vertical location (for example, in the superior-inferior) of the sample scanning to change in time. Figure 6 shows an eye, which is directed to move in the horizontal direction 605. Due to the vertical scanner, the scanned trajectory 610 covers a large portion of the eye 600. Scanning in the vertical and horizontal directions can produce a 3D-OCT. In some embodiments, continuous and/or regularly patterned A-scans are combined to form a full scan for example, B-scan or 3D-OCT. In other embodiments, discrete and/or random A-scans are combined to form the full scan. Systems configured such that users 114 are directed to move their eyes throughout a scan may include fewer scanners than comparable systems configured such that users 114 keep their eyes fixated at a stationary target. For example, instead of a system comprising both a vertical and a horizontal scanner, the user 114 may move his eyes in the horizontal direction, thereby eliminating the need for a horizontal scanner.
[0088] In other embodiments, two scanners (e.g., a vertical and a horizontal scanner) can be used. The design, capabilities and/or specifications for these scanners need not be the same. For example, one of the scanners may be faster and/or higher resolution than the other. Specifically, a vertical scanner may be used that scans more rapidly than a horizontal scanner, or vice versa. Scanners such as galvanometers having different speeds may be used in some example embodiments to scan continuously in the vertical direction and only occasionally increment along the horizontal (or vice versa). In some embodiments, for example, one of the scanners may be Vi to 1/500 as fast as the other scanner although values outside this range are possible. Similarly, in some embodiments, the 3D-OCT image may not contain as many pixels in one direction (e.g. horizontal) as the other direction (e.g.. vertical). hi some embodiments, for example, the 3D-OCT image may be 600 x 512 although other sizes are possible. Likewise, one scanner or galvanometer may have a reduced resolution compared to the other scanner. In instances wherein the specifications for one scanner include slower scan rates or less resolution than the other scanner, possibly a less expensive scanner or galvanometer may be used. Accordingly, the two scanners or galvanometers need not be the same type or grade. A relatively high performance (higher cost) and a relatively lower performance (lower cost) scanner may be used. Use of a lower performance/cost scanner or galvanometer instead of two scanners or galvanometers of equal quality and cost may reduce the overall cost of the instrument. Other variations are also possible. Some embodiments disclosed herein refer to one or more galvanometers. In some embodiments, a different kind of scanner may be used in place of the galvanometer.
J0089] Figure 6B shows an example of an A scan. The A scan comprises the signal strength (indicated by the brightness) as a function of depth for one horizontal and vertical position. Thus, an A-scan comprises a plurality of intensity values corresponding to different anterior-posterior positions. A plurality of A scans form a B scan. Figure 6C shows a B-scan, in which the largest portion of the bright signal corresponds to retinal tissue and the elevated region under the retina corresponds to diseased tissue within the eye. [0090] With reference to Figure 7A, there is illustrated an enlarged view depicting an embodiment of the main body 106 that is configured with a handle 1 18 for adjusting the eyepieces to conform to the user's interpupillary distance. In the illustrative embodiment, the main body 106 comprises a left eyepiece 712 and a right eyepiece 714 wherein each is connected to the other by interpupillary distance adjustment device 718. The interpupillary distance adjustment device 718 is coupled to the handle 1 18, wherein the handle 1 18 is configured to allow the user to engage the handle 118 to adjust the distance between the left and right eyepieces 712, 714 to match or substantially conform to the interpupillary distance between the eyes of the user.
[0091] Referring to Figure 7A, the user can rotate, turn, or twist the handle 118 to adjust the distance between the left and right eyepieces 712, 714 so as to match or substantially conform to the interpupillary distance between the eyes of the user. Alternatively, the handle 1 18 can be configured to move side to side to allow the user to adjust the distance between the left and right eyepieces 712, 714. Additionally, the handle 1 18 can be configured to move forward and backward to allow the user to adjust the distance between the left and right eyepieces 712, 714. In the alternative, the handle 1 18 can be configured to move up and down to allow the user to adjust the distance between the left and right eyepieces 712, 714. In another embodiment, the distance between the left and right eyepieces 712. 714 can be adjusted and/or controlled by a motor activated by the user. Alternatively, the motor can be configured to be controlled by computer system 104 to semi- automatically position the left and right eyepieces 712. 714 to match the interpupillary distance between the eyes of the user. In these instances, eye tracking devices may be included with a system described herein. In other embodiments, a combination of the foregoing are utilized to adjust the distance between the left and right eyepieces 712. 714 to match or substantially conform to the user's interpupillary distance.
[0092] A user 1 14 may adjust interpupillary distance based on the user's viewing of one or more fixation targets on one or more displays 215. For example, the displays 215 and the fixation targets may be configured such that the user views two aligned images, which may form a single, complete image when the interpupillary distance is appropriate for the user 114. The user 1 14 may adjust (for example, rotate) an adjustment control 204 to change the interpupillary distance based on the fixation target images, as shown in Figure 7A. Figures 7B-7F illustrate one embodiment of fixation targets as seen by the viewer under a plurality of conditions; however, other fixation targets are possible, including but not limited to a box configuration. Figure 7B shows a U-shaped fixation target 715a on the display 215a for the left eye. Figure 7C shows an upside-down U-shaped fixation target 715b on the display 215b for the right eye.
|0093] When the interpupillary distance is appropriately adjusted, the bottom and top images 715a and 715b are aligned, as shown in Figure 7D to form a complete H-shaped fixation target 715. When the interpupillary distance is too narrow, the fixation target 715a on the display 215a for the left eye appear shifted to the right and the fixation target on the display 215b for the right eye appear shifted to the left and the user sees the image shown in Figure 7E. Conversely, when the interpupillary distance is too wide, the fixation target 715a on the display 215a for the left eye appear shifted to the left and the fixation target on the display 215b for the right eye appear shifted to the right and the user sees the image shown in Figure 7F . Thus, the interpupillary distance may be adjusted based on these images.
[0094] In particular, in Figure 7D, the alignment image 715 is in the shape of an "H." Thus, when the interpupillary distance is properly adjusted, the fixation targets on the left and right displays overlap to form an "H". Other alignment images 715 may be provided.
[0095] With reference to Figure 8, there is illustrated an embodiment of the computer system 104. In the illustrated embodiment, the computer system 104 can comprise a scan control and analysis module 824 configured to control the scanning operations performed by the main body 106. The computer system 104 can also comprise a fixation marker control system 822 configured to display a fixation marker visible by the user from main body 106. In certain embodiments, the fixation marker is displayed as an ''X," a dot, a box, or the like. The fixation marker can be configured to move horizontally, vertically, diagonally, circularly, or a combination thereof. The fixation marker can be repositioned quickly to relocate the beam location on the retina as the eye repositions itself. The computer system 104 can also comprise a focus adjust module 820 for automatically adjusting the focusing lenses in the main body 106 as further discussed herein. The computer system 104 can also comprise a Z positioning module 818 for automatically adjusting the Z offset as herein discussed.
[0096] Referring to Figure 8, the computer system 104 comprises in the illustrative embodiment a disease risk assessment / diagnosis module 808 for storing and accessing information, data, and algorithms for determining, assessing the risk or likelihood of disease, and/or generating a diagnosis based on the data and/or measurements obtained from scanning the eyes of the user. In one embodiment, the scan control and analysis module 824 is configured to compare the data received from the main body 106 to the data stored in the disease risk assessment / diagnosis module 808 in order to generate a risk assessment and/or diagnosis of disease in the eyes of the user as further illustrated. The computer system 104 can also comprise an image/scans database configured to store images and/or scans generated by the main body 106 for a plurality of users, and to store a unique identifier associated with each image and/or scan. In certain embodiments, the scan control and analysis module 824 uses historical images and/or scans of a specific user to compare with current images and/or scans of the same user to detect changes in the eyes of the user. In certain embodiments, the scan control and analysis module 824 uses the detected changes to help generate a risk assessment and/or diagnosis of disease in the eyes of the user.
|0097} In the illustrative embodiment shown in Figure 8. the computer system 104 can comprise a user / patient database 802 for storing and accessing patient information, for example, user name, date of birth, mailing address, residence address, office address, unique identifier, age. affiliated doctor, telephone number, email address, social security number, ethnicity, gender, dietary history and related infoπnation, lifestyle and/or exercise history information, use of corrective lens, family health history, medical and/or ophthalmic history, prior procedures, or other similar user information. The computer system 104 can also comprise a physician referral database for storing and accessing physician information, for example, physician name, physician training and/or expertise/specialty, physician office address, physician telephone number and/or email address, physician scheduling availability, physician rating or quality, physician office hours, or other physician information.
|0098] In reference to Figure 8, the computer system 104 can also comprise a user interface module 805 (which can comprise without limitation commonly available input/output (I/O) devices and interfaces as described herein) configured to communicate, instruct, and/or interact with the user through audible verbal commands, a voice recognition interface, a key pad, toggles, a joystick handle, switches, buttons, a visual display, touch screen display, etc. or a combination thereof. In certain embodiments, the user interface module 805 is configured to instruct and/or guide the user in utilizing and/or positioning the main body 106 of the optical coherence tomography system 100. The computer system 104 can also comprise a reporting / output module 806 configured to generate, output, display, and/or print a report (for example, Figures 1OA and 10B) comprising the risk assessment and/or diagnosis generated by the disease risk assessment / diagnosis module 808. In other embodiments, the report comprises at least one recommended physician to contact regarding the risk assessment.
|0099 j Referring to Figure 8, the computer system 104 can also comprise an authentication module 816 for interfacing with user card reader system 1 12. wherein a user can insert a user identification card into the user card reader system 1 12. In certain embodiments, the authentication module 816 is configured to authenticate the user by reading the data from the identification card and compare and/or store the information with the data stored in the user / patient database 802. In certain embodiments, the authentication module 816 is configured to read or obtain the user's insurance information from the user's identification card through the user card reader system 1 12. The authentication module 816 can be configured to compare the user's insurance information with the data stored in the insurance acceptance database 828 to determine whether the user's insurance is accepted or whether the user's insurance company will pay for scanning the user's eyes. In other embodiments, the authentication module communicates with the billing module 810 to send a message and/or invoice to the user's insurance company and/or device manufacturer to request payment for performing a scan of the patient's eyes. The card can activate one or more functions of the machine allowing the user, for example, to have a test performed or receive output from the machine. In other embodiments, the billing module 810 is configured to communicate with the user interface module 805 to request payment from the user to pay for all or some (for example, co-pay) of the cost for performing the scan. In certain embodiments, the billing module 810 is configured to communicate with the user card reader system 112 to obtain card information from the user's credit card, debit card, gift card, or draw down credit stored on the user's identification card. Alternatively, the billing module 810 is configured to receive payment from the user by communicating and/or controlling an interface device for receiving paper money, coins, tokens, or the like. Alternatively, the billing module 810 is configured to receive payment from the user by communicating with the user's mobile device through Bluetooth® or other communications protocols/channels in order to obtain credit card information, billing address, or to charge the users mobile network service account (for example, the cellular carrier network).
|0100] With reference to Figure 8, the user card may be used by insurers to track which users have used the system. In one embodiment, the system can print (on the face of the card) or store (in a chip or magnetic stripe) the scan results, risk assessment, and/or report directly onto or into the card that the patient inserts into the system (wherein the card is returned to the user). The system can be configured to store multiple scan results, risk assessments, and/or reports, and/or clear prior scan results, risk assessments, and/or reports before storing new information on the magnetic stripe. In certain embodiments, the calculation of the risk assessment is performed by the system (for example, scanning analysis module 824). In certain embodiments, the calculated risk assessment is transmitted a centralized server system (for example, remote systems 1 10) in another location that provides the results via a web page to physicians, users, patients, or the like. The centralized server system (for example, remote system 1 10) allows the user, patients, or doctors to enter their card code to see the results which are saved in the centralized database.
[0101] In the example embodiment of Figure 8, the computer system 104 can comprise a network interface 812 and a firewall 814 for communicating with other remote systems U O through a communications medium 108. Other remote systems 1 10 can comprise without limitation a system for checking the status/accuracy of the optical coherence tomography system 100; a system for updating the disease risk assessment / diagnosis database 808, the insurance acceptance database 828, the physician referral database 804, and/or the scan control and analysis module 824. In certain embodiments, the computer system 104 can be configured to communicate with a remote system 110 to conduct a primary and/or secondary risk assessment based on the data from scanning the user's eyes with the main body 106.
[0102} Referring to Figure 8, the remote system 1 10 can be configured to remotely perform (on an immediate, delayed, and/or batch basis) a risk assessment and/or diagnosis and transmit through a network or communications medium the risk assessment, diagnosis, and/or report to the computer system 104 for output to the user using output device 102. In certain embodiments, the output device 102 is configured to display the risk assessment, diagnosis, and/or report as a webpage that can be printed, emailed, transmitted, and/or saved by the computer system 104. The remote system 1 10 can also be configured to transmit through a network or communications medium the risk assessment, diagnosis, and/or report to the user's (or doctor) cellular phone, computer, email account, fax, or the like.
10103] With reference to Figure 9, there is shown an illustrated method of using the optical coherence tomography system 100 to self-administer an OCT scan of the user's eyes and obtain a πsk assessment or diagnosis of various diseases and ailments. The process begins at block 901 wherein the user approaches the optical coherence tomography system 100 and activates the system, by for example pushing a button or typing in a activation code or anonymous identification number, hi other embodiments, the user interface module 805 instructs users at block 901 to first insert an identification card or anonymous coded screening card in user card reader system 112 to activate the system. The system can also be activated at block 901 when users insert their user identification card in user card reader system 112 Other means of activating the system are possible as well as, including without limitation, a motion sensor, a weight sensor, a radio frequency identification (RFID) device, or other actuator to detect the presence of the user. Alternatively, the optical tomography system 100 can be activated when the billing module 810 detects that the user has inserted paper money, coins, tokens, or the like into an interface device configured to receive such payment. Alternatively, the billing module 810 can also be configured to activate the optical tomography system 100 when the billing module 810 communicates with a user's mobile device in order to obtain the user's credit card information, billing address, or the like, or to charge the user's mobile network service account (for example, the cellular earner network) |0104] In referring to Figure 9 at block 902, the user interface module 805 is configured to direct the user to attach disposable eyecups onto the main body 106, and then position the main body 106 with the disposable eyecups near the eyes of the user and/or support the disposable eyecups against the user's eye socket. The user interface module 805 instructs the user to engage handle 1 18 to adjust the distance between the left and right eyepieces 612, 614 to match or substantially conform to the interpupillary distance of the user as described with respect to Figures 6A-6F. After the main body 106 and the interpupillary distance has been appropriately calibrated and/or adjusted by the user, the user inputs into or indicates to the user interface module 805 to begin the scan. The scan control and analysis module 824 substantially restricts movement or locks the position of the zero gravity arm and/or the distance between the left and right tubes 612, 614 to begin the scan.
[0105] Referring to Figure 9, the Z module 818 automatically adjusts the z-offset in the main body 106 at block 906 such that the OCT measurement will be obtained, for example, from tissue in the retina. The Z module 818 may identify and/or estimate a position of part of the sample (for example, part of an eye of a user 1 14) and adjust the location of one or more optical components based on the position. One of ordinary skill in the art will appreciate the multitude of ways to perform such an adjustment. For example, the Z module 818 may comprise a motor, such as a piezoelectric motor, to translate the reference mirror/s longitudinally such that the optical path length from the beam splitter to the retina is about equal to (within a coherence length of) the optical path length in the reference arm. This movement may enable light from the reference arm to interfere with light reflected by a desired portion of the sample (for example, the retina). At block 908, the illustrative method performs a focus adjustment using the focus adjustment module 820. Those of ordinary skill in the art will also appreciate the different techniques for performing such auto-focus calibration. Block 910 illustrates an optional test performed by the computer system 104 to determine the visual function and/or acuity of the user's eye. Such visual function and/or acuity tests will be appreciated by those skilled in the art. In one embodiment, the visual acuity test works with or is combined with the fixation marker control system 722, and can test both eyes simultaneously or one eye at time. For example, the fixation marker will initially appear small and then gradually increase in size until the user indicates through the user interface module 705 that the fixation marker is visible. Based on the size at which the user can clearly see the fixation marker, fixation marker control system 722 can estimate or determine or assess the visual acuity of the users eyes (for example, 20/20, 20/40, or the like).
[0106] With reference to Figure 9 at Block 912, the user interface module 805 instructs the user to follow the movement of the fixation marker that is visible to the user from the main body 106. In one embodiment, the fixation marker control 822 is configured to display a fixation marker that moves horizontally. In some embodiments, the horizontal movement of the fixation marker allows the scan control and analysis module 824 to scan the eye vertically as the eye moves horizontally, thus possibly obtaining a two-dimensional, volume, or raster scan of the eye tissue at issue. Alternatively, the scan control and analysis module 824 and/or the fixation marker control may cause the fixation marker or the beam to jump or move around to obtain measurements at different lateral locations on the eye.
[0107] During the scanning of the eye, the scan control and analysis module 824 could be configured to detect at block 913 whether there has been a shift in the position of the main body 106 relative to the user. In one embodiment, the scan control and analysis module 824 can detect (in real-time, substantially real-time, or with a delay) whether a shift has occurred based on what the values the module 824 expects to receive during the scanning process. For example, as the scan control and analysis module 824 scans the retina, the module 824 expects to detect a change in signal as the scanning process approaches the optic nerve (for example, based on the location of the fixation target and/or state of the scanner(s)). Alternatively, the expected values or the expected change in values can also be determined or generated using a nomogram. If the system does not detect an expected signal change consistent with a detection of the optic nerve and/or receives no signal change, then the module 824 can be configured to interpret such data as the user is not tracking properly. Other features, for example, the fovea, or the like, can be used to determine whether the expected signal is observed. If improper tracking occurs enough (based on, for example, a threshold), the system 100 may request that the user fixate again (using fixation marker control 822) for another scan. If the foregoing shift detection process does not occur in realtime or substantially real-time, then the system can be configured to complete the scan, perform data analysis, and during the analysis the system can be configured to detect whether a shift occurred during the scan. If a substantial shift is detected, then the user may be instructed (through visual, audible, or verbal instructions using the user interface module 805) to sit forward again so another scan can be performed. If the system detects a shift 2 or 3 or more times, the system can be configured to refer the user to a general eye doctor.
[0108] At the end of a scan, the scan control and analysis module 824 can be configured to produce a confidence value that indicates how likely the nomograms wil! be to apply to this patient. For example, if the patient had borderline fixation, the confidence value might be lower than a patient whose fixation appeared to be good.
[0109] hi the real-time embodiment, the system can be configured to perform rapid cross-correlations between adjacent A-scans or B-scans to make sure the eye is moving somewhat. In some embodiments, the foregoing can be advantageous for ANSI laser safety standards so as to avoid having users stare at the same location with laser energy bombarding the user's retina. Accordingly, in some embodiments, the system is configured with a laser time-out feature if the system detects no eye moment (for example, cross-correlations above a certain threshold). In some embodiments, to expedite this process and provide real time analysis in frequency domain OCT, signal data may be analyzed prior to performing an FFT. Other technologies can be used to deterrmne that the user has some eye movement.
[0110] If no fixation problem has been detected, the scan control and analysis module 824 completes the scan of the user's eyes, stores the image and/or scan data in the images/scans database 826, and analyzes the A-scan data at block 915 to generate/determine a nsk assessment and/or diagnosis at block 916 by accessing the data and/or algorithms stored m the disease nsk assessment/diagnosis database 808. In some embodiments, groups of A-scans, partial or full B scans, or partial or full 3D-OCT data can be analyzed.
[0111] As used herein the term "nomogram" generally refers to predictive tools, algorithms, and/or data sets. Nomograms in general can provide predictions for a user based on the comparison of characteristics of the user with the nomogram. The nomograms are deπved, generated, calculated, or computed from a number, for example, hundreds, thousands, or millions of users/patients who exhibited the same condition (normal or diseased). In some embodiments descπbed herein, nomograms compare the risk of having a disease based on physical characteristics. Accordingly, in some cases, nomograms can provide individualized predictions that are relative to risk groupings of patient populations who share similar disease characteristics. In some embodiments, nomograms can be used to provide the risk estimation or risk assessment on a 0-100% scale. Alternatively, nomograms used herein can provide an expected value, for example, at a certain position in the eye there is an expected eye thickness value of 100 microns.
[0112] Generally, nomograms have been developed and validated in large patient populations and are highly generalizable, and therefore, nomograms can provide the objective, evidence-based, individualized risk estimation or assessment. Accordingly, nomograms can be used as described herein to empower patients and allow them to better understand their disease. Further, nomograms as used herein can assist physicians with clinical decision-making and to provide consistent, standardized and reliable predictions.
[0113] In the illustrative method shown in Figure 9 at block 917, an eye health assessment or eye health grade report, as illustrated in Figures 1OA and ] 0B, is generated for the user by accessing the disease risk assessment / diagnosis database 808. At block 918, the physician referral database 804 is accessed to generate a recommendation of when the user should visit a physician (for example, within one to two weeks). The physician referral database 804 is also accessed to generate, compile a listing of physicians suitable for treating the patient. The physician referral list can be randomly generated or selected based on referral fee payments paid by physicians, insurance companies, or based on location of the physician relative to the user's present location or office/home address, or based on the type of detected disease, or based on the severity of the detected disease, based on the location or proximity of the system relative the location of the physician, or based on a combination thereof. At block 919, the report is displayed to the user by using reporting / output module 806 and output device 102. In certain embodiments, the report data is stored in the user/patient database 802 for future analysis or comparative analysis with future scans.
[0114] In some embodiments, the main body 106 is not supported by the user 1 14. For example, the main body 106 may be supported by a free-standing structure, as shown in Figure 1OA. The user 1 14 may look into the eyepiece(s). The user 1 14 may be seated on a seating apparatus, which may include a height-adjusting mechanism. The main body 106 may supported by a height- adjustable support.
J0115] In some embodiments, such as those shown in Figures 10B- 1OC, a strap 1005 is connected to the main body 106. The strap may function to fully or partly support the main body 106, as shown in Figure 1OB. The strap 905 may be excluded in some embodiments. The main body 106 may be hand held by the user. In some embodiments, the main body 106 may be supported on eyewear frames, hi some embodiments, all of the optics are contained within the main body 106 that is directly or indirectly supported by the user 114. For example, the main body 106 in Figure 1OB may include an optical coherence tomography system, an alignment system, and a data acquisition device. The data acquisition device may wirelessly transmit data to a network or computer system or may use a cable to transfer control signals. Figure 1OC is similar to that of Figure 1 and is supported by a separate support structure (for example, an zero gravity arm). In some embodiments, a strap, belt, or other fastener assists in the alignment of the main body 106 with one or both eyes of the user 1 14.
|0116] hi some embodiments, as shown in Figure 10D, the user wears an object 1010 connected to the eyepiece. The wearable object 1010 may include a head-mounted object, a hat or an object to be positioned on a user's head. As described above, in some embodiments, the main body 106 is supported on an eyewear frame worn by the user like glasses. The wearable object 1010 may fully or partly support the main body 106 and/or may assist in aligning the main body 106 with one or both eyes of the user 114.
|0117] Referring to Figure HA and HB. there are illustrated two example embodiments of the eye health grades and the eye health assessment reports. With reference to Figure 1 IA, the eye health grades report can comprise without limitation a numeric and/or letter grade for each eye of the user for various eye health categories, including but not limited to macular health, optic nerve health, eye clarity, or the like. The eye health grades report can also comprise at least one recommendation to see or consult a physician within a certain period of time, and can provide at least one possible physician to contact. Data for generating the recommendation infoπnation and the list of referral physicians are stored in the physician referral database 804. In reference to Figure 1 IB, the eye health assessment report can comprise a graphical representation for each eye of the user for various eye health categories. The report can be presented to the user on an electronic display, printed on paper, printed onto a card that the user inserted into the machine, electronically stored on the users identification card, emailed to the user, or a combination thereof.
[0118] With reference to Figure 12, there is illustrated another embodiment of the computer system 104 connected to remote system 110 and billing / insurance reporting and payment systems 1201. The billing module 810 can be configured to communicate with billing / insurance reporting payment systems 1201 through communications medium 108 in order to request or process an insurance claim for conducting a scan of the user's eyes. Based on communications with billing / insurance reporting and payment system 1201, the billing module 810 can also be configured to determine the amount payable or covered by the user's insurance company and/or calculate or determine the co-pay amount to be charge the consumer. In certain embodiments, the user can interact with the user interface module 805 to schedule an appointment with the one of the recommended physicians and/or schedule a reminder to be sent to the user to consult with a physician. The computer system 104 or a remote system 1 10 can be configured to send the user the reminder via email, text message. regular mail, automated telephone message, or the like. Computing System
{0119] In some embodiments, the systems, computer clients and/or servers described above take the form of a computing system 1300 shown in Fig. 13, which is a block diagram of one embodiment of a computing system (which can be a fixed system or mobile device) that is in communication with one or more computing systems 1310 and/or one or more data sources 1315 via one or more networks 1310. The computing system 1300 may be used to implement one or more of the systems and methods described herein. In addition, in one embodiment, the computing system 1300 may be configured to process image files. While Fig. 13 illustrates one embodiment of a computing system 1300. it is recognized that the functionality provided for in the components and modules of computing system 1300 may be combined into fewer components and modules or further separated into additional components and modules. Client / Server Module [0120] In one embodiment, the system 1300 comprises an image processing and analysis module 1306 that carries out the functions, methods, and/or processes described herein. The image processing and analysis module 1306 may be executed on the computing system 1300 by a central processing unit 1304 discussed further below. Computing System Components
[0121] In one embodiment, the processes, systems, and methods illustrated above may be embodied in part or in whole in software that is running on a computing device. The functionality provided for in the components and modules of the computing device may comprise one or more components and/or modules. For example, the computing device may comprise multiple central processing units (CPUs) and a mass storage device, such as may be implemented in an array of servers.
[0122] In general, the word "module," as used herein, refers to logic embodied in hardware or firmware, or to a collection of software instructions, possibly having entry and exit points, written in a programming language, such as, for example, Java, C or C++, or the like. A software module may be compiled and linked into an executable program, installed in a dynamic link library, or may be written in an interpreted programming language such as, for example, BASIC, Perl, Lua, or Python. It will be appreciated that software modules may be callable from other modules or from themselves, and/or may be invoked in response to detected events or interrupts. Software instructions may be embedded in firmware, such as an EPROM. It will be further appreciated that hardware modules may be comprised of connected logic units, such as gates and flip-flops, and/or may be comprised of programmable units, such as programmable gate arrays or processors. The modules described herein are preferably implemented as software modules, but may be represented in hardware or firmware. Generally, the modules described herein refer to logical modules that may be combined with other modules or divided into sub-modules despite their physical organization or storage.
|0123] In one embodiment, the computing system 1300 also comprises a mainframe computer suitable for controlling and/or communicating with large databases, performing high volume transaction processing, and generating reports from large databases. The computing system 1300 also comprises a central processing unit ("CPU'") 1304, which may comprise a conventional microprocessor. The computing system 1300 further comprises a memory 1305, such as random access memory ("RAM") for temporary storage of information and/or a read only memory ("ROM'") for permanent storage of information, and a mass storage device 1301, such as a hard drive, diskette, or optical media storage device. Typically, the modules of the computing system 1300 are connected to the computer using a standards based bus system, hi different embodiments, the standards based bus system could be Peripheral Component Interconnect (PCI), MicroChannel. SCSI, Industrial Standard Architecture (ISA) and Extended ISA (EISA) architectures, for example.
[0124] The example computing system 1300 comprises one or more commonly available input/output (I/O) devices and interfaces 1303, such as a keyboard, mouse, touchpad, and printer. In one embodiment, the I/O devices and interfaces 1303 comprise one or more display devices, such as a monitor, that allows the visual presentation of data to a user. More particularly, a display device provides for the presentation of GUIs, application software data, and multimedia presentations, for example. In the embodiment of Fig. 13, the I/O devices and interfaces 1303 also provide a communications interface to various external devices. The computing system 1300 may also comprise one or more multimedia devices 1302, such as speakers, video cards, graphics accelerators, and microphones, for example. Computing System Device/Operating System
10125] The computing system 1300 may run on a variety of computing devices, such as, for example, a server, a Windows server, a Structure Query Language server, a Unix server, a personal computer, a mainframe computer, a laptop computer, a cell phone, a personal digital assistant, a kiosk, an audio player, and so forth. The computing system 1300 is generally controlled and coordinated by operating system software, such as z/OS, Windows 95, Windows 98, Windows NT, Windows 2000, Windows XP, Windows Vista, Linux, BSD, SunOS, Solaris, or other compatible operating systems, hi Macintosh systems, the operating system may be any available operating system, such as MAC OS X. In other embodiments, the computing system 1300 may be controlled by a proprietary operating system. Conventional operating systems control and schedule computer processes for execution, perform memory management, provide file system, networking, and I/O services, and provide a user interface, such as a graphical user interface ('OUF"), among other things. Network
[0126] In the embodiment of Fig. 13, the computing system 1300 is coupled to a network 1310, such as a modem system using POTS/PSTN (plain old telephone service/public switched telephone network), ISDN, FDDI, LAN, WAN, or the Internet, for example, via a wired, wireless, or combination of wired and wireless, communication link 1315. The network 1310 communicates (for example, constantly, intermittently, periodically) with various computing devices and/or other electronic devices via wired or wireless communication links. In the example embodiment of Fig. 13, the network 1310 is communicating with one or more computing systems 1317 and/or one or more data sources 1319.
[0127] Access to the image processing and analysis module 1306 of the computer system 1300 by remote computing systems 1317 and/or by data sources 1319 may be through a web-enabled user access point such as the computing systems' 1317 or data source's 1319 personal computer, cellular phone, laptop, or other device capable of connecting to the network 1310. Such a device may have a browser module implemented as a module that uses text, graphics, audio, video, and other media to present data and to allow interaction with data via the network 1310.
[0128] The browser module or other output module may be implemented as a combination of an all points addressable display such as a cathode-ray tube (CRT), a liquid crystal display (LCD), a plasma display, or other types and/or combinations of displays. In addition, the browser module or other output module may be implemented to communicate with input devices 1303 and may also comprise software with the appropriate interfaces which allow a user to access data through the use of stylized screen elements such as, for example, menus, windows, dialog boxes, toolbars, and controls (for example, radio buttons, check boxes, sliding scales, and so forth). Furthermore, the browser module or other output module may communicate with a set of input and output devices to receive signals from the user.
[0129] The input device(s) may comprise a keyboard, roller ball, pen and stylus, mouse, trackball, voice recognition system, or pre-designated switches or buttons. The output device(s) may comprise a speaker, a display screen, a printer, or a voice synthesizer. In addition a touch screen may act as a hybrid input/output device, hi another embodiment, a user may interact with the system more directly such as through a system terminal connected to the score generator without communications over the Internet, a WAN, or LAN, or similar network.
[01301 In some embodiments, the system 1300 may comprise a physical or logical connection established between a remote microprocessor and a mainframe host computer for the express purpose of uploading, downloading, or viewing interactive data and databases online in real time. The remote microprocessor may be operated by an entity operating the computer system 1300, including the client server systems or the main server system, and/or may be operated by one or more of the data sources 1319 and/or one or more of the computing systems, hi some embodiments, terminal emulation software may be used on the microprocessor for participating in the micro-mainframe link.
[0131] In some embodiments, computing systems 1317 that are internal to an entity operating the computer system 1300 may access the image processing and analysis module 1306 internally as an application or process run by the CPU 1304. User Access Point
[0132] In one embodiment, a user access point comprises a personal computer, a laptop computer, a cellular phone, a GPS system, a Blackberry® device, a portable computing device, a server, a computer workstation, a local area network of individual computers, an interactive kiosk, a personal digital assistant, an interactive wireless communications device, a handheld computer, an embedded computing device, or the like. Other Systems
[0133] In addition to the systems that are illustrated in Fig. 13, the network 1310 may communicate with other data sources or other computing devices. The computing system 1300 may also comprise one or more internal and/or external data sources. In some embodiments, one or more of the data repositories and the data sources may be implemented using a relational database, such as DB2, Sybase, Oracle, CodeBase and Microsoft® SQL Server as well as other types of databases such as, for example, a flat file database, an entity- relationship database, and object-oriented database, and/or a record-based database. [0134] With reference to Figure 14A, there is illustrated an example method for determining or generating a risk assessment of a disease, such as an eye disease, thereby allowing the generation of a health grade and recommended time to see a physician. The example shown in Figure 14A is for retinal disease, however, the process and method illustrated can be used for other diseases or eye diseases. In this example, the scan control and analysis module 824 is configured to determine the thickness of the retina based on the A-scan data derived from the main body 106. This data may include but is not limited to A- scan data from different A-scans. The scan control and analysis module 824 can also be configured to access data and algorithms in the disease risk assessment / diagnosis database 808 to calculate the risk assessment of retinal disease based on the measured thickness of the retina as illustrated by the function curve in Figure 14A. The reporting / output module 806 can be configured to normalize the calculated risk assessment value into an eye health letter or numerical grade or score. The reporting / output module 806 can also be configured to access data and algorithms in the physician referral database 804 to calculate a recommended time to see a physician based on the calculated risk assessment value.
[0135] With reference to Figure 14B, there is illustrated another example method or process for determining or generating a risk assessment of disease by comparing the scan data to the disease risk assessment / diagnosis database 808 comprising, for example, minimum and maximum thickness data and algorithms, and such minimum and maximum thickness data and algorithms that can be based on or are in the form of nomograms. In certain embodiments, the system is configured to generate scan data for portions of the eye scanned to determine thickness of the retina at any one point, and compare such data to histograms and/or nomograms (for example, nomograms that show expected thickness at said location likelihood of or disease for a given thickness) to derive a risk assessment. The system can also be configured to generate an average thickness for the entire retina that is scanned, and compare such data to histograms and/or nomograms to derive a risk assessment.
[0136] The term "histogram" as used herein generally refers to an algorithm, curve, or data or other representation of a frequency distribution for a particular variable, for example, retinal thickness. In some cases, the variable is divided into ranges, interval classes, and/or points on a graph (along the X-axis) for which the frequency of occurrence is represented by a rectangular column or location of points; the height of the column and/or point along the Y-axis is proportional to or otherwise indicative of the frequency of observations within the range or interval. "Histograms," as referred to herein, can comprise measured data obtained, for example, from scanning the eyes of a user, or can comprise data obtained from a population of people. Histograms of the former case can be analyzed to determine the mean, minimum, or maximum values, and analyze changes in slope or detect shapes or curvatures of the histogram curve. Histograms of the latter case can be used to determine the frequency of observation of a measured value in a surveyed sample.
[0137] In the instance where an average thickness value is derived from the scan data, there are some conditions/diseases that may be indicated by thickening of the retina in a localized area. Accordingly, such a condition may not significantly affect the average thickness value (for example, if a substantial portion of the retina is of normal thickness). Therefore, the maximum thickness value may be needed to detect this abnormal thickening in the retina. In some embodiments, this maximum thickness value may be due to a segmentation error. Accordingly, a more stable way of determining the maximum value may also be to use the value corresponding to 95% (or any value between 75% and 99%) maximal thickness. The foregoing can also be applied to minimum retinal thickness or any other value, measurement, and/or detectable condition in the eye. For example, with minimum retinal thickness, if the user has a macular hole, there will only be a small area of zero thickness, and possibly not enough to significantly reduce the average thickness, but definitely an abnormality that may be detected.
|0138] In other embodiments, the system may be configured to create histograms of measured thickness and/or measured intensity values and/or slopes or derivatives of intensity values and/or variables to identify abnormalities. For example, changes or substantial changes in slope (calculated as the derivative of adjacent intensity values) may indicate hyporeflective or hyperreflective structures that may not affect mean or average intensity values, but may be indicative of disease or conditions. For example, the system can determine if the distribution of retinal thicknesses across the measured portion of the retina matches that of the normal population. Deviation from such a "normal" histogram would result in lower health grades/ higher risk assessments. [0139] In various embodiments, the methods or processes described herein can be used to determine or generate a risk assessment of maculopathy based, for example, on abnormal thickening of the retina or fovea, the presence of hyperreflective (bright or high intensity) or hyporeflective (dark or low intensity) structures in the outer half of the retina, the presence of hyporeflective (dark) structures in the inner half of the retina, the presence of irregularities in the contour of the retinal pigment epithelium that depart from the normal curvature of the eye, or of the presence of hypertransmissϊon of light through the retinal pigment epithelium when compared to a database of normal values stored in the disease risk assessment / diagnosis database 708.
|0140] As described above, there are several ways to detect or generate a risk assessment for several diseases or conditions. In certain embodiments, scan data is compared to data found in normal people to identify similarities or differences from a nomogram and/or histogram. In other embodiments, scan data is compared to data found in people with diseases to identify similarities or differences from nomograms and/or histograms. The pathognomonic disease features could be indicated by similarity to nomograms, for example, images, histograms, or other data, etc. from diseased patients.
[0141] In one embodiment, "normal'" data (for example, histograms) are created for retinal thickness in each region of the retina (optic nerve, fovea, temporal retina) and compare to measured, detected, scanned, or encountered values to these "normal" data (for example, histograms) to determine relative risks of retinal disease or other diseases. The same can be performed for nerve fiber layer (NFL) thickness to detect glaucoma. In other embodiments, the detection or generation of a risk assessment for glaucoma is performed or generated by analyzing collinear A-scan data to see if curvilinear thinning indicates the presence of glaucoma because glaucoma tends to thin the NFL in curvilinear bundles. The NFL radiates out from the optic nerve in a curvilinear fashion like iron filings around a magnet. Measuring and analyzing a sequence of A-scan data that follow such a curvilinear path may be useful to identify such thinning that is characteristic of glaucoma. The analysis could be centered on and/or around the optic nerve or centered on and/or around the fovea or elsewhere. In another embodiment, the detection and/or generation of a risk assessment for glaucoma is performed or generated by analyzing the inner surface of the optic nerve to determine the optic disc cup volume.
[0142] The system can also be configured to detect and/or generate a risk assessment for optical clarity wherein the system integrates A-scan data in the Z direction and compares some or all the A-scan data to a nomogram value or values, or, for example, a histogram. In general, darker A-scans will probably indicate the presence of media opacities, for example, cataracts, that decrease optical clarity {therefore, increase the subject's risk of having an optical clarity problem, for example, cataracts).
|0143} The system can also be configured to detect or generate risk assessments for retinal pigment epithelium (RPE) features that depart from the norma] curvature of the eye (drusen, retinal pigment epithelial detachments). Such RPE features can be detected by fitting the detected RPE layer to a polynomial curve that mimics the expected curvature for the eye, and using a computer algorithm to analyze, compare, or examine the difference between these curves. For example with respect to Figure 15, the system can be configured to subtract the polynomial curve that mimics the expected curvature of the RPE layer 1502 from the detected RPE layer curve 1504, and analyze and/or compare the resulting difference/value 1506 with the values (for example, in a histogram or nomogram) from normal and/or diseased eyes to generate a diagnosis or risk assessment. The foregoing method and process is similar to a measure of tortuosity in that a bumpy RPE detection will generally have more deviations from a polynomial curve than smooth RPE detections, which are common in young, healthy people.
[0144J Such RPE detection can also be used to detect increased transmission through the RPE which is essentially synonymous with RPE degeneration or atrophy. In certain embodiments, the system is configured to analyze the tissue layer beyond or beneath the RPE layer. Using imaging segmentation techniques, the RPE layer can be segmented. In certain embodiments, the system is configured to add up all of the intensity values beneath the RPE detection. When atrophy is present, there are generally many high values beneath the RPE line, which makes the integral value high and would increase the patient's risk of having a serious macular condition, such as geographic atrophy. [0145] With reference to Figure 16, the system can also be used to detect or generate risk factors for abnormal intensities within the retina. In certain embodiments, the system is configured to divide the retina into an inner 1602 and outer 1604 half based on the midpoint between the internal limiting membrane (ILM) detection 1606 and the RPE detection lines 1608. In some instances, a blur filter (for example, a Gaussian blur, radial blur, or the like) is applied to the retinal tissue to remove speckle noise and/or other noise. For each the inner and outer retina regions, a first derivative of the intensity values (with respect to position, for example, d/dx, d/dy, or the like) can be calculated to determine the slope of the curve to differentiate the areas where there are large changes from dark to bright or vice versa across lateral dimensions of the tissue. For example, intensities or derivatives within the retina can be compared to, for example, normal histograms, wherein inner retinal hypointensity can be an indicator of cystoid macular edema; or wherein outer retinal hypointensity can be an indicator of cystoid macular edema, subretinal fluid, or diffuse macular edema; or wherein outer retinal hyperintensity can be an indication of diabetes (which may be the cause of diabetic retinopathy, or damage to the retina due to, for example, complications of diabetes mellitus), or age-related macular degeneration.
[0146] Data from normal patients can used to compile histograms of intensity and/or slope (derivative) data to indicate expected values for normal people. Data from people with various diseases can also be placed into histograms of intensity and/or derivative (slope) values to indicate expected values for those people with diseases, hi certain embodiments, a relative risk will then be developed for each entry on the histogram such that this risk can be applied to unknown cases. For example, in some instances, people with 10% of their outer retinal intensity values equal to 0 have an 85% chance of having a retinal problem. Accordingly, such users may receive a health grade of 15. In another example, people with any inner retinal points less than 10 have a 100% chance of disease, and therefore such users may receive a health grade of 5.
[0147] Alternatively, as discussed herein, the foregoing method or process can also be used to determine or generate a risk assessment of glaucoma based on patterns of thinning of the macular and/or peripapillary nerve fiber layer or enlarged cupping of the optic nerve head as compared to a database of normal and abnormal values stored in the disease risk assessment / diagnosis database 708. Similarly, to detect or develop a risk assessment for uveitis, a histogram of expected intensity values above the inner retinal surface (in the vitreous), for example, can be used. The presence of large, bright specks (for example, high intensity areas) in the vitreous cavity would indicate possible uveitis and would likely indicate a need for referral. The foregoing method and process can also be used to determine or generate a risk of eye disease based on the intensity levels of the image signal as compared to a database of normal and abnormal values stored in the disease risk assessment / diagnosis database 708.
[0148] In other embodiments, the foregoing method and process can also be used to determine or generate a risk assessment of uveitis based on hyperreflective features in the vitreous cavity as compared to normal and abnormal hyperreflective features stored in the disease risk assessment / diagnosis database 708. The foregoing method and process can also be used to determine or generate a risk assessment of anterior eye disease based on detection of pathognomonic disease features, such as cystoid retinal degeneration, outer retinal edema, subretinal fluid, subretinal tissue, macular holes, drusen, retinal pigment epithelial detachments, and/or retinal pigment epithelial atrophy, wherein the detected features are compared with such pathognomonic disease features stored in the disease risk assessment / diagnosis database 708. In certain embodiments, the system is configured to perform template matching wherein the system detects, compares, and/or matches characteristics from A-scans generated from scanning a user, also known as unknown A-scans, with a database of patterns known to be associated with disease features, such as subretinal fluid, or the like.
[0149] With reference to Figures 1. 8 and 9, the optical coherence tomography system 100 is configured to allow the user to self-administer an OCT scan of the user's eyes without dilation of the eyes, and obtain a risk assessment or diagnosis of various diseases and ailments without the engaging or involving a doctor and/or technician to align the user's eyes with the system, administer the OCT scan and/or interpret the data from the scan to generate or determine a risk assessment or diagnosis. In one embodiment, the optical coherence tomography system 100 can perform a screening in less than two minutes, between 2-3 minutes, or 2-5 minutes. In certain embodiments, the use of the binocular system allows the user to self-align the optical coherence tomography system 100. The optical coherence system 100 with a binocular system is faster since it scans both eyes without repositioning and can allow the optical coherence tomography system 100 to scan a person's bad eye because the person's bad eye will follow the person's good eye as the latter tracks the fixation marker. Accordingly, the optical coherence tomography system 100 reduces the expense of conducting an OCT scan, thereby making OCT scanning more accessible to more people and/or users, and saving millions of people from losing their eye sight due to eye diseases or ailments that are preventable through earlier detection. In one embodiment, the optical coherence tomography system 100 is configured to have a small-foot print and/or to be portable, such that the optical coherence tomography system 100 can be installed or placed in drug stores, retail malls or stores, medical imaging facilities, grocery stores, libraries, and/or mobile vehicles, buses, or vans, a general practitioner's or other doctor's office, such that the optical coherence tomography system 100 can be used by people who do not have access to a doctor.
[0150] Additional features may be added to the optical coherence tomography system 100. hi some instances, the additional features may enhance performance of the system 100.
(0151] Figures 17A-C show B-scans obtained when the OCT system is positioned too far anterior, at a position that provides increased field of view, or too far posterior with respect to the eye. As shown, when the OCT system is too far anterior or too far posterior with respect to the eye the field of view (here the size or width of the B-scan) is reduced.
[0152] Figures 18A-C further show how a field of view of the system 100 can be affected by the location of the OCT system with respect to the eye. Figures 18A-C each show two probe beams 2005a and 2005b emitted from an optical coherence tomography system along different trajectories, as shown, by, for example, rotating a galvanometer 280 to probe different portions of the retina. For example, rotation of the galvanometer 280 may cause light to be emitted along different trajectories as described above. The trajectories may intersect with each other at rotation point 2010. Movement, e.g., rotation, of the galvanometer 280, may cause the trajectory of the probe beam 2005 to rotate about the rotation point 2010. Typically, a plurality of beams 2005 will be emitted by the system, such that the eye tissue can be sufficiently imaged. Thus, in some embodiments, numerous other beams are emitted between beams 2005a and 2005b. The beams are shown to intersect with each other at a rotation point or common point 2010. In some embodiments, the location of this point may coincide with a focus of the beams. Each of the beams 2005a and 2005b and the beams therebetween (not shown) can cause structures of the eye to reflect light, such that A-scan data can be beams associated with each beam. Figures 18A-C show a region 2015 that can be imaged by the plurality of beams. Thus, the emitted light may sweep across a swath of points of the retina. The position of the rotation point 2010 may influence the lateral dimension (e.g., length or width) of this region 2015. The region 2015 may be described as a field of view and may be correlated with the amount of data within a B-scan or set of A-scans that is above a threshold intensity.
[0153) hi Figure 18A, the rotation point 2010 is located behind/ posterior to the pupil 2030. Light beams 2020a and 2020b incident at high incident angles will therefore be unable to enter the eye, as the will be blocked by the iris 2025. The angle of incidence and therefore the region 2015 of the eye that can be imaged are limited in this situation.
|0154] In Figure 18B, the rotation point 2010 is located at a pupil plane at the pupil 2030 (e.g., in the plane of the pupil). Because the light beams intersect at the rotation point 2010, no incident light will be blocked by the iris 2025. Therefore, the region 2015 of the eye that can be imaged is not limited to obstruction by the iris as shown. A larger field of view is thereby provided.
[0155] In Figure 18C, the rotation point 2010 is located in front of/ anterior to the pupil 2030. As in Figure 18A, light beams 2020a and 2020b incident at high incident angles will therefore be unable to enter the eye, as the will be blocked by the iris 2025. The angle of incidence and therefore the region 2015 of the eye that can be imaged are limited in this situation. Accordingly, the regions 2015 probed in Figures 18A and 18C are shown reduced in comparison to the region 2015 probed in Figure 18B.
[0156] Referring again to Figures 17A-C, examples of how B-scans can be affected by the position of one or more movable components are shown. When the rotation point 2010 is too far anterior (Figure 17A) or too far posterior (Figure 17C), less tissue is imaged than if the rotation point 2010 is positioned at a more optimal position (Figure 17B). In each case, light from the center of the eye is reflected back towards the OCT system. However, when the rotation point 2010 is at a non-optimal location, light from the more extreme positions of the eye is not reflected back towards the OCT system. It is theorized that this light is instead scattered by the iris before it ever enters the eye, as illustrated in Figures 18A-C. By analyzing the resultant B-scans obtained for different positions of one or more movable components, it may thus be possible to determine a position that improves the field of view and thus the imaging capabilities of the OCT system. A risk assessment or diagnosis may then (e.g., automatically) be performed by the OCT system using an improved field of view, the improved field of view being obtained when the movable components are in a first position, and the improved field of view being larger than a field of view obtained from when the movable components are in a different second position.
[0157] Accordingly, it can be advantageous in some embodiments to position the intersection/rotation point 2010 at a specific location of the eye to, for example, improve the field of view and/or to reduce obstruction of incident light by the iris 2025. The location may comprise, for example, a position in or near the pupil of the user eye, a location in a plane of the iris of the user eye, a location within the lens of the user eye, or a location posterior to the pupil of the user eye. Other locations are possible. Additionally, certain embodiments may not include a well-defined intersection/rotation point 2010 at all. In some embodiments, OCT system 100 is configured to adjust an anterior-posterior distance of the OCT system with respect to the eye or a working distance of the OCT system. Figure 19 shows at least one movable optical component (e.g., lens 205) of the optical coherence tomography system 100. The position of the at least one movable optical component can at least partly determine the position of the rotation point 2010 and the working distance of the OCT system. In certain embodiments, for example, the working distance may at least partly determine the position of the rotation point 2010. The working distance 3005 may be measured, for example, as the distance between the outermost lens or window of the eyepiece and the rotation point 2010 or a position of the eyecup 120 and the rotation point 2010. (Other reference locations on the OCT system 100 can be used.) Thus, increasing the working distance can move the rotation point 2010 further anterior. In some embodiments, the OCT system may be moved with respect to the eye. As changing either the position of the at least one movable component, the working distance of the OCT system 100, or of the OCT system itself can change the position of the rotation point 2010 with respect to the eye, these changes may affect an angular field of view of the retina 3010, for reasons described above in relation to Figures 18A-C.
[0158] As described above, it may be desirable to position the intersection point or rotation point 2010, or another region of the emitted probe beam in order to reduce such blocking. In one instance, a field of view (a size of a set of A-scans, a B-scan or a region of the eye that can be imaged) is monitored as the galvanometer 280 is rotated. A moveable or adjustable optical component, such as one or more of the lens 205, adjustable optics 210, eyecup 120, and the eyepiece 203, may be moved to adjust a working distance, the location of the eyepiece 203 and/or OCT system 100 (in whole or part) with respect to the eye, which may at least partly control the field of view. The adjustment may change a rotation point to, for example, position the rotation point in or near the plane of the pupil. The adjustment may allow more light (e.g., a wider range of probe beam trajectories) to enter the eye from the system than would otherwise occur, thereby increasing a field of view. For example, the adjustment may increase the number of probe beam orientations that can enter the eye across a B-scan by reducing the light blocked by one or more structures of the eye (e.g., the iris).
[0159] Translation stages and other actuators or movement devices may be employed to position the eye or the optics of the OCT system in the anterior-posterior direction or otherwise adjusted to provide movement in a longitudinal direction along the optical axis of the OCT instrument. Thus, the position (e.g., longitudinally along the optical axis of the OCT instrument) of the movable component may determine the anterior-posterior position of the rotation point.
[0160] In certain embodiments, a translation stage such as a stage configured to move laterally (e.g. horizontally) may be included. Such a translation stage or actuator may determine the horizontal position of the rotation point. For example, if the stage was positioned too far medial or lateral, the iris may block a portion of the light from entering the eye that would be used to form a medial or lateral portion of, for example, a B-scan. In some instances, the iris may block a medial portion of the scan if the stage is too far medial, while in others, it may block a lateral portion. Accordingly, the field of view (e.g., B-scans) for the left and right eye can be compared. If one is smaller than the other, the translation stage for the eye with the smaller field of view can be adjusted to increase the field of view of that eye.
[0161] Accordingly, in some instances, a B-scan or other OCT measurement may be analyzed or a plurality of scans or measurements are compared to determine whether a lateral (e.g., horizontal) movement of the stage or actuator is advantageous. The stage or actuator may be adjusted for example after a user-conducted interpupillary distance alignment process using for example a fixation target, such as that described above. Additional alignment may be performed subsequent to adjustment of the stage or actuator which may affect interpupillary distance. Moreover, movement of the stage or actuator may affect the position of the rotation point or the portion of the sample being imaged and additional modifications of the positions of the system components may be made to account for this effect. In various embodiments, movement of the stage may move one or more of the components of the OCT system. For example, in certain embodiments, the stage may support lens 205, adjustable optics 210, beamsplitter 230 and/or mirror 260. One such stage may be included for each of the eyes. In some embodiments, the components supported by the stage are those such that stage movement does not affect the angle at which the beam is output from the device. In some instances, one or more movements or adjustments (e.g., of a horizontal stage) may be asymmetric across the two eyes, such that a movement associated with one eye is unparalleled or is different than a movement associated with the other eye.
|01621 In order to determine an appropriate adjustment, the one or more of the movable or adjustable optical components may be positioned (e.g., systematically) in a plurality of positions, and data (e.g., optical coherence data) may be obtained at these positions. The one or more movable/adjustable optical components may then be adjusted to be positioned at a desired position, the desired position being based on a comparison the optical coherence data obtained at each of the positions. The image data, e.g., B-scan, may be obtained at the desired position.
[0163) In one instance, one or more sets of A-scans or one or more B-scans are analyzed to determine a position of one or more movable/adjustable components. Each of the B-scans or the sets of A-scans can be associated with a distinct position/setting of the one or more movable/adjustable components. A property of the scans (e.g.. an image quality measure or signal intensity value) may be compared across the B-scans or sets of A-scans in order to, for example, determine a preferred position or setting of the one or more movable/adjustable components. In one instance, the sum total of the integrated intensities across the B-scans or sets of A-scans are compared. In another instance, the intensities (e.g. sum total of integrated intensities) at a particular point or location within the A-scans comprising a B-scan or set of A-scans is used in the comparison. For example, a variable may be defined as the sum of the intensity at the approximate location of the retina across all A-scans within a B-scan or set of A-scans. This variable may then be compared across sets of A-scans or B-scans. A resultant position/setting of the one or more movable/adjustable components may be defined as the position/setting with a set of A-scans or a B-scan having a value for the variable that is above a threshold or is maximum (e.g., greatest total intensity). Other values may be measured, calculated or considered and other approaches may be used to determine the desired position/setting and thereby increase the field of view.
|0164] hi some instances, a plurality (e.g., a predetermined number) of B-scans or sets of A-scans are obtained and a preferred position/setting of the one or more movable/adjustable components is determined as a position/setting associated with one of the B-scans or sets of A-scans. In another instance, the data is used to predict a preferred position/setting that may or may not be a position/setting associated with the collected data. For example, extrapolation or interpolation may be employed, hi some instances, the B-scan or A-scan set data is dynamically collected. For example, if a shift of the one or more movable components along an axis from a first position to a second position caused a preferable change in a variable, then subsequent movements may avoid drastic changes in the opposite direction. In another example, the one or more movable components may repeatedly be adjusted until a variable crosses a threshold. Other approaches and methods may be used.
|0165] It may be desirable to position or set the one or more movable or adjustable components such that a rotation point of the probe beams emitted from the optical coherence tomography system are at or close to the pupil plane. If the probe beams are rotated around a position not at the pupil plane but instead shifted longitudinally towards the retina, then some of the light emitted from the optical coherence tomography system may be blocked by, for example, the iris before reaching the focal point. If the probe beam is rotated around a position not at the pupil but instead shifted towards the cornea, then some of the light emitted from the optical coherence tomography system may be blocked by, for example, the iris after reaching the focal point but before reaching other ocular structures such as the retina. By rotating light beams emitted by the OCT system around a point at the pupil, ocular structures, such as the iris, which surround the pupil might not block input light. Thus, rotating the probe beams at the pupil may increase or maximize the amount of tissue that may be imaged by the instrument.
[0166] In some instances, a position of the retina is determined for a plurality of B-scans or sets of A-scans, each associated with a different position or setting of the one or more movable or adjustable components. The position of the retina may then be used to predict the position of the pupil, and the movable/adjustable components may be positioned/set such that a rotation point is at or near the predicted pupil position. In some instances, a position of an structure of the eye, such as the cornea, iris, retina, vitreous, anterior chamber, or tear film interface, or from another anatomical feature, such as the orbital rim, nasal bridge, cheekbone (maxilla), frontal bone, eyelid or skin surface, is determined and the movable/adjustable components are positioned/set based on the determined location. The determined location may be used to predict a location of another structure, such as that of the pupil.
[0167] in some instances, a desire position or setting of the one or more movable or adjustable components is not based on optical coherence tomography data obtained for a specific patient. For example, the position or setting may be selected based on normative data which may comprise, for example, a normative position or setting that is determined based on population-based measurements or data. For example, for each of a plurality of patients, field of view measurements may be made for each of a plurality of positions or settings of the one or more movable/adjustable components. In a first instance, a preferred position or setting is determined for each patient. A normative position/setting may be, for example, a mean, median or mode of the preferred positions/setting across patients, hi a second instance, the measurements are compared to a threshold for all patients. The normative position or setting may then be determined as a position or setting for which, for example, the measurements exceeded this threshold across the most patients. In some embodiments, the movable or adjustable components are fixed at a normative position or setting. The movable/adjustable component may be fixed in the same system or in different systems at the position or setting determined based on normative data. This position/setting may be used as the selected position/setting or may be used as a starting point for measuring different fields of view for different positions/setting as described above to determine the position/setting having an increased field of view.
[0168] In some embodiments, the normative position/setting is not determined based on optical coherence tomography data but is instead based on anatomical data otherwise obtained. For example, the normative position may be determined based on an average distance between a pupil and a retina, an average anterior-posterior distance between an eye socket and a pupil, an average anterior-posterior distance between the cornea and the pupil, or an average anterior-posterior distance between a chin and a pupil. The normative position/setting may be separately determined for different patient groups. For example, the normative position/setting may be based on a person's age, gender or race.
(0169] In some instances, the position/setting of the one or more movable/adjustable components may be based at least partly on sensor data. For example, a sensor may detect a position of the patient or a patient feature (e.g., an eye, a pupil, an iris, a chin, an eye socket), and this position may be used to determine the position or setting of the eyepiece 203 or OCT system 100. In one instance, the detected position is used to predict the position of the pupil, which is used to determine the position of the one or more movable components.
[0170] Accordingly, in some embodiments, an optical coherence tomography system (e.g., that of Figure 1 or 3) comprises a sensor or tracker. The sensor or tracker may determine a position of the user, one or two eyes of the user, and/or one or more structures (e.g., a retina, pupil, cornea, or lens) of the user's eye. In some embodiments, the sensor or tracker is positioned on or attached to main body 106, zero gravity arm 116, or even eyecup 120. In some embodiments, the sensor or tracker is a device separate from the main body 106. In some embodiments, the sensor or tracker is attached to or comprised within the system shown in Figure 3. [0171] In one instance, the sensor emits light or ultrasound from a light source and detects light reflected back. The light may be reflected back from a structure of the user's eye, such as the cornea, iris, pupil, retina, vitreous, anterior chamber, or tear film interface, or from another anatomical feature, such as the orbital rim, nasal bridge, cheekbone (maxilla), frontal bone, eyelid or skin surface. The sensor may determine the position of the structure based on the time difference between the time the light (e.g., a pulse) was emitted and the time the light was detected. In other instances, other types of sensors or trackers may be used. For example, an optical coherence tomography instrument may determine the position of an eye structure based on interference or reflectance results.
[0172] In some embodiments, the position/setting of the one or more movable/adjustable components is based on a combination of approaches. For example, the position/setting may be determined based on non-optical coherence tomography sensor data and optical coherence tomography data. The position/setting may be determined based on field-of-view data and sensor data and/or a determined normative position. The position may be determined based on sensor data and a determined nonnative position. In certain embodiments, at least one of normative data or sensor data may be used to assist in determining a starting point for multiple OCT measurements that are subsequently employed to determine a position or setting which provides a further increased field of view.
[0173] Other approaches are possible. In some embodiments, for example, an optical coherence tomography system 100 comprises a chin rest. In such instances, the system 100 may be configured to automatically adjust or to allow for manual adjustment between the main body (and/or the eyepiece) and the patient's eyes. The adjustment may be fine, on the order of about 0.5, 1, 2, 3, 4, 5, 10, 20, 30 or 50 millimeters. The adjustment may comprise any adjustment described herein, such as an adjustment of one or more moveable optical components to, for example, improve a field of view of the system 100. In one instance, the distance between the main body and/or an optical component of in the main body and the patient's eye is systematically adjusted from a first distance to a second distance. The chin rest may move in certain embodiments although in other embodiments the chin rest may be fixed. The distance may be based at least partly on nonnative values, such as an average offset (e.g., in the anterior-posterior direction) between a chin and a pupil or an average distance between a pupil and an eyecup. In some instances, the distance is detennined based at least partly on a sensor reading. For example, a sensor may detect a position of the user's eye, pupil or iris. The sensor may comprise an optical tomography instrument or may comprise another optical or ultrasonic instrument. For example, as described above, the sensor may emit a light and determine the time elapsed between the emission and that at which reflected light (e.g., a pulse) is received. The sensor may comprise a weight sensor to sense, for example, a location of the patient's chin. A sensor may detect a position or weight of the user's chin. In certain embodiments the chin rest may move or the main body and/or eyepiece of the OCT system may move with respect to the chin rest and the field of view monitored as described above to determine a suitable location of the eye. Other variations are possible.
[0174] In some embodiments, a position/setting of one or more moveable/adjustable optical components can be manually adjusted by the patient. The patient may be instructed, for example, to adjust the position/setting based on one or more images seen by the patient. For example, the patient may be instructed to adjust the position until two or more images (e.g., working distance images) are aligned. Alignment may correspond to an appropriate distance of the eye to the OCT instrument. Other designs are also possible.
10175] All of the methods and processes described above may be embodied in, and fully automated via, software code modules executed by one or more general purpose computers or processors. The code modules may be stored in any type of computer-readable medium or other computer storage device. Some or all of the methods may alternatively be embodied in specialized computer hardware.
|0176] While the invention has been discussed in terms of certain embodiments, it should be appreciated that the invention is not so limited. The embodiments are explained herein by way of example, and there are numerous modifications, variations and other embodiments that may be employed that would still be within the scope of the present invention. Components can be added, removed, and/or rearranged. Additionally, processing steps may be added, removed, or reordered. A wide variety of designs and approaches are possible. |0177] For purposes of this disclosure, certain aspects, advantages, and novel features of the invention are described herein. It is to be understood that not necessarily all such advantages may be achieved in accordance with any particular embodiment of the invention. Thus, for example, those skilled in the art will recognize that the invention may be embodied or carried out in a manner that achieves one advantage or group of advantages as taught herein without necessarily achieving other advantages as may be taught or suggested herein.

Claims

WHAT IS CLAIMED IS:
1. An optical coherence tomography instrument for providing a self-administered optical coherence tomography test comprising: an eyepiece for receiving at least one eye of a user; a light source that outputs a probe beam that is directed through the eyepiece into the user's eye; an interferometer configured to produce optical interference using light reflected from the user's eye; an optical detector disposed so as to detect said optical interference; a processor configured to analyze optical coherence tomography measurements obtained using said interferometer; and at least one adjustable optical component that may be adjusted to change the field of view of the optical coherence tomography instrument; wherein said processor is configured to monitor a field of view of the optical coherence tomography instrument and direct adjustment of the at least one adjustable optical component based on said monitored field of view.
2. The optical coherence tomography instrument of Claim 1, further comprising a rotational unit configured to rotate the trajectory of the probe beam into the user's eye.
3. The optical coherence tomography instrument of Claim 2, wherein the rotational unit comprises a galvanometer.
4. The optical coherence tomography instrument of Claim 2, wherein the rotational unit causes the probe beam to propagate along a plurality of different trajectories that intersect at a common location and said at least one adjustable optical component is adjustable so as to place said common location in or near the pupil of the user eye.
5. The optical coherence tomography instrument of Claim 1, wherein the position of the optical coherence tomography instrument is adjustable along an anterior- posterior axis of the user so as to adjust the position of said at least one adjustable optical component so as to adjust the field of view.
6. The optical coherence tomography instrument of Claim 1 , wherein said at least one adjustable optical component has a position that is adjustable along an optical axis of said optical coherence tomography instrument so as to adjust said field of view.
7. The optical coherence tomography instrument of Claim 1 , wherein said optical coherence tomography instrument has a working distance and adjustment of said at least one optical adjustable component alters the work distance of said optical coherence tomography instrument.
8. The optical coherence tomography of Claim 1 , wherein the at least one optical component comprises at least one lens.
9. The optical coherence tomography instrument of Claim 1, wherein the processor is configured to compare a plurality of sets of A-scans, each set being associated with a different state of the adjustable optical component.
10. The optical coherence tomography instrument of Claim 9, wherein the processor is configured to compare the field of view of different A-scan sets.
1 1. The optical tomography instrument of Claim 9, wherein the processor is configured to determine a setting of the adjustable optical component selected based on monitored field of view data associated with each of the number of settings of the adjustable optical component.
12. The optical tomography instrument of Claim 1 , wherein the processor is configured to determine a position of the adjustable optical component that increases the portion of the user's eye that can be imaged by the optical coherence tomography instrument.
13. The optical coherence tomography of Claim 1 , further comprising an output device configured to output a result of the analysis to the user.
14. The optical tomography instrument of Claim 1 , further comprising a sensor for determining a position of a structure of the user's eye, said processor being configured to control the position of the adjustable optical component based at least in part on the determined position of structure of the eye.
15. The optical coherence tomography instrument of Claim 14, wherein the structure comprises the pupil, the iris, the cornea, the anterior chamber, the lens, the vitreous or the retina.
16. A method for providing a self-administered optical coherence tomography test of a user's eye, the method comprising: automatically evaluating a plurality of fields of view of the optical coherence tomography instrument, each field of view corresponding to a different state of at least one optical component of said optical coherence tomography instrument; automatically adjusting said at least one optical component at least partly based on said evaluation; and with said at least one optical component adjusted, performing the optical coherence tomography test, wherein the optical coherence tomography instrument is configured to scan the at least one user eye.
37. The method of Claim 16, wherein automatically evaluating a plurality of fields of view of the optical coherence tomography instrument comprises rotating a galvanometer to direct a probe beam along different trajectories into the user eye.
18. The method of Claim 17, wherein said trajectories intersect at a common location.
19. The method of Claim 18, wherein said common location comprises at least one of a location in or near the pupil of the user eye, a location in a plane of the iris of the user eye, a location within the lens of the user eye, and a location posterior to the pupil of the user eye.
20. The method of Claim 16, wherein said adjusting said at least one optical component comprises positioning the optical coherence tomography instrument along an anterior-posterior axis of the user.
23. The method of Claim 16, wherein said adjusting said at least one optical component comprises positioning said at least one optical component along an optical axis of said optical coherence tomography instrument.
22. The method of Claim 16, wherein said optical coherence tomography instrument has a working distance and said adjusting said at least one optical component alters the working distance of said optical coherence tomography instrument.
23. The method of Claim 16, wherein the at least one optical component comprises at Jeast one lens.
24. The method of Claim 16, wherein said evaluating the fields of view comprises collecting A-scan data.
25. The method of Claim 16, wherein said evaluating the fields of view comprises collecting B-scan data.
26. The method of Claim 16, wherein the fields of view comprise angular fields of view.
27. The method of Claim 16, further comprising providing the user with an output of the optical coherence tomography test from an output device.
28. The method of Claim 16, further comprising: receiving the user eye via an eyepiece of the optical coherence tomography instrument; directing light from a source through the eyepiece into the user eye; producing optical interference using light reflected from the user eye; and detecting optical interference.
29. The method of Claim 28, further comprising: automatically performing a diagnosis or risk assessment based on said detected optical interference, wherein said output comprises the diagnosis or risk assessment.
30. The method of Claim 28, further comprising: identifying one or more diseases or abnormal findings that the user may manifest indications of based on said detected optical interference, wherein said output comprises a notification said identified diseases or abnormal findings.
31. A method for providing a self-administered optical coherence tomography test of a user's eye, the method comprising: determining a longitudinal position of the user relative to an optical coherence tomography instrument; automatically adjusting at least one optical component of said optical coherence tomography instrument at least partly based on said determined position thereby increasing the field of view of the optical coherence tomography instrument; and with said at least one optical component adjusted, performing the optical coherence tomography test, wherein the optical coherence tomography instrument is configured to scan the at least one user eye.
32. The method of Claim 31. wherein said determining a position of the user comprises detecting a portion of the user based on a sensor measurement.
33. The method of Claim 31, wherein determining a position of the user comprises determining the position of the user's eye.
34. The method of Claim 31, wherein determining a position of the user comprises radiating a signal to the user and detecting the signal returned from the user.
35. The method of Claim 34, wherein said signal comprises light or ultrasound.
36. The method of Claim 34, wherein said signal is reflected off of at least one of a cornea, an iris, a retina, a vitreous, an anterior chamber, and a tear film interface of the user.
37. The method of Claim 34, wherein said signal is reflected off of at least one of an orbital rim, a nasal bridge, a cheekbone (maxilla), a frontal bone, an eyelid and a skin surface.
38. The method of Claim 31, wherein determining a position of the user comprises determining the position of the user's eye cornea or tear film interface.
39. The method of Claim 31, wherein determining a position of the user comprises determining the position of the user's eye pupil or iris.
40. The method of Claim 31, wherein determining a position of the user comprises determining the position of the user's retina.
41. The method of Claim 31, wherein said determining a position of the user comprises analyzing optical coherence tomography data obtained by the optical coherence tomography instrument.
42. The method of Claim 31, further comprising automatically evaluating a plurality of fields of view of the optical coherence tomography instrument, each field of view corresponding to a different state of at least one optical component of said optical coherence tomography instrument, said automatically adjusting said at least one optical component being at least partly based on said evaluation.
43. The method of Claim 31, wherein the optical coherence tomography instrument comprises rotating a galvanometer that directs a probe beam along different trajectories into the user eye, said trajectories intersecting at a common location.
44. The method of Claim 43, wherein said automatically adjusting said at least one optical component positions said common location in a pupil of the user eye.
45. The method of Claim 31, wherein said adjusting said at least one optical component comprises positioning the optical coherence tomography instrument along an anterior-posterior axis of the user.
46. The method of Claim 31 , wherein said adjusting said at least one optical component comprises positioning said at least one optical component along an optical axis of said optical coherence tomography instrument.
47. The method of Claim 31, wherein said optical coherence tomography instrument has a working distance and said adjusting said at least one optical component alters the work distance of said optical coherence tomography instrument.
48. The method of Claim 31, wherein the at least one optical component comprises at least one lens.
49. The method of Claim 31, further comprising providing the user with an output of the optica] coherence tomography test from an output device.
50. An optical coherence tomography instrument for providing a self-administered optical coherence tomography test comprising: an eyepiece for receiving at least one eye of a user; a sensor for determining a position of a structure of the user eye with respect to the optical coherence tomography instrument; a light source that outputs light that is directed through the eyepiece into the user's eye; an interferometer configured to produce optical interference using light reflected from the user's eye; an optical detector disposed so as to detect said optical interference; and at least one adjustable optical component that may be adjusted to change the field of view of the optical coherence tomography instrument; and a processor configured to analyze optical coherence tomography measurements obtained using said interferometer, wherein the processor is further configured to direct adjustment of the at least one adjustable optical component based at least in part on said position.
51. The optical coherence tomography instrument of Claim 50, wherein said sensor comprises an optica] sensor or an ultrasound sensor.
52. The optical coherence tomography instrument of Claim 51, further comprising an ultrasound emitter.
53. The optical coherence tomography instrument of Claim 51, further comprising a light emitter that directs an optical signal to said eye to be detected by said sensor.
54. The optical coherence tomography instrument of Claim 50, wherein the structure comprises a pupil, iris or lens.
55. The optical coherence tomography instrument of Claim 50, wherein the structure comprises a cornea or anterior chamber.
56. The optical coherence tomography instrument of Claim 50, wherein the structure comprises a retina or the vitreous.
57. The optical coherence tomography of Claim 50, wherein the at least one adjustable optical component comprises at least one lens.
58. The optical coherence tomography of Claim 50, wherein the at least one adjustable optical component comprises an eyepiece.
59. The optical coherence tomography of Claim 50, further comprising an output device configured to output a result of the analysis to said user.
60. An optica] coherence tomography instrument for providing a self-administered optical coherence tomography test comprising: an eyepiece for receiving at least one eye of a user; a chin rest for receiving a chin of the user, said chin rest movable with respect to said eyepiece to alter the longitudinal distance therebetween; a light source that outputs light that is directed through the eyepiece into the user s eye; an interferometer configured to produce optical interference using light reflected from the user's eye; an optical detector disposed so as to detect said optical interference; a processor configured to analyze optical coherence tomography measurements obtained using said interferometer, wherein said processor is further configured to control the position of the chin rest relative to the eyepiece so as to increase the field of view of the optical coherence tomography instrument.
63. The optical coherence tomography instrument of Claim 60, wherein the processor is configured to position the chin rest with respect to the eyepiece based at least partly on a normative position based on across a patient population.
62. The optical coherence tomography instrument of Claim 60, wherein said processor is configured to monitor a field of view of the optical coherence tomography instrument and position the chin rest with respect to the eyepiece based at least partly on said monitored field of view.
63. The optical coherence tomography instrument of Claim 60, wherein the processor is configured to compare a plurality of sets of A-scans, each data set being associated with a different position of the chin rest relative to the eyepiece.
64. The optical coherence tomography instrument of Claim 63, wherein the processor is configured to position the chin rest relative to the eyepiece at least partly based on a comparison of a plurality of fields of view from each of the plurality of A-scan data sets.
65. The optical coherence tomography instrument of Claim 60, further comprising a sensor to sense the user or part of the user.
66. The optical coherence tomography instrument of Claim 65, wherein the sensor is configured to sense a position of the user's eye or structure of the user's eye.
67. The optical coherence tomography instrument of Claim 65, wherein the controller is configured to position the chin rest relative to the eyepiece based at least partly on data collected by the sensor.
68. The optical coherence tomography instrument of Claim 60, further comprising a movable horizontal stage supporting at least one component of the optical coherence tomography instrument sensor, the stage being configured to adjust a medial-lateral position of the light being directed through the eyepiece into the user's eye.
PCT/US2009/037448 2008-03-27 2009-03-17 Optical coherence tomography device, method, and system WO2009120543A1 (en)

Applications Claiming Priority (8)

Application Number Priority Date Filing Date Title
US4008608P 2008-03-27 2008-03-27
US4008408P 2008-03-27 2008-03-27
US61/040,086 2008-03-27
US61/040,084 2008-03-27
US12/111,894 2008-04-29
US12/111,894 US8348429B2 (en) 2008-03-27 2008-04-29 Optical coherence tomography device, method, and system
US8216708P 2008-07-18 2008-07-18
US61/082,167 2008-07-18

Publications (1)

Publication Number Publication Date
WO2009120543A1 true WO2009120543A1 (en) 2009-10-01

Family

ID=40637957

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2009/037448 WO2009120543A1 (en) 2008-03-27 2009-03-17 Optical coherence tomography device, method, and system

Country Status (2)

Country Link
TW (1) TW200940026A (en)
WO (1) WO2009120543A1 (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2485345A (en) * 2010-11-05 2012-05-16 Queen Mary & Westfield College Optical coherence tomography scanning to identify a region of interest in a sample
EP2478830A3 (en) * 2011-01-20 2012-11-14 Canon Kabushiki Kaisha Optical coherence tomographic imaging method and apparatus
WO2012146711A3 (en) * 2011-04-27 2013-01-17 Carl Zeiss Meditec Ag Systems and methods for improved ophthalmic imaging
US9149182B2 (en) 2008-03-27 2015-10-06 Doheny Eye Institute Optical coherence tomography device, method, and system
US9492079B2 (en) 2008-07-18 2016-11-15 Doheny Eye Institute Optical coherence tomography-based ophthalmic testing methods, devices and systems
EP2983604A4 (en) * 2013-04-11 2016-12-28 Novartis Ag Method and system to detect ophthalmic tissue structure and pathologies
US9907466B2 (en) 2014-05-08 2018-03-06 Mimo Ag Optical coherence tomography imaging device for imaging a retina of a human subject
CN113168668A (en) * 2018-12-04 2021-07-23 哈格-斯特莱特有限公司 Method and measuring system for performing a movement correction during the measurement of an eye
WO2022053387A1 (en) * 2020-09-10 2022-03-17 Ivis Technologies S.R.L. Medical device for automatic exam acquisition
US11510567B2 (en) 2008-03-27 2022-11-29 Doheny Eye Institute Optical coherence tomography-based ophthalmic testing methods, devices and systems

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA2817963A1 (en) * 2010-11-17 2012-05-24 Optovue, Inc. 3d retinal disruptions detection using optical coherence tomography
US10366269B2 (en) * 2016-05-06 2019-07-30 Qualcomm Incorporated Biometric system with photoacoustic imaging
TWI647427B (en) * 2018-01-10 2019-01-11 緯創資通股份有限公司 Object distance estimation method and electronic device
TWI664582B (en) * 2018-11-28 2019-07-01 靜宜大學 Method, apparatus and system for cell detection
CN111812341A (en) * 2020-07-22 2020-10-23 英华达(上海)科技有限公司 Automatic equipment detection system and method for detecting internal operation of automatic equipment

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH11225958A (en) * 1998-02-18 1999-08-24 Koonan:Kk Chin rest device for ophthalmologic instrument and ophthalmogic instrument
US20040196432A1 (en) * 2000-06-13 2004-10-07 Wei Su Digital eye camera
US20070153233A1 (en) * 2005-12-31 2007-07-05 Campin John A Determining optimal positioning of ophthalmic devices by use of image processing and autofocusing techniques
EP1864608A1 (en) * 2006-06-09 2007-12-12 Kabushiki Kaisha TOPCON A fundus observation device, an ophthalmologic image processing unit, an ophthalmologic image processing method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH11225958A (en) * 1998-02-18 1999-08-24 Koonan:Kk Chin rest device for ophthalmologic instrument and ophthalmogic instrument
US20040196432A1 (en) * 2000-06-13 2004-10-07 Wei Su Digital eye camera
US20070153233A1 (en) * 2005-12-31 2007-07-05 Campin John A Determining optimal positioning of ophthalmic devices by use of image processing and autofocusing techniques
EP1864608A1 (en) * 2006-06-09 2007-12-12 Kabushiki Kaisha TOPCON A fundus observation device, an ophthalmologic image processing unit, an ophthalmologic image processing method

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11510567B2 (en) 2008-03-27 2022-11-29 Doheny Eye Institute Optical coherence tomography-based ophthalmic testing methods, devices and systems
US11291364B2 (en) 2008-03-27 2022-04-05 Doheny Eye Institute Optical coherence tomography device, method, and system
US10165941B2 (en) 2008-03-27 2019-01-01 Doheny Eye Institute Optical coherence tomography-based ophthalmic testing methods, devices and systems
US10945597B2 (en) 2008-03-27 2021-03-16 Doheny Eye Institute Optical coherence tomography-based ophthalmic testing methods, devices and systems
US9149182B2 (en) 2008-03-27 2015-10-06 Doheny Eye Institute Optical coherence tomography device, method, and system
US11839430B2 (en) 2008-03-27 2023-12-12 Doheny Eye Institute Optical coherence tomography-based ophthalmic testing methods, devices and systems
US9492079B2 (en) 2008-07-18 2016-11-15 Doheny Eye Institute Optical coherence tomography-based ophthalmic testing methods, devices and systems
GB2485345A (en) * 2010-11-05 2012-05-16 Queen Mary & Westfield College Optical coherence tomography scanning to identify a region of interest in a sample
US8517537B2 (en) 2011-01-20 2013-08-27 Canon Kabushiki Kaisha Optical coherence tomographic imaging method and optical coherence tomographic imaging apparatus
EP2478830A3 (en) * 2011-01-20 2012-11-14 Canon Kabushiki Kaisha Optical coherence tomographic imaging method and apparatus
US9918634B2 (en) 2011-04-27 2018-03-20 Carl Zeiss Meditec, Inc. Systems and methods for improved ophthalmic imaging
US9055892B2 (en) 2011-04-27 2015-06-16 Carl Zeiss Meditec, Inc. Systems and methods for improved ophthalmic imaging
WO2012146711A3 (en) * 2011-04-27 2013-01-17 Carl Zeiss Meditec Ag Systems and methods for improved ophthalmic imaging
US9955865B2 (en) 2013-04-11 2018-05-01 Novartis Ag Method and system to detect ophthalmic tissue structure and pathologies
AU2014251290B2 (en) * 2013-04-11 2019-04-11 Alcon Inc. Method and system to detect ophthalmic tissue structure and pathologies
EP2983604A4 (en) * 2013-04-11 2016-12-28 Novartis Ag Method and system to detect ophthalmic tissue structure and pathologies
US9907466B2 (en) 2014-05-08 2018-03-06 Mimo Ag Optical coherence tomography imaging device for imaging a retina of a human subject
CN113168668A (en) * 2018-12-04 2021-07-23 哈格-斯特莱特有限公司 Method and measuring system for performing a movement correction during the measurement of an eye
CN113168668B (en) * 2018-12-04 2022-11-25 哈格-斯特莱特有限公司 Method and measuring system for performing a movement correction during the measurement of an eye
WO2022053387A1 (en) * 2020-09-10 2022-03-17 Ivis Technologies S.R.L. Medical device for automatic exam acquisition

Also Published As

Publication number Publication date
TW200940026A (en) 2009-10-01

Similar Documents

Publication Publication Date Title
US11291364B2 (en) Optical coherence tomography device, method, and system
US10945597B2 (en) Optical coherence tomography-based ophthalmic testing methods, devices and systems
US11510567B2 (en) Optical coherence tomography-based ophthalmic testing methods, devices and systems
WO2009120543A1 (en) Optical coherence tomography device, method, and system
WO2010117386A1 (en) Ophthalmic testing methods, devices and systems
US11800977B2 (en) Apparatus and method for self-administration of optical scanning of a person's eye optical system
JP5512520B2 (en) Optical alignment apparatus and optical alignment method
TW201014571A (en) Optical coherence tomography device, method, and system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 09726148

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 09726148

Country of ref document: EP

Kind code of ref document: A1