WO2016123448A2 - Systems and method for mapping the ocular surface usually obstructed by the eyelids - Google Patents

Systems and method for mapping the ocular surface usually obstructed by the eyelids Download PDF

Info

Publication number
WO2016123448A2
WO2016123448A2 PCT/US2016/015579 US2016015579W WO2016123448A2 WO 2016123448 A2 WO2016123448 A2 WO 2016123448A2 US 2016015579 W US2016015579 W US 2016015579W WO 2016123448 A2 WO2016123448 A2 WO 2016123448A2
Authority
WO
WIPO (PCT)
Prior art keywords
eye
ocular surface
individual
images
pattern
Prior art date
Application number
PCT/US2016/015579
Other languages
French (fr)
Other versions
WO2016123448A3 (en
Inventor
Scott Paul CATANZARITI
Original Assignee
Catanzariti Scott Paul
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Catanzariti Scott Paul filed Critical Catanzariti Scott Paul
Publication of WO2016123448A2 publication Critical patent/WO2016123448A2/en
Publication of WO2016123448A3 publication Critical patent/WO2016123448A3/en

Links

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B3/00Apparatus for testing the eyes; Instruments for examining the eyes
    • A61B3/10Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions
    • A61B3/107Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for determining the shape or measuring the curvature of the cornea
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • G06T9/001Model-based coding, e.g. wire frame
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B3/00Apparatus for testing the eyes; Instruments for examining the eyes
    • A61B3/0008Apparatus for testing the eyes; Instruments for examining the eyes provided with illuminating means
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B3/00Apparatus for testing the eyes; Instruments for examining the eyes
    • A61B3/0091Fixation targets for viewing direction

Definitions

  • the present disclosure relates to three dimensional imaging using the structured light approach, and more particularly to ocular surface measurement over the anterior surfaces of the cornea and sclera.
  • Cornea is responsible for about 70% of the refractive power of the eye and therefore the corneal topography has a great importance in determining the quality of vision. It is commonly used for diagnosis of keratoconus, for selecting appropriate soft contact lenses, for fitting scleral lenses, and for topography guided Laser-Assisted in situ Keratomileusis (LAS IK).
  • LAS IK Laser-Assisted in situ Keratomileusis
  • Placido disks are not capable of measuring the scleral topography, which is important for custom fitting of scleral lenses.
  • Orbscan II by Bausch & Lomb Scheimpflug systems such as the Galilei by Zeimer Ophthalmic Systems, as well as rastersterographic systems such as the CTS by Par Technlogies and ESP by Eaglet Eye have been used for corneal topography. While these systems provide better measurement of the corneal apex they lack the coverage of the corneal periphery and sclera needed for scleral lens fitting.
  • the present disclosure describes an advanced measurement system for mapping the complete three dimensional anterior ocular surface.
  • Commonly used ocular topography measurements including Placido-disk measurements, scanning slit beam measurements, and rastersterographic measurements have focused on the central region of the cornea. Extending such measurements to the entire anterior surface, specifically to include the sclera, introduces a new set of challenges. The obstruction of the upper and lower sclera with eyelids is one of the major challenges in the mapping of the eye surface.
  • a sample embodiment of the technology disclosed herein provides a system and method for obtaining the corneal and scleral maps under the eyelids.
  • the surface of the eye is coated with a fluorescent dye and two or more projectors are configured to project light in a wavelength that overlaps the excitation band of the said dye.
  • the resulting fluorescent light emitted from the dye covering the ocular surfaces is detected using one or more imaging sensors.
  • the projected light may be comprised of a single pattern or a coded light sequence.
  • the pattern may be comprised of vertical or horizontal lines, a combination of vertical and horizontal lines, or an arbitrary pattern with known geometrical characteristics that can be used to determine the ocular shape.
  • the coded light sequence may contain series of structured light patterns which can be interpreted as time-series measurements, where the projected intensity at a given location over time has a unique pattern for each pixel or subregion of pixels, allowing accurate, unambiguous identification of triangulation points in each member of the stereo-photogrammetric or stereo-projection pair.
  • Said triangulation measurements performed in order to obtain the three dimensional surface of the eye can be performed between two cameras, between two projectors or between one of the cameras and the projector.
  • the abovementioned triangulation measurements can be performed separately or in any combination of the above in order to increase measurement redundancy and reduce the measurement errors.
  • Said coded-light sequences allow higher resolution measurements than conventional raster- stereographic methods by coding all pixels in the pattern region, instead of requiring interpolation between grid lines or grid intersection points.
  • Another embodiment may be comprised of a system for determining the topography of a curved surface, comprising a projection device means for projecting patterns of lines on the surface, the projection device means including two independent light projectors disposed at an angle relative to each other, each projector having a projection axis and including a grating having parallel straight lines, wherein each grating is positioned at a right angle to the projection axis of its respective projector, a rectangular diaphragm having the long sides parallel to the lines of the grating, a detection device means for registering an image formed on the surface and a flash light source synchronized in sequence with the detection device means, wherein the detection device means comprises a frame grabber means for separately registering a projected grating of each projector in said sequence for explicit digital image analysis to obtain the topography of the curved surface.
  • the detection device may be comprised of a TV-camera that is synchronized for triggering with the flash light source of the first projector at the end of a first half raster period and for triggering the flash light source of the second projector at the beginning of a second half raster period, the flash light sources illuminating a complete TV-raster wherein the half rasters are adapted to complement each other.
  • both projectors may illuminate the eye at the same time and the Moire pattern resulting from the combination of two projections may be analyzed in order to determine the surface shape of the eye.
  • detection device may be further comprised of a data processor means for analyzing the image to obtain a height function according to Fourier transform, demodulation, and inverse Fourier transform, on one or more half rasters simultaneously in a one or two dimensional manner for reconstructing the topography.
  • the projectors may be angled to cause the projected gratings to intersect on the curved surface, wherein the frame grabber is capable of distinguishing the intersecting gratings and separately registering the gratings for individual digital analysis via the data processor means.
  • Microsaccadic movements are involuntarily movements of the eye that occur once or twice every second. Therefore, it is advantageous to take these movements into account during the three dimensional mapping of the eye.
  • One embodiment of the disclosed technology uses multiple wavelength illumination of the eye during the measurement.
  • One wavelength can be used for excitation of the fluorescent dye during coded light imaging, while another wavelength or wavelengths can be selected so the features on the eye surface, including but not limited to the blood vessels and limbus can be resolved in the recorded images.
  • the said light of different wavelengths can be projected onto the eye simultaneously or in sequence.
  • the light of the said wavelengths can be projected before and after the structured light pattern or sequence is projected onto the eye.
  • locations of the ocular surface features in the images before and after the projection of the coded light can be compared to ensure that eye did not move during the measurement and the data set can be used for three dimensional surface reconstruction.
  • Embodiments herein allow for measurement of the complete anterior ocular surface, including regions which may be occluded by the eyelids even when said eyelids are retracted manually, by combining multiple measurements of the ocular surface taken for different gaze directions of the eye.
  • Each of said multiple measurements produces a partial surface model of the visible region of the ocular surface, including three dimensional surface coordinates and color or intensity measurements.
  • Feature information from the color or intensity component of the model is used to aid convergence of the model registration processes, allowing the smooth surfaces of the multiple said partial surface models to be uniquely registered in space into a complete model of the full anterior ocular surface.
  • the object being mapped need not be a human or animal eye and may be any other type of surface.
  • the disclosed systems and methods may be advantageous for mapping a surface of a moving or unstable object (biologic or non-biologic).
  • FIG. la is a schematic illustration of an example of a stereo- photogrammetic triangulation between single camera and single projector
  • FIG. lb is a schematic illustration of an example of a stereo- photogrammetric triangulation between two cameras.
  • FIG. lc is a schematic illustration of an example of a stereo- photogrammetric triangulation with two projectors.
  • FIG. 2 is an illustration of an example of a system for measuring the ocular surface.
  • FIG. 3 is an illustration of use of fixation targets to acquire data for multiple gaze directions for the purpose of mapping the ocular area under the eyelids
  • FIG. 4 is a flow chart illustrating an example data acquisition process.
  • FIG. 5 is a flow chart depicting a sample stitching algorithm for combining three dimensional surface models obtained at multiple gaze directions for the purpose of mapping the surface of the eye under the eyelids.
  • FIG. 6 is an illustration demonstrating examples of intermediate and final results of the stitching algorithm for mapping the surface of the eye under the eyelids.
  • FIG. 7 is a flow chart depicting a sample iterative data acquisition process for operating in Discrete Station Mode.
  • FIG. 8 is a flow chart depicting an example processing tree for
  • FIG. 9 is a flow chart depicting a sample measurement process for Operation of Continuous Processing Mode
  • the present disclosure is directed toward systems and methods for performing surface measurement, mapping, and modeling of the complete anterior surface of the eye, including the corneal and scleral regions of the eye.
  • ocular topography measurements have focused on corneal topography and include: Placido disk measurements which using concentric illuminated rings to map the surface slope and infer topography elevation measurements from the slopes based on specific assumptions, scanning slit measurements which use using moving laser line or slit beam triangulation methods to calculate the elevation topography, and rastersterographic measurements which triangulate points from a static pattern projected on the ocular surface and interpolate between the triangulated points.
  • Scanning slit and rasterstereographic methods typically employ stereo-photogrammetric measurement pairs which may comprise a light-source 0102 and an imaging detector 0103 with known orientations relative to the surface to be measured 0101, as depicted schematically in Figure la, or may comprise two imaging detectors 0105 and 0106 with known orientations relative to the surface to be measured 0104, as depicted in Figure lb, or may comprise two or more light sources 0108 and 0109 and a single camera 0110 Figure lc.
  • Calibration of the stereo-photogrammetric pairs allows accurate triangulation of points in three dimensional space from the two-dimensional locations of the points in the imaging planes of the respective light sources or imaging detectors.
  • a system for mapping and modeling of the ocular surface can include an optical measurement head comprised of one or more pattern projection system, one or more cameras, and a fixation target array for fixing the gaze direction of the subject during the series of measurements, a mounting system for the optical measurement head comprising a mounting stand attached to manipulator with a chin and forehead rest for controlling the relative orientations of the eye of the subject to be measured and the optical measurement head, a computing device connected to the measurement head for controlling measurement acquisition and processing the acquired data, and a display screen.
  • Figure 2 shows the schematic drawing of one embodiment of the technology described herein, including an optical head 0201, a manipulator comprising a moveable optical stand 0202, a chin rest 0205, a forehead rest 0206, an attached computing device 0203, and a display screen 0204 (which may include computer processing hardware configured to implement the analysis methods disclosed herein).
  • the technology can be in communication with a scleral contact lens manufacturing system 0207, which can utilize information related to the topographic map of the eye generated by the technology for diagnosis or customized treatment of the subject's eye or for manufacturing of a scleral contact lens for the subject's eye.
  • the scleral contact lens manufacturing system 0207 can be geographically remote from the rest of the system, and the topographic map information can be communicated to the manufacturing system 0207 via a network connection (e.g., the Internet, a local or wide area network, etc.).
  • a network connection e.g., the Internet, a local or wide area network, etc.
  • the projected structured light pattern sequence also comprises additional flat-field frames which may precede the projected structured light pattern grids, follow them, or both.
  • the eye is illuminated by a uniform or almost uniform light with a wavelength that overlaps the transmission wavelength of the fluorescence emission filter located in the optical path of the cameras but does not overlap the excitation wavelength of the fluorescent substance introduced into the tear film.
  • a single image or several images are recorded by each camera during the said flat-field illumination of the eye.
  • these emission wavelength flat-field illumination images recorded before and after an excitation wavelength sequence of projected structured light patterns can be used to verify that the eye has not moved during the measurement and for correction of measurement artifacts caused by any such eye movement.
  • emission wavelength illumination imaging can be used to verify that the eye hasn't moved during the measurement sequence.
  • Some embodiments can be used for mapping the surface of the eye in order to create a custom back surface of the scleral lens for a comfortable fitting.
  • a measurement of the ocular surface is performed within a 12mm to 22mm diameter circle centered at the cornea apex.
  • a fixation target array can be used to guide the gaze direction of the subject during the surface measurement process such that measurements are taken with the eye oriented at each of a plurality of gaze directions, such that each measurement comprises data of a different portion of the ocular surface.
  • each element in the fixation target array is chosen so that there is a significant overlap between measurements that can be used for later manual or automated stitching of the plurality of measurements in order to obtain a single composite model of the eye surface within 12mm to 22mm diameter region centered at the corneal apex, which may include regions of the ocular surface normally hidden by eyelids.
  • Figure 3 depicts a measurement process comprising measurements of the eye 0602 fixed at three gazed directions corresponding to three different illuminated elements in the gaze fixation target array 603, 607, 611. Flat- field illumination images captured at each gaze direction are depicted in 0601 0605 and 0609.
  • the manual or automated registration and stitching of the resulting three dimensional datasets taken at each gaze direction can be performed either by optimizing correspondence between the three dimensional shapes and features in each dataset as measured using fluorescence wavelength illumination, or by combining said optimization of correspondence with three dimensional shapes with color or monochrome intensity feature information for optical surface features such as the corneal limbus and blood vessels obtained during flat-field illumination.
  • the intensity of the fluorescence can be evaluated during or after the focusing process and a visual or auditory indication can be used to alert the operator of the state of the fluorescent dye on the ocular surface. In the case of insufficient dye the operator can be prompted to add more dye to the eye before continuing with the measurement.
  • the intensity of the illumination for the fluorescence and emission wavelength is selected so that it does not exceed the safety limits established by appropriate standards.
  • the pattern projection system can be capable of projecting one or more sequential predetermined two- dimensional patterns, whereas first and second cameras may be still or video CCD, CMOS or other cameras.
  • the herein topography system comprises between one and three simultaneous stereo measurements.
  • between one and three abovementioned stereo measurement methods can be used simultaneously to provide between one and three distinct measurements of the ocular surface.
  • the obtained measurements can then be combined, averaged, or otherwise mathematically manipulated in order to increase the measurement accuracy at the entire measurement region or at a specific portion of the ocular surface.
  • the triangulation between the projector and the right camera can be used for measurement of the right side of the eye and the triangulation between the projector and the left camera can be used for measurement of the left side of the eye, while the triangulation between the two cameras can be used for the central portion of the eye, and overlap between the measurement regions can be used to achieve error reduction.
  • FIG. 4 The measurement process for the some embodiments is depicted in Figure 4, wherein the subject's eye is prepared with fluorescent dye 0901 and positioned with respect to the optical head 0902.
  • the computing device calculates and checks focus and data quality metrics 0903 before initiating the formal data acquisition.
  • Formal data acquisition takes the form of a loop 0904 wherein a frame of the structure light pattern sequence is projected onto the ocular surface 0905 and the emitted or reflected light from the ocular surface is captured by each imaging detector 0906 and transferred to the controlling computing device.
  • the acquired images are processed into three dimensional topographical models of the anterior ocular surface 0907.
  • the encoding schemes are based on grids of parallel lines which shift in space on the projector frame during the encoding process.
  • Such encoding sequences can be parallel vertical lines and parallel horizontal lines in sequential or alternating series where the shifting vertical lines are used to encode the column values and the shifting horizontal lines are used to encode the row values.
  • the encoding sequence can be a shifting Cartesian grid composed of parallel vertical lines and parallel horizontal lines superimposed on the same pattern frame. In the latter case, the horizontal and vertical components of the pattern can be isolated algorithmically via Fourier methods.
  • a full three dimensional topographical model of the anterior ocular surface can be obtained from a single measurement taken with the gaze direction of the eye fixed on a single fixation point.
  • the lids of the eye are retracted manually by the practitioner to expose the desired extent of the scleral and corneal regions of the anterior surface to be imaged.
  • a full three dimensional topographical model of the anterior ocular surface is created as a composite model from a plurality of individual three dimensional topographical models of the anterior ocular surface where each individual three dimensional topographical model is calculated from a measurement taken with the eye fixed on each of a plurality of fixation points.
  • registration of the various topography models is accomplished using feature information gleaned from flat-field illumination of the eye captured in conjunction with the coded light patterns. This can be done before or after using emission wavelength, or can be done simultaneously by using non-overlapping light source and color camera.
  • Scale Invariant Feature Transform SIFT
  • BM Block Match
  • SIFT Scale Invariant Feature Transform
  • BM Block Match
  • Comparison of component features in each feature based model description allows direct registration of models with respect to one another.
  • An iterative registration algorithm is then used to refine the fit, where ⁇ , ⁇ , ⁇ components, and feature proximity are all used as optimization parameters.
  • the stitching registration process by which the individual three dimensional topography models are combined into a composite three dimensional topography model of the full extents of the anterior ocular surface is depicted in Figure 5.
  • Each of the series of individual three dimensional topography models 1301 is analyzed 1302 to compute a rough orientation vector, from which the estimated overlap regions between each model can be computed 1303, which are then used to determine the order by which models will be registered to one another 1304.
  • models with greater overlap are registered to each other, then subsequent models are registered to the growing composite model. This maximizes the overlap between models at each registration step, increasing the reconstruction accuracy.
  • the models comprising the series of individual three dimension topography models are then registered and stitched in a loop 1305 according to the calculated stitching order.
  • Models u-1 1306 and u 1307 are registered to each other by using their calculated feature descriptions 1308 and 1309 to identify the subset of features in each model feature description common to both models, and computing and applying a transform 1310 from the model u to model u-1.
  • An iterative closest points algorithm 1312 is then applied which optimizes coefficients derived from matching nearest neighbors in each model, using a coordinate space which includes X, Y, and Z three dimensional coordinates as well as color feature information as basis vectors.
  • the stitched model then becomes model u-1, and the next model in the series according to the computed stitching order takes the role of model u.
  • FIG. 6 depicts an example of the stitching process results for a series of measurement comprising three gaze directions.
  • the flat-field illumination frames from the structured light pattern illumination measurements 1401, 1403, and 1405 are included to indicate the show direction of the eye relative to the imaging detector.
  • the individual three dimensional topographical models resulting from the measurements at each gaze direction are displayed as 1402, 1404, and 1406.
  • the final stitched composite three dimensional topographical model is shown in 1407.
  • the registration process efficiency and robustness is aided by accurate estimation of the gaze orientation vector prior to application of nearest neighbor techniques such as feature matching or iterative closest points algorithms.
  • Accurate gaze estimation allows estimation of overlap regions between models, allowing search segmentation.
  • Such gaze estimation is accomplished by identifying the corneal limbus by leveraging the contrast between the predominantly white sclera and the iris pigmentation.
  • a plane can be fit to the three-dimensional limbus points, for example using a least squares criterion. The normal vector to the plane closely approximates the gaze orientation vector.
  • the limbus plane is also used as a clipping plane to exclude features under the transparent corneal membrane, which are distorted by the optical properties of said corneal membrane, from the feature classification and matching algorithms.
  • speed and accuracy of the model registration process is enhanced by excluding all features not pertaining to the scleral surface from the feature classification and matching algorithms.
  • a primary interfering feature is represented by eyelashes which protrude into the optical path of the imaging detector.
  • eyelashes can be identified by generating disparity maps based on images obtained during flat-field illumination of the eye during the measurement process.
  • full three dimensional topographical models of the anterior ocular surface can be created either by manually retracting the lids of the eye to expose the entire portion of the ocular surface to be measured or by combining a plurality of individual three dimensional topography models each containing some segment of the entire portion of the ocular surface to be measured into a composite three dimensional topographical model of the anterior ocular surface.
  • the technology can operate in distinct modes: a Discrete Station Mode and a Continuous Processing Mode.
  • the operation of the device is similar to the operation described previously and described in Figure 4.
  • the steps 0902 through 0908 are operated in a loop where each iteration of the loop is realized by fixing gaze direction of the subject on a distinct element of the gaze fixation target array.
  • the modified process is depicted in Figure 7,
  • the image sequence comprising the data from each measurement of the ocular surface is then processed into an individual three dimensional topographical model by the computing device, and the plurality of individual three dimensional topographical models are then combined into a composite model of the full extent of the anterior ocular surface portion measured.
  • the locations of the elements in the gaze fixation target array are chosen such that the portion of the ocular surface measured when the gaze direction is fixed on element N- 1 overlaps significantly with the portion of the ocular surface measured when the gaze direction is fixed on element N to allow optimal registration of individual models with respect to one another.
  • Figure 8 details the Discrete Station Mode processing tree.
  • Scan data comprised of a sequence of images of the eye subject to a sequence of structured light patterns and flat- field illuminations is acquired for a given gaze direction 1601.
  • a stability check 1602 is then performed by comparing flat-field illumination images comprising part of said pattern sequence and a stability metric is calculated. If the stability metric meets established criteria, the acquired data is processed through path A, including surface point triangulation by means of Coded-light reconstruction 1604 and Rasterstereographic reconstruction 1605.
  • a preliminary feature classification is performed on each reconstruction result 1606, and weighting coefficients and surface constraints are applied to reconcile the various surface reconstructions 1607, before the flat-field illumination images are used in conjunction with SIFT and Block Match algorithms to create a detailed feature-based compliment to the three dimensional surface topology 1608 that is stored to the computer memory 1609 for future use.
  • the data is processed through path B which uses a combination of slit-spline reconstruction 1610 square wave phase-shift reconstruction 1611 on individual image frames or pairs of image frames to mitigate the impact of eye motion by restricting the effective measurement windows to small fractions of the entire measurement window.
  • a feature classification is then performed on each of the surface reconstructions 1612 using the flat-field illumination frame that preceded the structured light sequence. The process then repeats using last two frames or frame pairs in the structured light sequence in conjunction with the flat- field frame that follows the structured light sequence.
  • frames N-1 and N of the structured light sequence are processed using the slit-spline 1613 and phase-shift reconstruction 1614 techniques and the feature classifications are performed using the flat-field frame 1615.
  • the reconstruction results for Frames 1 and 2 are reconciled to each other using surface constraints and weighting coefficients 1616, and the same process is applied to the results for Frames N-2 and N-1, before the two reconciled results are combined and reconciled to one another 1618.
  • a final feature classification step is performed to create the feature-based compliment 1620 and the topology and feature information are stored to the computer memory for future use 1621.
  • each image frame or pair of image frames in the acquire data image sequence may be processed as a separate measurement into a separate individual three dimensional topography model with its own feature-based compliment as well, as opposed to processing only the beginning and end portions of the sequence.
  • Discrete Station Mode is also available for real-time control of the measurement system hardware by the attached computing device.
  • the data are re-acquired 1622.
  • the data acquisition and data processing functions are directly coupled by means of the attached computing device to improve data quality and measurement speed.
  • the gaze fixation target array is initialized by the computer, then the ocular surface is illuminated by a flat-field illumination frame and a structured light sequence truncated to one or a few frames of structured light patterns and image sequences of the illuminated ocular surface are captured by the imaging sensors.
  • the computer processes the acquired image sequence into an individual three dimensional topography model with a feature-based compliment description, orientation vector, and model extents, then computes a desired gaze direction and updates the illumination of the gaze fixation target array. After each subsequent data acquisition, convergence, coverage, and quality metrics are calculated as part of the computation of the desired gaze direction for the next acquisition. This process continues until the calculated metrics meet established criteria, and the controlling computer algorithm ends the scan.
  • the operation of Continuous Processing Mode for this embodiment is detailed in Figure 9.
  • the process begins with the subject's eye being prepared with fluorescent dye 1701 and fixed on the gaze fixation target array in its initial state 1702.
  • the controlling algorithm enters the control loop 1703, during which it acquires image data by illuminating the eye with the structured light sequence and capturing images of the illuminate ocular surface 1704, processes the acquired image data using the Templated Grid Search algorithm 1705 and then applies one or both of Slit-spline reconstruction and square wave phase reconstruction 1706, computes the feature -based description compliment and the coverage and quality metrics 1707, and reconciles the topology and feature results 1708.
  • this reconciled topology and feature result becomes the composite model
  • the composite model is updated through another reconciliation process with the newly measured components 1709.
  • Coverage and quality metrics are then calculated 1710 and compared against established criteria 1711, after which a new gaze direction is determined and indicated on the fixation target array 1712, triggering the next scan iteration of the loop.
  • the new measurement is registered to the previous measurements by means of the feature-based registration stitching method described previously. The process repeats until the coverage and quality metrics meet the established convergence criteria.
  • Mode offers speed and data quality improvements by providing real-time feedback on data quality and allowing the processing computer to correct for deficiencies in the acquired data during the initial measurement process, minimizing the possibility of repeating the measurement at a later time.
  • the data quality and speed are both improved significantly by simultaneous projection of the structured light pattern sequence and flat-field illumination frame using non-overlapping wavelength bands.
  • a display screen attached to the computing device displays provides operational feedback to the user. This feedback includes real-time views of the acquired imagery for use in alignment and focusing of the measurement system with respect to the surface to be measured, intermediate stage progress indicators including focusing quality indicators, as well as visualizations of the three dimensional topographical models and optometrically useful realizations of said models and quantities derived from them.
  • Additional embodiments can be used to create a three dimensional model of the eye surface and electronically transmit it to the scleral and contact lens manufacturing facility for designing and building a custom lens that is specifically fit to a patient's eye.
  • Both the system and method embodiments disclosed herein may be used independently or in can be combined with a Placido disk based corneal topography measurement within the same embodiment.
  • the traditional Placido disk measurement approach can be used to provide a rapid corneal measurement without the fluorescent substance, while the structured light system can be used for simultaneous measurement of the ocular surface measurement in the corneal and scleral regions. Measurements made by either approach may stand alone, may be registered into a common data set to complement one another, or may be incorporated as algorithmic constraints to one or both datasets to improve the accuracy of a single composite model of the ocular surface
  • Some examples of the disclosure may have more than two cameras arranged in such fashion that more than three triangulation pairs can be created during the device operation and data analysis.
  • systems and methods disclosed herein may achieve one, some, or all of the following advantages and/or provide some or all of the following functionality.
  • three simultaneous independent measurements are used, which may advantageously provide error reduction in overlap regions through averaging, error reduction in overlap regions through constraints, wider field-of-view by having cameras out at angles to the surface to be measure, and/or more accurate tracking, because it may be difficult to use corneal points for tracking.
  • coded structured light is used to map the surface of the object (e.g., the anterior surface of the eye).
  • Pattern sequences can offer higher spatial resolution than single patterns, and unique pixel encoding can eliminate iterative point searching, which advantageously can increase speed and/or accuracy of the mapping.
  • multiple patterns may be avoided because of eye movements. Tracking and processing segmentation can allow for correction for eye movements.
  • trimming coded light bins to single stripes at bin edges can reduce dynamic range requirements and/or exposure modification requirements of the cameras.
  • trimming coded light sequence to moving set of identical patterns can provide processing flexibility. For example, with smoothness constraints for the surface, the systems can perform true coded light methods.
  • the techniques can include Standard Coded Light (e.g., using all frames or subset of frames), phase-shift or scanning slit using single frames, etc.
  • the systems and methods can be implemented to allow selection between techniques based on eye stability. Multiple techniques to constrain solutions and improve surface accuracy can be adopted.
  • use of identical patterns means each frame, or any subset, can be processed by itself using alternate methods to battle eye movement during the mapping sequence.
  • a fluorescent dye can be applied to the object (e.g., the anterior surface of the eye) to deal with differences in surface reflectivities.
  • real-time brightness/quality indicator can be used to permit analysis of the dye coverage and fluorescence intensity during focusing or mapping, and optionally, after measurement.
  • the systems and methods can be configured to warn if scans need to be repeated.
  • the disclosed focusing methods of the camera and projector geometry may allow a simple focusing/alignment indication by matching the projected focusing pattern with a fixed reference display pattern.
  • the measurements in multiple wavelengths allow obtaining position and intensity (e.g., XYZI) data.
  • position and intensity e.g., XYZI
  • Such data may allow eye movement tracking, stitching registration of multiple partial datasets (e.g., using intensity domain features to constrain algorithms for stitching smooth surfaces, limbus detection for scleral lens fitting, and/or detection of problem spots such as scaring to avoid during scleral lens fitting.
  • simultaneous pattern projections in multiple wavelengths can be used.
  • flat field and structured light can be projected simultaneously.
  • Red, green, blue (RGB) coded light is can be used in ophthalmic or non-ophthalmic settings.
  • substantially the entire sclera can be mapped by moving the gaze direction, taking partial datasets, and then combining the datasets.
  • the systems and methods utilize discrete gaze directions and the processing starts from the XYZI models collected (including third party data).
  • Continuous Processing Mode the systems and methods utilize automated, guided data acquisition. The gaze direction moves around in field following an indicator directed by the algorithm until convergence is obtained.
  • any of the systems and methods disclosed herein can be combined with a Placido disk for two types of independent corneal measurements.
  • a system for measuring an anterior surface topography of an eye comprising: a pattern projection system configured to emit light towards an ocular surface of the eye, wherein the pattern projection system is configured to project a sequence of patterns onto the ocular surface; one or more image sensors configured to record one or more images of the patterns resulting from the projected pattern sequence, an analysis system comprising computing hardware configured to determine a topographic map of the ocular surface from the one or more images of the patterns.
  • the system of aspect 1 or aspect 2 wherein the patterns in the sequence are projected in two or more wavelength bands.
  • the system of any one of aspects 1-3 wherein one or more patterns in the projected sequence of patterns are emitted in a wavelength band at least partially overlapping an excitation wavelength of a fluorescent substance adapted to be applied to the eye.
  • the system of any one of aspects 1-5 wherein the pattern projection system is configured to produce structured light patterns in one or more of three modes which can be operated either simultaneously or sequentially, such that: in a first mode, an emitted wavelength range of the pattern projection system overlaps an excitation wavelength of a fluorescent substance used to prepare the ocular surface but does not overlap with a fluorescence wavelength of the fluorescent substance; in a second mode, an emitted wavelength range of the pattern projection system overlaps a fluorescence wavelength of a fluorescent substance used to prepare the ocular surface but does not overlap an excitation wavelength of the fluorescent substance; and in a third mode, an emitted wavelength range of the pattern projection system overlaps neither an excitation wavelength of a fluorescent substance used to prepare the ocular surface nor a fluorescence wavelength of the fluorescent substance.
  • the system of any one of aspects 1-6 wherein the system is configured to perform a measurement of the ocular surface in a measurement duration less than about 0.5 seconds between the microsaccadic movements of the eye.
  • the pattern projection system is configured to simultaneously project multiple individual patterns from the sequence of patterns in a coded light sequence, wherein each individual pattern is projected in a non-overlapping wavelength band.
  • the system of aspect 8 wherein the one or more image sensors comprise a multi-color imaging detector configured to record each individual pattern in a separate recorded color channel, whereby system is configured to project and record the coded light sequence in one or more exposures.
  • the system of any one of aspects 1-9 wherein illumination levels incident on the ocular surface are less than 3.9 x 10 "3 Joules of radiant energy as measured through a 7-mm aperture located within 5 mm of the projector focus.
  • the system of any one of aspects 1-10 further comprising a fixation target system configured to permit a gaze of the eye to be sequentially fixed at a plurality of gaze directions.
  • the fixation target system comprises one or both of: a plurality of targets to be illuminated in sequence or an emissive screen configured to display stationary or moving gaze fixation targets.
  • the system of aspect 14 wherein the pattern projection system is configured to illuminate the ocular surface with a flat field in a first wavelength and one or more structured light patterns in second wavelength.
  • system of any one of aspects 1-15 further comprising a display device configured to display a representation of the topographic map of the ocular surface or a representation of one or more optometric values derived from the topographic map.
  • system of any one of aspects 1-16 further comprising a scleral contact lens manufacturing system, wherein the system is configured to communicate information related to the topographic map of the ocular surface to the scleral contact lens manufacturing system.
  • a method for calculating a three-dimensional topographical model of an anterior ocular surface of an eye comprising: under control of an ocular topographic mapping system comprising computer hardware: receiving images of projected structured light patterns that are reflected or emitted from the anterior ocular surface, the images obtained from a system comprising a plurality of imaging sensors configured to record images projected on the anterior ocular surface by a pattern projection system; analyzing the received images using one or more of the following techniques: coded light triangulation between any one of the plurality of imaging sensors and the pattern projection system, or rastersterographic triangulation between any two of the plurality of imaging sensors, slit-spline surface reconstruction, or phase-shift surface reconstruction; and determining, based at least in part on the analyzed images, a composite measurement of topography of at least a portion of the anterior ocular surface.
  • the method of aspect 18, wherein determining the composite measurement of topography of at least a portion of the anterior ocular surface comprises combining a plurality of individual topography measurements of portions of the anterior surface of the eye taken at a plurality of orientations of an optical axis of the eye.
  • each of the plurality of individual topography measurements is created from a series of images of projected structured light pattern sequences reflected from the anterior ocular surface, wherein each pattern sequence comprises: at least one image of the eye where the eye is illuminated in a wavelength range that overlaps a fluorescence wavelength of a fluorescent dye used to prepare the ocular surface but does not overlap an excitation wavelength of the fluorescent dye, and at least one projected structured light pattern where the projected pattern is illuminated in a wavelength range that overlaps the excitation wavelength of the fluorescent dye used to prepare the ocular surface but does not overlap the fluorescence wavelength of the fluorescent dye.
  • determining the composite measurement of topography of at least a portion of the anterior ocular surface comprises: analyzing each of the plurality of individual topography measurements to provide a respective individual three-dimensional topographical model of a segment of the ocular surface, wherein the individual three-dimensional topographical model comprises three dimensional coordinate data, color intensity data, and a feature-based description of the individual three-dimensional topographical model produced from analyzing captured images of a reflected or a fluorescent pattern sequence.
  • each of the plurality of individual topography measurements is taken in a Discrete Station Mode wherein an individual topography measurement is acquired with an optical axis of the eye directed at one of a plurality of fixed location fixation targets, wherein each individual topography measurement is created from a series of images of projected structured light pattern sequences reflected from the ocular surface, where each pattern sequence comprises: at least two flat-field images of the eye where the eye is illuminated in a wavelength range that overlaps a fluorescence wavelength of a fluorescent dye used to prepare the ocular surface but does not overlap an excitation wavelength of the fluorescent dye, and a sequence of at least one projected structured light pattern where the projected pattern is illuminated in wavelength range that overlaps the excitation wavelength of the fluorescent dye used to prepare the ocular surface but does not overlap the fluorescence wavelength of the fluorescent dye, and wherein at least one of the at least two flat-field images of the eye precedes the sequence of structured light patterns, and at least one
  • the method of aspect 22 further comprising analyzing at least one of the flat-field images which preceded the sequence of structured light patterns and at least one of the flat-field images which followed the structured light patterns to compute a metric describing apparent motion of the eye during a measurement period.
  • the method of aspect 23 further comprising determining, based at least in part on the computed metric, at least one processing technique for constructing a respective individual three-dimensional topographical model.
  • each of the individual topography measurements is taken in Continuous Processing Mode wherein an individual topography measurement is acquired and processed in a continuous loop until a desired convergence metric or a time threshold is reached, the method further comprising: computing a composite three-dimensional topography model of the anterior ocular surface from combinations of individual three-dimensional topography measurements acquired during the continuous loop in a measurement window, while the orientation of the eye is allowed to change during the measurement window.
  • each of the individual topography measurements taken in Continuous Processing Mode comprises: at least one flat- field image of the eye where the eye is illuminated by a wavelength range that overlaps the fluorescence wavelength of the fluorescent dye used to prepare the ocular surface but does not overlap the excitation wavelength of the fluorescent dye, and a sequence of at least one projected structured light patterns where the projected pattern is illuminated in a wavelength range that overlaps the excitation wavelength of the fluorescent dye used to prepare the ocular surface but does not overlap the fluorescence wavelength of the fluorescent dye, the method further comprising: projecting the at least one flat-field image and the sequence of at least one projected structured light patterns at least partially overlapped in time, if a color camera is used, or projecting the at least one flat-field image and the sequence of at least one projected structured light patterns sequentially, if a monochromatic cameras is used.
  • the method of aspect 26 further comprising: processing each of the individual topography measurements into a rough individual three-dimensional topography along with extents and orientation of each rough individual three-dimensional topography model; combining each individual rough three-dimensional topography model with previous rough three-dimensional topography models taken during a measurement period of a subject into a rough composite three-dimensional topography model of a measured portion of the ocular surface; evaluating the extents and surface metrics of the measured portion of the ocular surface to provide a gauge of measurement quality and completeness; and communicating the measurements of quality and completeness in real time such that a gaze direction of the subject can be adjusted to facilitate completion of the measurement of the eye of the subject.
  • the method of any one of aspects 18-27, wherein the projected structured light patterns comprise grids of parallel lines or square wave patterns.
  • the method of any one of aspects 18-29 wherein the one or more of the following techniques comprise at least two of the techniques, and the at least two techniques are applied to constrain or refine the composite measurement of topography.
  • the method of any one of aspects 18-30 further comprising: analyzing received images of reflected flat-field illumination to compute a feature-based description of individual three-dimensional topographical models of the measured portion of the ocular surface for registering the individual three- dimensional topographical models in Cartesian space; and analyzing, based at least in part on the feature-based description, rotation and translation of an individual three- dimensional topographical model with respect to another individual three-dimensional topographical model or with respect to that same individual three-dimensional topographical model over the course of the measurement.
  • the method of aspect 31 wherein the feature- based description contains points which correspond to the corneal limbus of the eye, the method further comprising creating a masking region which excludes non- topographical features in the corneal region from the feature-based description of the individual three-dimensional topographical models to prevent optical properties of the cornea from skewing the analysis of the rotation and translation of an individual three-dimensional topographical model, and wherein a plane-fit to the corneal limbus points is used to determine an approximate orientation for the optical axis of the eye.
  • the method of any one of aspects 18-32 further comprising: receiving flat-field images captured simultaneously on more than one of the plurality of imaging sensors; and determining regions of the images that are occluded in one or more received image by protruding eyelashes.
  • optical measurement head does not imply that the components or functionality described or claimed as part of the optical measurement head are all configured in a common package. Indeed, any or all of the various components of an optical measurement head, whether control logic or other components, can be combined in a single package or separately maintained and can further be distributed in multiple groupings or packages or across multiple locations.
  • Each of the processes, methods, and algorithms described herein and/or depicted in the attached figures may be embodied in, and fully or partially automated by, code modules executed by one or more physical computing systems, computer processors, application-specific circuitry, and/or electronic hardware configured to execute computer instructions.
  • computing systems can include general purpose computers configured with specific executable instructions for performing the disclosed methods or special purpose computers, servers, desktop computers, laptop or notebook computers or tablets, personal mobile computing devices, mobile telephones, and so forth.
  • a code module may be stored in non- transitory computer memory, compiled and linked into an executable program, installed in a dynamic link library, or may be written in an interpreted programming language.
  • Code modules may be stored on any type of non-transitory computer-readable medium, such as physical computer storage including hard drives, solid state memory, random access memory (RAM), read only memory (ROM), optical disc, volatile or non- volatile storage, combinations of the same and/or the like.
  • the methods and modules may also be transmitted as generated data signals (e.g., as part of a carrier wave or other analog or digital propagated signal) on a variety of computer-readable transmission mediums, including wireless-based and wired/cable- based mediums, and may take a variety of forms (e.g., as part of a single or multiplexed analog signal, or as multiple discrete digital packets or frames).
  • the results of the disclosed processes and process steps may be stored, persistently or otherwise, in any type of non-transitory, tangible computer storage or may be communicated via a computer-readable transmission medium.
  • Any processes, blocks, states, steps, or functionalities in flow diagrams described herein and/or depicted in the attached figures should be understood as potentially representing code modules, segments, or portions of code which include one or more executable instructions for implementing specific functions (e.g., logical or arithmetical) or steps in the process.
  • the various processes, blocks, states, steps, or functionalities can be combined, rearranged, added to, deleted from, modified, or otherwise changed from the illustrative examples provided herein.
  • additional or different computing systems or code modules may perform some or all of the functionalities described herein.
  • the processes, methods, and systems may be implemented in a network (or distributed) computing environment.
  • Network environments include enterprise-wide computer networks, intranets, local area networks (LAN), wide area networks (WAN), personal area networks (PAN), cloud computing networks, crowd- sourced computing networks, the Internet, and the World Wide Web.
  • the network may be a wired or a wireless network (e.g., a terrestrial and/or satellite network) or any other type of communication network.
  • any reference to "one embodiment” or “some embodiments” or “an embodiment” means that a particular element, feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment.
  • the appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment.
  • Conditional language used herein, such as, among others, “can,” “could,” “might,” “may,” “e.g.,” and the like, unless specifically stated otherwise, or otherwise understood within the context as used, is generally intended to convey that certain embodiments include, while other embodiments do not include, certain features, elements and/or steps.
  • the articles “a” or “an” as used in this application and the appended claims are to be construed to mean “one or more” or “at least one” unless specified otherwise.
  • a condition A or B is satisfied by any one of the following: A is true (or present) and B is false (or not present), A is false (or not present) and B is true (or present), or both A and B are true (or present).
  • a phrase referring to "at least one of a list of items refers to any combination of those items, including single members.
  • "at least one of: A, B, or C” is intended to cover: A, B, C, A and B, A and C, B and C, and A, B, and C.
  • Conjunctive language such as the phrase "at least one of X, Y and Z," unless specifically stated otherwise, is otherwise understood with the context as used in general to convey that an item, term, etc. may be at least one of X, Y or Z. Thus, such conjunctive language is not generally intended to imply that certain embodiments require at least one of X, at least one of Y, and at least one of Z to each be present.
  • a system for measuring an anterior surface topography of an eye comprising:
  • a projection system configured to emit light towards an ocular surface of the eye, wherein the projection system may contain one or more projectors, each configured to project a single pattern or a sequence of patterns onto the ocular surface;
  • one or more image sensors configured to record one or more images of the patterns resulting from the projected pattern or sequence
  • an analysis system comprising computing hardware configured to determine a topographic map of the ocular surface from the one or more images of the patterns.
  • fixation target system comprises one or both of: a plurality of targets to be illuminated in sequence or an emissive screen configured to display stationary or moving gaze fixation targets.
  • At least one of the one or more image sensors is configured to simultaneously record at least one of the one or more images in a plurality of wavelengths.
  • the pattern projection system is configured to illuminate the ocular surface with a flat field in a first wavelength and one or more structured light patterns in second wavelength.
  • a method for calculating a three-dimensional topographical model of an anterior ocular surface of an eye comprising:
  • determining the composite measurement of topography of at least a portion of the anterior ocular surface comprises combining a plurality of individual topography measurements of portions of the anterior surface of the eye taken at a plurality of orientations of an optical axis of the eye.
  • each of the plurality of individual topography measurements is created from one or series of images of projected structured light pattern sequences reflected from the anterior ocular surface, wherein each pattern sequence comprises:
  • the projected pattern is illuminated in a wavelength range that overlaps the excitation wavelength of the fluorescent dye used to prepare the ocular surface but does not overlap the fluorescence wavelength of the fluorescent dye.
  • determining the composite measurement of topography of at least a portion of the anterior ocular surface comprises:
  • each of the plurality of individual topography measurements to provide a respective individual three-dimensional topographical model of a segment of the ocular surface.
  • the individual three-dimensional topographical model comprises three dimensional coordinate data, color intensity data, and a feature-based description of the individual three-dimensional topographical model produced from analyzing captured images of a reflected or a fluorescent pattern sequence.
  • each of the plurality of individual topography measurements is taken in a Discrete Station Mode wherein an individual topography measurement is acquired with an optical axis of the eye directed at one of a plurality of fixed location fixation targets, wherein each individual topography measurement is created from a series of images of projected structured light pattern sequences reflected from the ocular surface, where each pattern sequence comprises:
  • a sequence of at least one projected structured light pattern where the projected pattern is illuminated in wavelength range that overlaps the excitation wavelength of the fluorescent dye used to prepare the ocular surface but does not overlap the fluorescence wavelength of the fluorescent dye, and wherein at least one of the at least two flat-field images of the eye precedes the sequence of structured light patterns, and at least one of the at least two flat-field images of the eye follows the sequence of structured light patterns;
  • the method further comprises processing each of the plurality of individual topography measurements into a respective individual three- dimensional topography model.
  • the projected structured light patterns comprise grids of parallel lines or square wave patterns.
  • the method of claim 18 wherein the feature -based description contains points which correspond to the corneal limbus of the eye, the method further comprising creating a masking region which excludes non-topographical features in the corneal region from the feature -based description of the individual three- dimensional topographical models to prevent optical properties of the cornea from skewing the analysis of the rotation and translation of an individual three-dimensional topographical model, and wherein a plane-fit to the corneal limbus points is used to determine an approximate orientation for the optical axis of the eye.

Abstract

Apparatus and method for producing accurate elevation models and representations of the ocular surface by combining multiple partial measurements to circumvent interference by the eyelids. For each partial measurement of the eye, elevation and feature models of the measured surface are created and analyzed to create an accurate composite representation of the full anterior eye surface. The measurements provide a three-dimensional map of the surface which can be used for a variety of ophthalmic and optometric applications from astigmatism and keratoconus diagnostics to contact lens fitting.

Description

SYSTEMS AND METHODS FOR MAPPING THE OCULAR SURFACE USUALLY OBSTRUCTED BY THE EYELIDS
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This application claims the benefit of and priority to U.S.
Provisional Patent Application No. 62110362, filed January 30, 2014, entitled "Systems and Methods for Mapping the Ocular Surface Usually Obstructed by the Eyelids," which is hereby incorporated by reference herein in its entirety.
Background
Technical Field
[0002] The present disclosure relates to three dimensional imaging using the structured light approach, and more particularly to ocular surface measurement over the anterior surfaces of the cornea and sclera.
Description of Related Art
[0003] The accurate knowledge of the corneal surface is very important for diagnostics and treatment of a number of ocular conditions. Cornea is responsible for about 70% of the refractive power of the eye and therefore the corneal topography has a great importance in determining the quality of vision. It is commonly used for diagnosis of keratoconus, for selecting appropriate soft contact lenses, for fitting scleral lenses, and for topography guided Laser-Assisted in situ Keratomileusis (LAS IK).
[0004] Currently the majority of the corneal topography is performed using a Placido disk. The concept of the Placido disk was introduced by Antonio Placido in 1880 and since then it has been the primary method for corneal topography. The method is based on viewing or imaging the corneal reflection of series of concentric bright and dark rings positioned in front of the cornea. By increasing the number of concentric disks and placing them on a concave surface around the eye, it is possible to measure a large section of the cornea. But in most cases the data on the central zone needs to be interpolated and the data on the corneal periphery is often missing due to limited reflection. Additionally, Placido disks are not capable of measuring the scleral topography, which is important for custom fitting of scleral lenses.
[0005] In addition to the Placido disk, scanning slit systems, such as the
Orbscan II by Bausch & Lomb, Scheimpflug systems such as the Galilei by Zeimer Ophthalmic Systems, as well as rastersterographic systems such as the CTS by Par Technlogies and ESP by Eaglet Eye have been used for corneal topography. While these systems provide better measurement of the corneal apex they lack the coverage of the corneal periphery and sclera needed for scleral lens fitting.
[0006] In all the instruments mentioned above the mapping of the anterior surface of the eye is performed it the regions that are normally not obstructed by the upper and lower eyelid, therefore limiting the extent of the measurement and often not capturing a sufficient portion of the anterior surface on the upper and lower portions of the said anterior surface.
Summary
[0007] The present disclosure describes an advanced measurement system for mapping the complete three dimensional anterior ocular surface. Commonly used ocular topography measurements, including Placido-disk measurements, scanning slit beam measurements, and rastersterographic measurements have focused on the central region of the cornea. Extending such measurements to the entire anterior surface, specifically to include the sclera, introduces a new set of challenges. The obstruction of the upper and lower sclera with eyelids is one of the major challenges in the mapping of the eye surface.
[0008] A sample embodiment of the technology disclosed herein provides a system and method for obtaining the corneal and scleral maps under the eyelids.
[0009] In the said embodiment the surface of the eye is coated with a fluorescent dye and two or more projectors are configured to project light in a wavelength that overlaps the excitation band of the said dye. The resulting fluorescent light emitted from the dye covering the ocular surfaces is detected using one or more imaging sensors. Using the above fluorescent imaging method solves the complication of dissimilar reflective properties of scleral and corneal surfaces.
[0010] In the said embodiment the projected light may be comprised of a single pattern or a coded light sequence. The pattern may be comprised of vertical or horizontal lines, a combination of vertical and horizontal lines, or an arbitrary pattern with known geometrical characteristics that can be used to determine the ocular shape. The coded light sequence may contain series of structured light patterns which can be interpreted as time-series measurements, where the projected intensity at a given location over time has a unique pattern for each pixel or subregion of pixels, allowing accurate, unambiguous identification of triangulation points in each member of the stereo-photogrammetric or stereo-projection pair. Said triangulation measurements performed in order to obtain the three dimensional surface of the eye can be performed between two cameras, between two projectors or between one of the cameras and the projector. The abovementioned triangulation measurements can be performed separately or in any combination of the above in order to increase measurement redundancy and reduce the measurement errors. Said coded-light sequences allow higher resolution measurements than conventional raster- stereographic methods by coding all pixels in the pattern region, instead of requiring interpolation between grid lines or grid intersection points.
[0011] Another embodiment may be comprised of a system for determining the topography of a curved surface, comprising a projection device means for projecting patterns of lines on the surface, the projection device means including two independent light projectors disposed at an angle relative to each other, each projector having a projection axis and including a grating having parallel straight lines, wherein each grating is positioned at a right angle to the projection axis of its respective projector, a rectangular diaphragm having the long sides parallel to the lines of the grating, a detection device means for registering an image formed on the surface and a flash light source synchronized in sequence with the detection device means, wherein the detection device means comprises a frame grabber means for separately registering a projected grating of each projector in said sequence for explicit digital image analysis to obtain the topography of the curved surface. [0012] In the said embodiment the detection device may be comprised of a TV-camera that is synchronized for triggering with the flash light source of the first projector at the end of a first half raster period and for triggering the flash light source of the second projector at the beginning of a second half raster period, the flash light sources illuminating a complete TV-raster wherein the half rasters are adapted to complement each other.
[0013] In another embodiment both projectors may illuminate the eye at the same time and the Moire pattern resulting from the combination of two projections may be analyzed in order to determine the surface shape of the eye.
[0014] In the said embodiment detection device may be further comprised of a data processor means for analyzing the image to obtain a height function according to Fourier transform, demodulation, and inverse Fourier transform, on one or more half rasters simultaneously in a one or two dimensional manner for reconstructing the topography.
[0015] In the said embodiment the projectors may be angled to cause the projected gratings to intersect on the curved surface, wherein the frame grabber is capable of distinguishing the intersecting gratings and separately registering the gratings for individual digital analysis via the data processor means.
[0016] Microsaccadic movements are involuntarily movements of the eye that occur once or twice every second. Therefore, it is advantageous to take these movements into account during the three dimensional mapping of the eye. One embodiment of the disclosed technology uses multiple wavelength illumination of the eye during the measurement. One wavelength can be used for excitation of the fluorescent dye during coded light imaging, while another wavelength or wavelengths can be selected so the features on the eye surface, including but not limited to the blood vessels and limbus can be resolved in the recorded images. The said light of different wavelengths can be projected onto the eye simultaneously or in sequence. In one embodiment of the disclosed technology the light of the said wavelengths can be projected before and after the structured light pattern or sequence is projected onto the eye. In said embodiment locations of the ocular surface features in the images before and after the projection of the coded light can be compared to ensure that eye did not move during the measurement and the data set can be used for three dimensional surface reconstruction.
[0017] Embodiments herein allow for measurement of the complete anterior ocular surface, including regions which may be occluded by the eyelids even when said eyelids are retracted manually, by combining multiple measurements of the ocular surface taken for different gaze directions of the eye. Each of said multiple measurements produces a partial surface model of the visible region of the ocular surface, including three dimensional surface coordinates and color or intensity measurements. Feature information from the color or intensity component of the model is used to aid convergence of the model registration processes, allowing the smooth surfaces of the multiple said partial surface models to be uniquely registered in space into a complete model of the full anterior ocular surface.
[0018] In other implementations, the object being mapped need not be a human or animal eye and may be any other type of surface. The disclosed systems and methods may be advantageous for mapping a surface of a moving or unstable object (biologic or non-biologic).
[0019] Other features and aspects of the disclosed technology will become apparent from the following detailed description, taken in conjunction with the accompanying drawings, which illustrate, by way of example, the features in accordance with embodiments of the disclosed technology. The summary is not intended to limit the scope of any disclosures described herein, which are defined solely by the claims attached hereto.
Brief Description of the Drawings
[0020] Throughout the drawings, reference numbers may be re-used to indicate correspondence between referenced elements. The drawings are provided to illustrate example embodiments described herein and are not intended to limit the scope of the disclosure.
[0021] FIG. la is a schematic illustration of an example of a stereo- photogrammetic triangulation between single camera and single projector [0022] FIG. lb is a schematic illustration of an example of a stereo- photogrammetric triangulation between two cameras.
[0023] FIG. lc is a schematic illustration of an example of a stereo- photogrammetric triangulation with two projectors.
[0024] FIG. 2 is an illustration of an example of a system for measuring the ocular surface.
[0025] FIG. 3 is an illustration of use of fixation targets to acquire data for multiple gaze directions for the purpose of mapping the ocular area under the eyelids
[0026] FIG. 4 is a flow chart illustrating an example data acquisition process.
[0027] FIG. 5 is a flow chart depicting a sample stitching algorithm for combining three dimensional surface models obtained at multiple gaze directions for the purpose of mapping the surface of the eye under the eyelids.
[0028] FIG. 6 is an illustration demonstrating examples of intermediate and final results of the stitching algorithm for mapping the surface of the eye under the eyelids.
[0029] FIG. 7 is a flow chart depicting a sample iterative data acquisition process for operating in Discrete Station Mode.
[0030] FIG. 8 is a flow chart depicting an example processing tree for
Discrete Station Mode
[0031] FIG. 9 is a flow chart depicting a sample measurement process for Operation of Continuous Processing Mode
Detailed Description
[0032] The present disclosure is directed toward systems and methods for performing surface measurement, mapping, and modeling of the complete anterior surface of the eye, including the corneal and scleral regions of the eye. [0033] Commonly use ocular topography measurements have focused on corneal topography and include: Placido disk measurements which using concentric illuminated rings to map the surface slope and infer topography elevation measurements from the slopes based on specific assumptions, scanning slit measurements which use using moving laser line or slit beam triangulation methods to calculate the elevation topography, and rastersterographic measurements which triangulate points from a static pattern projected on the ocular surface and interpolate between the triangulated points. Scanning slit and rasterstereographic methods typically employ stereo-photogrammetric measurement pairs which may comprise a light-source 0102 and an imaging detector 0103 with known orientations relative to the surface to be measured 0101, as depicted schematically in Figure la, or may comprise two imaging detectors 0105 and 0106 with known orientations relative to the surface to be measured 0104, as depicted in Figure lb, or may comprise two or more light sources 0108 and 0109 and a single camera 0110 Figure lc. Calibration of the stereo-photogrammetric pairs allows accurate triangulation of points in three dimensional space from the two-dimensional locations of the points in the imaging planes of the respective light sources or imaging detectors.
[0034] According to some embodiments of the technology described herein, a system for mapping and modeling of the ocular surface can include an optical measurement head comprised of one or more pattern projection system, one or more cameras, and a fixation target array for fixing the gaze direction of the subject during the series of measurements, a mounting system for the optical measurement head comprising a mounting stand attached to manipulator with a chin and forehead rest for controlling the relative orientations of the eye of the subject to be measured and the optical measurement head, a computing device connected to the measurement head for controlling measurement acquisition and processing the acquired data, and a display screen. Figure 2 shows the schematic drawing of one embodiment of the technology described herein, including an optical head 0201, a manipulator comprising a moveable optical stand 0202, a chin rest 0205, a forehead rest 0206, an attached computing device 0203, and a display screen 0204 (which may include computer processing hardware configured to implement the analysis methods disclosed herein). Optionally, in some embodiments, the technology can be in communication with a scleral contact lens manufacturing system 0207, which can utilize information related to the topographic map of the eye generated by the technology for diagnosis or customized treatment of the subject's eye or for manufacturing of a scleral contact lens for the subject's eye. The scleral contact lens manufacturing system 0207 can be geographically remote from the rest of the system, and the topographic map information can be communicated to the manufacturing system 0207 via a network connection (e.g., the Internet, a local or wide area network, etc.).
[0035] In some embodiments of the technology described herein, the projected structured light pattern sequence also comprises additional flat-field frames which may precede the projected structured light pattern grids, follow them, or both. In these flat-field frames the eye is illuminated by a uniform or almost uniform light with a wavelength that overlaps the transmission wavelength of the fluorescence emission filter located in the optical path of the cameras but does not overlap the excitation wavelength of the fluorescent substance introduced into the tear film. A single image or several images are recorded by each camera during the said flat-field illumination of the eye. In some embodiments, these emission wavelength flat-field illumination images recorded before and after an excitation wavelength sequence of projected structured light patterns can be used to verify that the eye has not moved during the measurement and for correction of measurement artifacts caused by any such eye movement. Since it has been reported that the microsaccadic movements of the eye involuntarily occur with a periodicity between 0.3 and 1 seconds it can be advantageous that the entire measurement sequence should last less than about 0.5 second. In embodiments where multiple frame structured light sequences are employed, emission wavelength illumination imaging can be used to verify that the eye hasn't moved during the measurement sequence.
[0036] Some embodiments can be used for mapping the surface of the eye in order to create a custom back surface of the scleral lens for a comfortable fitting. For said application it can be advantageous that a measurement of the ocular surface is performed within a 12mm to 22mm diameter circle centered at the cornea apex. For such large diameters the portions of the scleral and corneal surfaces may be hidden behind the eyelids, therein complicating the measurement procedure. [0037] In some embodiments, a fixation target array can be used to guide the gaze direction of the subject during the surface measurement process such that measurements are taken with the eye oriented at each of a plurality of gaze directions, such that each measurement comprises data of a different portion of the ocular surface. In said embodiments the location of each element in the fixation target array is chosen so that there is a significant overlap between measurements that can be used for later manual or automated stitching of the plurality of measurements in order to obtain a single composite model of the eye surface within 12mm to 22mm diameter region centered at the corneal apex, which may include regions of the ocular surface normally hidden by eyelids. Figure 3 depicts a measurement process comprising measurements of the eye 0602 fixed at three gazed directions corresponding to three different illuminated elements in the gaze fixation target array 603, 607, 611. Flat- field illumination images captured at each gaze direction are depicted in 0601 0605 and 0609.
[0038] In the abovementioned embodiments the manual or automated registration and stitching of the resulting three dimensional datasets taken at each gaze direction can be performed either by optimizing correspondence between the three dimensional shapes and features in each dataset as measured using fluorescence wavelength illumination, or by combining said optimization of correspondence with three dimensional shapes with color or monochrome intensity feature information for optical surface features such as the corneal limbus and blood vessels obtained during flat-field illumination.
[0039] In some embodiments the intensity of the fluorescence can be evaluated during or after the focusing process and a visual or auditory indication can be used to alert the operator of the state of the fluorescent dye on the ocular surface. In the case of insufficient dye the operator can be prompted to add more dye to the eye before continuing with the measurement.
[0040] During the operation of the device in all of the abovementioned embodiments the intensity of the illumination for the fluorescence and emission wavelength is selected so that it does not exceed the safety limits established by appropriate standards. [0041] In one embodiment described herein the pattern projection system can be capable of projecting one or more sequential predetermined two- dimensional patterns, whereas first and second cameras may be still or video CCD, CMOS or other cameras. The herein topography system comprises between one and three simultaneous stereo measurements.
[0042] In said embodiments, between one and three abovementioned stereo measurement methods can be used simultaneously to provide between one and three distinct measurements of the ocular surface. The obtained measurements can then be combined, averaged, or otherwise mathematically manipulated in order to increase the measurement accuracy at the entire measurement region or at a specific portion of the ocular surface. For example, the triangulation between the projector and the right camera can be used for measurement of the right side of the eye and the triangulation between the projector and the left camera can be used for measurement of the left side of the eye, while the triangulation between the two cameras can be used for the central portion of the eye, and overlap between the measurement regions can be used to achieve error reduction.
[0043] Several other embodiments can be made by adding capabilities, features, or by changing the ocular target to other objects with potentially varying surface topographies.
[0044] The measurement process for the some embodiments is depicted in Figure 4, wherein the subject's eye is prepared with fluorescent dye 0901 and positioned with respect to the optical head 0902. The computing device then calculates and checks focus and data quality metrics 0903 before initiating the formal data acquisition. Formal data acquisition takes the form of a loop 0904 wherein a frame of the structure light pattern sequence is projected onto the ocular surface 0905 and the emitted or reflected light from the ocular surface is captured by each imaging detector 0906 and transferred to the controlling computing device. At the conclusion of the formal data acquisition loop, the acquired images are processed into three dimensional topographical models of the anterior ocular surface 0907. These three dimensional topographical models are then stored in the memory of the computing device 0908 and may subsequently combined with other measurements or used to calculate optometric data products of various types. [0045] In some embodiments the encoding schemes are based on grids of parallel lines which shift in space on the projector frame during the encoding process. Such encoding sequences can be parallel vertical lines and parallel horizontal lines in sequential or alternating series where the shifting vertical lines are used to encode the column values and the shifting horizontal lines are used to encode the row values. Or the encoding sequence can be a shifting Cartesian grid composed of parallel vertical lines and parallel horizontal lines superimposed on the same pattern frame. In the latter case, the horizontal and vertical components of the pattern can be isolated algorithmically via Fourier methods.
[0046] In one embodiment, a full three dimensional topographical model of the anterior ocular surface can be obtained from a single measurement taken with the gaze direction of the eye fixed on a single fixation point. In this embodiment the lids of the eye are retracted manually by the practitioner to expose the desired extent of the scleral and corneal regions of the anterior surface to be imaged.
[0047] In other embodiments, a full three dimensional topographical model of the anterior ocular surface is created as a composite model from a plurality of individual three dimensional topographical models of the anterior ocular surface where each individual three dimensional topographical model is calculated from a measurement taken with the eye fixed on each of a plurality of fixation points.
[0048] In one such embodiment, registration of the various topography models is accomplished using feature information gleaned from flat-field illumination of the eye captured in conjunction with the coded light patterns. This can be done before or after using emission wavelength, or can be done simultaneously by using non-overlapping light source and color camera.
[0049] In said embodiment Scale Invariant Feature Transform (SIFT) and Block Match (BM) algorithms may be used to identify features on the scleral surface and create a feature based description of each model which orients it in space. Comparison of component features in each feature based model description allows direct registration of models with respect to one another. An iterative registration algorithm is then used to refine the fit, where Χ,Υ,Ζ components, and feature proximity are all used as optimization parameters. [0050] The stitching registration process by which the individual three dimensional topography models are combined into a composite three dimensional topography model of the full extents of the anterior ocular surface is depicted in Figure 5. Each of the series of individual three dimensional topography models 1301 is analyzed 1302 to compute a rough orientation vector, from which the estimated overlap regions between each model can be computed 1303, which are then used to determine the order by which models will be registered to one another 1304. In practice, models with greater overlap are registered to each other, then subsequent models are registered to the growing composite model. This maximizes the overlap between models at each registration step, increasing the reconstruction accuracy. The models comprising the series of individual three dimension topography models are then registered and stitched in a loop 1305 according to the calculated stitching order. Models u-1 1306 and u 1307 are registered to each other by using their calculated feature descriptions 1308 and 1309 to identify the subset of features in each model feature description common to both models, and computing and applying a transform 1310 from the model u to model u-1. An iterative closest points algorithm 1312 is then applied which optimizes coefficients derived from matching nearest neighbors in each model, using a coordinate space which includes X, Y, and Z three dimensional coordinates as well as color feature information as basis vectors. The stitched model then becomes model u-1, and the next model in the series according to the computed stitching order takes the role of model u. After registering all M models into a composite three dimensional topography model, additional iterations can be performed where each of the M models is compared to the composite model and it's position refined to minimize a weighting function. By tracking the weighting functions through successive iterations of looping through the M models, convergence can be determined. Figure 6 depicts an example of the stitching process results for a series of measurement comprising three gaze directions. The flat-field illumination frames from the structured light pattern illumination measurements 1401, 1403, and 1405 are included to indicate the show direction of the eye relative to the imaging detector. The individual three dimensional topographical models resulting from the measurements at each gaze direction are displayed as 1402, 1404, and 1406. The final stitched composite three dimensional topographical model is shown in 1407. [0051] In said embodiment the registration process efficiency and robustness is aided by accurate estimation of the gaze orientation vector prior to application of nearest neighbor techniques such as feature matching or iterative closest points algorithms. Accurate gaze estimation allows estimation of overlap regions between models, allowing search segmentation. Such gaze estimation is accomplished by identifying the corneal limbus by leveraging the contrast between the predominantly white sclera and the iris pigmentation. A plane can be fit to the three-dimensional limbus points, for example using a least squares criterion. The normal vector to the plane closely approximates the gaze orientation vector. The limbus plane is also used as a clipping plane to exclude features under the transparent corneal membrane, which are distorted by the optical properties of said corneal membrane, from the feature classification and matching algorithms.
[0052] In said embodiment speed and accuracy of the model registration process is enhanced by excluding all features not pertaining to the scleral surface from the feature classification and matching algorithms. A primary interfering feature is represented by eyelashes which protrude into the optical path of the imaging detector. When a valid camera-camera triangulation pair is present eyelashes can be identified by generating disparity maps based on images obtained during flat-field illumination of the eye during the measurement process.
[0053] In some embodiments, full three dimensional topographical models of the anterior ocular surface can be created either by manually retracting the lids of the eye to expose the entire portion of the ocular surface to be measured or by combining a plurality of individual three dimensional topography models each containing some segment of the entire portion of the ocular surface to be measured into a composite three dimensional topographical model of the anterior ocular surface. To produce the composite topographical model, the technology can operate in distinct modes: a Discrete Station Mode and a Continuous Processing Mode.
[0054] When said embodiment is operating in Discrete Station Mode, the operation of the device is similar to the operation described previously and described in Figure 4. In Discrete Station Mode, the steps 0902 through 0908 are operated in a loop where each iteration of the loop is realized by fixing gaze direction of the subject on a distinct element of the gaze fixation target array. The modified process is depicted in Figure 7, The image sequence comprising the data from each measurement of the ocular surface is then processed into an individual three dimensional topographical model by the computing device, and the plurality of individual three dimensional topographical models are then combined into a composite model of the full extent of the anterior ocular surface portion measured.
[0055] When said embodiment is operating in Discrete Station Mode the locations of the elements in the gaze fixation target array are chosen such that the portion of the ocular surface measured when the gaze direction is fixed on element N- 1 overlaps significantly with the portion of the ocular surface measured when the gaze direction is fixed on element N to allow optimal registration of individual models with respect to one another.
[0056] Figure 8 details the Discrete Station Mode processing tree. Scan data comprised of a sequence of images of the eye subject to a sequence of structured light patterns and flat- field illuminations is acquired for a given gaze direction 1601. A stability check 1602 is then performed by comparing flat-field illumination images comprising part of said pattern sequence and a stability metric is calculated. If the stability metric meets established criteria, the acquired data is processed through path A, including surface point triangulation by means of Coded-light reconstruction 1604 and Rasterstereographic reconstruction 1605. A preliminary feature classification is performed on each reconstruction result 1606, and weighting coefficients and surface constraints are applied to reconcile the various surface reconstructions 1607, before the flat-field illumination images are used in conjunction with SIFT and Block Match algorithms to create a detailed feature-based compliment to the three dimensional surface topology 1608 that is stored to the computer memory 1609 for future use.
[0057] If the stability metric 1602 fails to meet the established criteria, the data is processed through path B which uses a combination of slit-spline reconstruction 1610 square wave phase-shift reconstruction 1611 on individual image frames or pairs of image frames to mitigate the impact of eye motion by restricting the effective measurement windows to small fractions of the entire measurement window. A feature classification is then performed on each of the surface reconstructions 1612 using the flat-field illumination frame that preceded the structured light sequence. The process then repeats using last two frames or frame pairs in the structured light sequence in conjunction with the flat- field frame that follows the structured light sequence. Namely, frames N-1 and N of the structured light sequence are processed using the slit-spline 1613 and phase-shift reconstruction 1614 techniques and the feature classifications are performed using the flat-field frame 1615. The reconstruction results for Frames 1 and 2 are reconciled to each other using surface constraints and weighting coefficients 1616, and the same process is applied to the results for Frames N-2 and N-1, before the two reconciled results are combined and reconciled to one another 1618. A final feature classification step is performed to create the feature-based compliment 1620 and the topology and feature information are stored to the computer memory for future use 1621. In embodiments where flat- field illumination is acquired simultaneously with the structured light pattern illumination data by using color cameras and non-overlapping wavelength bands, each image frame or pair of image frames in the acquire data image sequence may be processed as a separate measurement into a separate individual three dimensional topography model with its own feature-based compliment as well, as opposed to processing only the beginning and end portions of the sequence.
[0058] An additional processing path for this embodiment operating in
Discrete Station Mode is also available for real-time control of the measurement system hardware by the attached computing device. In said path C, if the stability metric fails to meet the established criteria, the data are re-acquired 1622.
[0059] When some embodiments are operating in Continuous
Processing Mode, the data acquisition and data processing functions are directly coupled by means of the attached computing device to improve data quality and measurement speed. In Continuous Processing Mode, the gaze fixation target array is initialized by the computer, then the ocular surface is illuminated by a flat-field illumination frame and a structured light sequence truncated to one or a few frames of structured light patterns and image sequences of the illuminated ocular surface are captured by the imaging sensors. The computer processes the acquired image sequence into an individual three dimensional topography model with a feature-based compliment description, orientation vector, and model extents, then computes a desired gaze direction and updates the illumination of the gaze fixation target array. After each subsequent data acquisition, convergence, coverage, and quality metrics are calculated as part of the computation of the desired gaze direction for the next acquisition. This process continues until the calculated metrics meet established criteria, and the controlling computer algorithm ends the scan.
[0060] The operation of Continuous Processing Mode for this embodiment is detailed in Figure 9. The process begins with the subject's eye being prepared with fluorescent dye 1701 and fixed on the gaze fixation target array in its initial state 1702. The controlling algorithm enters the control loop 1703, during which it acquires image data by illuminating the eye with the structured light sequence and capturing images of the illuminate ocular surface 1704, processes the acquired image data using the Templated Grid Search algorithm 1705 and then applies one or both of Slit-spline reconstruction and square wave phase reconstruction 1706, computes the feature -based description compliment and the coverage and quality metrics 1707, and reconciles the topology and feature results 1708. For the first iteration of the loop, this reconciled topology and feature result becomes the composite model, for subsequent iterations of the loop, the composite model is updated through another reconciliation process with the newly measured components 1709. Coverage and quality metrics are then calculated 1710 and compared against established criteria 1711, after which a new gaze direction is determined and indicated on the fixation target array 1712, triggering the next scan iteration of the loop. For each measurement after the first, the new measurement is registered to the previous measurements by means of the feature-based registration stitching method described previously. The process repeats until the coverage and quality metrics meet the established convergence criteria.
[0061] In some embodiments, operating in Continuous Processing
Mode offers speed and data quality improvements by providing real-time feedback on data quality and allowing the processing computer to correct for deficiencies in the acquired data during the initial measurement process, minimizing the possibility of repeating the measurement at a later time. For embodiments which take advantage of color imaging sensors, the data quality and speed are both improved significantly by simultaneous projection of the structured light pattern sequence and flat-field illumination frame using non-overlapping wavelength bands. [0062] In some embodiments, a display screen attached to the computing device displays provides operational feedback to the user. This feedback includes real-time views of the acquired imagery for use in alignment and focusing of the measurement system with respect to the surface to be measured, intermediate stage progress indicators including focusing quality indicators, as well as visualizations of the three dimensional topographical models and optometrically useful realizations of said models and quantities derived from them.
[0063] Additional embodiments can be used to create a three dimensional model of the eye surface and electronically transmit it to the scleral and contact lens manufacturing facility for designing and building a custom lens that is specifically fit to a patient's eye.
[0064] Both the system and method embodiments disclosed herein may be used independently or in can be combined with a Placido disk based corneal topography measurement within the same embodiment. In such embodiment, the traditional Placido disk measurement approach can be used to provide a rapid corneal measurement without the fluorescent substance, while the structured light system can be used for simultaneous measurement of the ocular surface measurement in the corneal and scleral regions. Measurements made by either approach may stand alone, may be registered into a common data set to complement one another, or may be incorporated as algorithmic constraints to one or both datasets to improve the accuracy of a single composite model of the ocular surfaceSome examples of the disclosure may have more than two cameras arranged in such fashion that more than three triangulation pairs can be created during the device operation and data analysis.
Additional Examples and Embodiments
[0065] In various embodiments, the systems and methods disclosed herein may achieve one, some, or all of the following advantages and/or provide some or all of the following functionality.
[0066] In some embodiments, three simultaneous independent measurements are used, which may advantageously provide error reduction in overlap regions through averaging, error reduction in overlap regions through constraints, wider field-of-view by having cameras out at angles to the surface to be measure, and/or more accurate tracking, because it may be difficult to use corneal points for tracking.
[0067] In some embodiments, coded structured light is used to map the surface of the object (e.g., the anterior surface of the eye). Pattern sequences can offer higher spatial resolution than single patterns, and unique pixel encoding can eliminate iterative point searching, which advantageously can increase speed and/or accuracy of the mapping.
[0068] In some implementations, multiple patterns may be avoided because of eye movements. Tracking and processing segmentation can allow for correction for eye movements. In some embodiments, trimming coded light bins to single stripes at bin edges can reduce dynamic range requirements and/or exposure modification requirements of the cameras. In some embodiments, trimming coded light sequence to moving set of identical patterns can provide processing flexibility. For example, with smoothness constraints for the surface, the systems can perform true coded light methods.
[0069] Various implementations may provide multiple processing techniques. For example, the techniques can include Standard Coded Light (e.g., using all frames or subset of frames), phase-shift or scanning slit using single frames, etc. The systems and methods can be implemented to allow selection between techniques based on eye stability. Multiple techniques to constrain solutions and improve surface accuracy can be adopted. In some embodiments, use of identical patterns means each frame, or any subset, can be processed by itself using alternate methods to battle eye movement during the mapping sequence.
[0070] In various implementations, a fluorescent dye can be applied to the object (e.g., the anterior surface of the eye) to deal with differences in surface reflectivities. In some embodiments, real-time brightness/quality indicator can be used to permit analysis of the dye coverage and fluorescence intensity during focusing or mapping, and optionally, after measurement. The systems and methods can be configured to warn if scans need to be repeated.
[0071] The disclosed focusing methods of the camera and projector geometry may allow a simple focusing/alignment indication by matching the projected focusing pattern with a fixed reference display pattern.
[0072] In some implementations, the measurements in multiple wavelengths allow obtaining position and intensity (e.g., XYZI) data. Such data may allow eye movement tracking, stitching registration of multiple partial datasets (e.g., using intensity domain features to constrain algorithms for stitching smooth surfaces, limbus detection for scleral lens fitting, and/or detection of problem spots such as scaring to avoid during scleral lens fitting.
[0073] In some implementations, simultaneous pattern projections in multiple wavelengths can be used. For example, flat field and structured light can be projected simultaneously. Red, green, blue (RGB) coded light is can be used in ophthalmic or non-ophthalmic settings.
[0074] In some applications, substantially the entire sclera can be mapped by moving the gaze direction, taking partial datasets, and then combining the datasets. In one example of a Discrete Station Mode, the systems and methods utilize discrete gaze directions and the processing starts from the XYZI models collected (including third party data). In one example of Continuous Processing Mode, the systems and methods utilize automated, guided data acquisition. The gaze direction moves around in field following an indicator directed by the algorithm until convergence is obtained.
[0075] In various implementations, any of the systems and methods disclosed herein can be combined with a Placido disk for two types of independent corneal measurements.
Additional Examples of Aspects of the Disclosure
[0076] In a first aspect, a system for measuring an anterior surface topography of an eye, the system comprising: a pattern projection system configured to emit light towards an ocular surface of the eye, wherein the pattern projection system is configured to project a sequence of patterns onto the ocular surface; one or more image sensors configured to record one or more images of the patterns resulting from the projected pattern sequence, an analysis system comprising computing hardware configured to determine a topographic map of the ocular surface from the one or more images of the patterns.
[0077] In a 2nd aspect, the system of aspect 1, wherein the patterns in the sequence are projected in a single wavelength band.
[0078] In a 3rd aspect, the system of aspect 1 or aspect 2, wherein the patterns in the sequence are projected in two or more wavelength bands. [0079] In a 4th aspect, the system of any one of aspects 1-3, wherein one or more patterns in the projected sequence of patterns are emitted in a wavelength band at least partially overlapping an excitation wavelength of a fluorescent substance adapted to be applied to the eye.
[0080] In a 5th aspect, the system of any one of aspects 1-4, wherein one or more patterns in the projected sequence of patterns are emitted in a wavelength band not overlapping an excitation wavelength of a fluorescent substance adapted to be applied to the eye.
[0081] In a 6th aspect, the system of any one of aspects 1-5, wherein the pattern projection system is configured to produce structured light patterns in one or more of three modes which can be operated either simultaneously or sequentially, such that: in a first mode, an emitted wavelength range of the pattern projection system overlaps an excitation wavelength of a fluorescent substance used to prepare the ocular surface but does not overlap with a fluorescence wavelength of the fluorescent substance; in a second mode, an emitted wavelength range of the pattern projection system overlaps a fluorescence wavelength of a fluorescent substance used to prepare the ocular surface but does not overlap an excitation wavelength of the fluorescent substance; and in a third mode, an emitted wavelength range of the pattern projection system overlaps neither an excitation wavelength of a fluorescent substance used to prepare the ocular surface nor a fluorescence wavelength of the fluorescent substance.
[0082] In a 7th aspect, the system of any one of aspects 1-6, wherein the system is configured to perform a measurement of the ocular surface in a measurement duration less than about 0.5 seconds between the microsaccadic movements of the eye.
[0083] In an 8th aspect, the system of aspect 7, wherein the pattern projection system is configured to simultaneously project multiple individual patterns from the sequence of patterns in a coded light sequence, wherein each individual pattern is projected in a non-overlapping wavelength band.
[0084] In a 9th aspect, the system of aspect 8, wherein the one or more image sensors comprise a multi-color imaging detector configured to record each individual pattern in a separate recorded color channel, whereby system is configured to project and record the coded light sequence in one or more exposures. [0085] In a 10th aspect, the system of any one of aspects 1-9, wherein illumination levels incident on the ocular surface are less than 3.9 x 10"3 Joules of radiant energy as measured through a 7-mm aperture located within 5 mm of the projector focus.
[0086] In an 11th aspect, the system of any one of aspects 1-10, further comprising a fixation target system configured to permit a gaze of the eye to be sequentially fixed at a plurality of gaze directions.
[0087] In a 12th aspect, the system of aspect 11, wherein the fixation target system comprises one or both of: a plurality of targets to be illuminated in sequence or an emissive screen configured to display stationary or moving gaze fixation targets.
[0088] In a 13th aspect, the system of any one of aspects 1-12, further comprising a Placido disk corneal topographer system.
[0089] In a 14th aspect, the system of any one of aspects 1-13 wherein at least one of the one or more image sensors is configured to simultaneously record at least one of the one or more images in a plurality of wavelengths.
[0090] In a 15th aspect, the system of aspect 14, wherein the pattern projection system is configured to illuminate the ocular surface with a flat field in a first wavelength and one or more structured light patterns in second wavelength.
[0091] In a 16th aspect, the system of any one of aspects 1-15, further comprising a display device configured to display a representation of the topographic map of the ocular surface or a representation of one or more optometric values derived from the topographic map.
[0092] In a 17th aspect, the system of any one of aspects 1-16, further comprising a scleral contact lens manufacturing system, wherein the system is configured to communicate information related to the topographic map of the ocular surface to the scleral contact lens manufacturing system.
[0093] In an 18th aspect, a method for calculating a three-dimensional topographical model of an anterior ocular surface of an eye, the method comprising: under control of an ocular topographic mapping system comprising computer hardware: receiving images of projected structured light patterns that are reflected or emitted from the anterior ocular surface, the images obtained from a system comprising a plurality of imaging sensors configured to record images projected on the anterior ocular surface by a pattern projection system; analyzing the received images using one or more of the following techniques: coded light triangulation between any one of the plurality of imaging sensors and the pattern projection system, or rastersterographic triangulation between any two of the plurality of imaging sensors, slit-spline surface reconstruction, or phase-shift surface reconstruction; and determining, based at least in part on the analyzed images, a composite measurement of topography of at least a portion of the anterior ocular surface.
[0094] In a 19th aspect, the method of aspect 18, wherein determining the composite measurement of topography of at least a portion of the anterior ocular surface comprises combining a plurality of individual topography measurements of portions of the anterior surface of the eye taken at a plurality of orientations of an optical axis of the eye.
[0095] In a 20th aspect, the method of aspect 19, wherein each of the plurality of individual topography measurements is created from a series of images of projected structured light pattern sequences reflected from the anterior ocular surface, wherein each pattern sequence comprises: at least one image of the eye where the eye is illuminated in a wavelength range that overlaps a fluorescence wavelength of a fluorescent dye used to prepare the ocular surface but does not overlap an excitation wavelength of the fluorescent dye, and at least one projected structured light pattern where the projected pattern is illuminated in a wavelength range that overlaps the excitation wavelength of the fluorescent dye used to prepare the ocular surface but does not overlap the fluorescence wavelength of the fluorescent dye.
[0096] In a 21st aspect, the method of aspect 20, wherein determining the composite measurement of topography of at least a portion of the anterior ocular surface comprises: analyzing each of the plurality of individual topography measurements to provide a respective individual three-dimensional topographical model of a segment of the ocular surface, wherein the individual three-dimensional topographical model comprises three dimensional coordinate data, color intensity data, and a feature-based description of the individual three-dimensional topographical model produced from analyzing captured images of a reflected or a fluorescent pattern sequence.
[0097] In a 22nd aspect, the method of any one of aspects 19-21, wherein each of the plurality of individual topography measurements is taken in a Discrete Station Mode wherein an individual topography measurement is acquired with an optical axis of the eye directed at one of a plurality of fixed location fixation targets, wherein each individual topography measurement is created from a series of images of projected structured light pattern sequences reflected from the ocular surface, where each pattern sequence comprises: at least two flat-field images of the eye where the eye is illuminated in a wavelength range that overlaps a fluorescence wavelength of a fluorescent dye used to prepare the ocular surface but does not overlap an excitation wavelength of the fluorescent dye, and a sequence of at least one projected structured light pattern where the projected pattern is illuminated in wavelength range that overlaps the excitation wavelength of the fluorescent dye used to prepare the ocular surface but does not overlap the fluorescence wavelength of the fluorescent dye, and wherein at least one of the at least two flat-field images of the eye precedes the sequence of structured light patterns, and at least one of the at least two flat-field images of the eye follows the sequence of structured light patterns; and wherein the method further comprises processing each of the plurality of individual topography measurements into a respective individual three-dimensional topography model.
[0098] In a 23rd aspect, the method of aspect 22, further comprising analyzing at least one of the flat-field images which preceded the sequence of structured light patterns and at least one of the flat-field images which followed the structured light patterns to compute a metric describing apparent motion of the eye during a measurement period.
[0099] In a 24th aspect, the method of aspect 23, further comprising determining, based at least in part on the computed metric, at least one processing technique for constructing a respective individual three-dimensional topographical model.
[00100] In a 25th aspect, the method of any one of aspects 19-24, wherein each of the individual topography measurements is taken in Continuous Processing Mode wherein an individual topography measurement is acquired and processed in a continuous loop until a desired convergence metric or a time threshold is reached, the method further comprising: computing a composite three-dimensional topography model of the anterior ocular surface from combinations of individual three-dimensional topography measurements acquired during the continuous loop in a measurement window, while the orientation of the eye is allowed to change during the measurement window. [00101] In a 26th aspect, the method of aspect 25, wherein each of the individual topography measurements taken in Continuous Processing Mode comprises: at least one flat- field image of the eye where the eye is illuminated by a wavelength range that overlaps the fluorescence wavelength of the fluorescent dye used to prepare the ocular surface but does not overlap the excitation wavelength of the fluorescent dye, and a sequence of at least one projected structured light patterns where the projected pattern is illuminated in a wavelength range that overlaps the excitation wavelength of the fluorescent dye used to prepare the ocular surface but does not overlap the fluorescence wavelength of the fluorescent dye, the method further comprising: projecting the at least one flat-field image and the sequence of at least one projected structured light patterns at least partially overlapped in time, if a color camera is used, or projecting the at least one flat-field image and the sequence of at least one projected structured light patterns sequentially, if a monochromatic cameras is used.
[00102] In a 27th aspect, the method of aspect 26, further comprising: processing each of the individual topography measurements into a rough individual three-dimensional topography along with extents and orientation of each rough individual three-dimensional topography model; combining each individual rough three-dimensional topography model with previous rough three-dimensional topography models taken during a measurement period of a subject into a rough composite three-dimensional topography model of a measured portion of the ocular surface; evaluating the extents and surface metrics of the measured portion of the ocular surface to provide a gauge of measurement quality and completeness; and communicating the measurements of quality and completeness in real time such that a gaze direction of the subject can be adjusted to facilitate completion of the measurement of the eye of the subject.
[00103] In a 28th aspect, the method of any one of aspects 18-27, wherein the projected structured light patterns comprise grids of parallel lines or square wave patterns.
[00104] In a 29th aspect, the method of any one of aspects 18-28 wherein the projected structured light patterns are chosen to minimize frame-to-frame variation of incident intensity of illumination striking the ocular surface.
[00105] In a 30th aspect, the method of any one of aspects 18-29, wherein the one or more of the following techniques comprise at least two of the techniques, and the at least two techniques are applied to constrain or refine the composite measurement of topography.
[00106] In a 31st aspect, the method of any one of aspects 18-30, further comprising: analyzing received images of reflected flat-field illumination to compute a feature-based description of individual three-dimensional topographical models of the measured portion of the ocular surface for registering the individual three- dimensional topographical models in Cartesian space; and analyzing, based at least in part on the feature-based description, rotation and translation of an individual three- dimensional topographical model with respect to another individual three-dimensional topographical model or with respect to that same individual three-dimensional topographical model over the course of the measurement.
[00107] In a 32nd aspect, the method of aspect 31, wherein the feature- based description contains points which correspond to the corneal limbus of the eye, the method further comprising creating a masking region which excludes non- topographical features in the corneal region from the feature-based description of the individual three-dimensional topographical models to prevent optical properties of the cornea from skewing the analysis of the rotation and translation of an individual three-dimensional topographical model, and wherein a plane-fit to the corneal limbus points is used to determine an approximate orientation for the optical axis of the eye.
[00108] In a 33rd aspect, the method of any one of aspects 18-32, further comprising: receiving flat-field images captured simultaneously on more than one of the plurality of imaging sensors; and determining regions of the images that are occluded in one or more received image by protruding eyelashes.
[00109] Although descriptions of the embodiments herein have focused on measurement of the anterior surface of the human or animal eye, some embodiments of the technology may be equally applicable to the measurement of surfaces of other objects of biologic or non-biologic nature.
[00110] The presence of broadening words and phrases such as "one or more," "at least," "but not limited to" or other like phrases in some instances shall not be read to mean that the narrower case is intended or required in instances where such broadening phrases may be absent. The use of the term "optical measurement head" does not imply that the components or functionality described or claimed as part of the optical measurement head are all configured in a common package. Indeed, any or all of the various components of an optical measurement head, whether control logic or other components, can be combined in a single package or separately maintained and can further be distributed in multiple groupings or packages or across multiple locations.
[00111] Additionally, the various embodiments set forth herein are described in terms of example block diagrams, flow charts and other illustrations. As will become apparent to one of ordinary skill in the art after reading this document, the illustrated embodiments and their various alternatives can be implemented without confinement to the illustrated examples. For example, block diagrams and their accompanying description should not be construed as mandating a particular architecture or configuration.
[00112] While various embodiments of the present disclosure have been described above, it should be understood that they have been presented by way of example only, and not of limitation. Likewise, the various diagrams may depict an example architectural or other configuration for the disclosure, which is done to aid in understanding the features and functionality that can be included in the disclosure. The disclosure is not restricted to the illustrated example architectures or configurations, but the desired features can be implemented using a variety of alternative architectures and configurations. Indeed, it will be apparent to one of skill in the art how alternative functional, logical or physical partitioning and configurations can be implemented to implement the desired features of the present disclosure. Also, a multitude of different constituent module names other than those depicted herein can be applied to the various partitions. Additionally, with regard to flow diagrams, operational descriptions and method claims, the order in which the steps are presented herein shall not mandate that various embodiments be implemented to perform the recited functionality in the same order unless the context dictates otherwise.
[00113] Each of the processes, methods, and algorithms described herein and/or depicted in the attached figures may be embodied in, and fully or partially automated by, code modules executed by one or more physical computing systems, computer processors, application-specific circuitry, and/or electronic hardware configured to execute computer instructions. For example, computing systems can include general purpose computers configured with specific executable instructions for performing the disclosed methods or special purpose computers, servers, desktop computers, laptop or notebook computers or tablets, personal mobile computing devices, mobile telephones, and so forth. A code module may be stored in non- transitory computer memory, compiled and linked into an executable program, installed in a dynamic link library, or may be written in an interpreted programming language. Further, certain implementations of the functionality of the present disclosure are sufficiently mathematically, computationally, or technically complex that application-specific hardware or one or more physical computing devices (utilizing appropriate executable instructions) may be necessary to perform the functionality, for example, due to the volume or complexity of the calculations involved (e.g., computing ocular topography) or to provide results substantially in real-time.
[00114] Code modules may be stored on any type of non-transitory computer-readable medium, such as physical computer storage including hard drives, solid state memory, random access memory (RAM), read only memory (ROM), optical disc, volatile or non- volatile storage, combinations of the same and/or the like. The methods and modules may also be transmitted as generated data signals (e.g., as part of a carrier wave or other analog or digital propagated signal) on a variety of computer-readable transmission mediums, including wireless-based and wired/cable- based mediums, and may take a variety of forms (e.g., as part of a single or multiplexed analog signal, or as multiple discrete digital packets or frames). The results of the disclosed processes and process steps may be stored, persistently or otherwise, in any type of non-transitory, tangible computer storage or may be communicated via a computer-readable transmission medium.
[00115] Any processes, blocks, states, steps, or functionalities in flow diagrams described herein and/or depicted in the attached figures should be understood as potentially representing code modules, segments, or portions of code which include one or more executable instructions for implementing specific functions (e.g., logical or arithmetical) or steps in the process. The various processes, blocks, states, steps, or functionalities can be combined, rearranged, added to, deleted from, modified, or otherwise changed from the illustrative examples provided herein. In some embodiments, additional or different computing systems or code modules may perform some or all of the functionalities described herein. The methods and processes described herein are also not limited to any particular sequence, and the blocks, steps, or states relating thereto can be performed in other sequences that are appropriate, for example, in serial, in parallel, or in some other manner. Tasks or events may be added to or removed from the disclosed example embodiments. Moreover, the separation of various system components in the implementations described herein is for illustrative purposes and should not be understood as requiring such separation in all implementations. It should be understood that the described program components, methods, and systems can generally be integrated together in a single computer or software product or packaged into multiple computer or software products. Many implementation variations are possible.
[00116] The processes, methods, and systems may be implemented in a network (or distributed) computing environment. Network environments include enterprise-wide computer networks, intranets, local area networks (LAN), wide area networks (WAN), personal area networks (PAN), cloud computing networks, crowd- sourced computing networks, the Internet, and the World Wide Web. The network may be a wired or a wireless network (e.g., a terrestrial and/or satellite network) or any other type of communication network.
[00117] The various elements, features and processes described herein may be used independently of one another, or may be combined in various ways. All possible combinations and subcombinations are intended to fall within the scope of this disclosure. Further, nothing in the foregoing description is intended to imply that any particular feature, element, component, characteristic, step, module, method, process, task, or block is necessary or indispensable. The example systems and components described herein may be configured differently than described. For example, elements or components may be added to, removed from, or rearranged compared to the disclosed examples.
[00118] As used herein any reference to "one embodiment" or "some embodiments" or "an embodiment" means that a particular element, feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment. The appearances of the phrase "in one embodiment" in various places in the specification are not necessarily all referring to the same embodiment. Conditional language used herein, such as, among others, "can," "could," "might," "may," "e.g.," and the like, unless specifically stated otherwise, or otherwise understood within the context as used, is generally intended to convey that certain embodiments include, while other embodiments do not include, certain features, elements and/or steps. In addition, the articles "a" or "an" as used in this application and the appended claims are to be construed to mean "one or more" or "at least one" unless specified otherwise.
[00119] As used herein, the terms "comprises," "comprising," "includes," "including," "has," "having" or any other variation thereof, are open- ended terms and intended to cover a non-exclusive inclusion. For example, a process, method, article, or apparatus that comprises a list of elements is not necessarily limited to only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Further, unless expressly stated to the contrary, "or" refers to an inclusive or and not to an exclusive or. For example, a condition A or B is satisfied by any one of the following: A is true (or present) and B is false (or not present), A is false (or not present) and B is true (or present), or both A and B are true (or present). As used herein, a phrase referring to "at least one of a list of items refers to any combination of those items, including single members. As an example, "at least one of: A, B, or C" is intended to cover: A, B, C, A and B, A and C, B and C, and A, B, and C. Conjunctive language such as the phrase "at least one of X, Y and Z," unless specifically stated otherwise, is otherwise understood with the context as used in general to convey that an item, term, etc. may be at least one of X, Y or Z. Thus, such conjunctive language is not generally intended to imply that certain embodiments require at least one of X, at least one of Y, and at least one of Z to each be present.
[00120] Although the disclosure is described above in terms of various exemplary embodiments and implementations, it should be understood that the various features, aspects and functionality described in one or more of the individual embodiments are not limited in their applicability to the particular embodiment with which they are described, but instead can be applied, alone or in various combinations, to one or more of the other embodiments of the disclosure, whether or not such embodiments are described and whether or not such features are presented as being a part of a described embodiment. Thus, the breadth and scope of the present disclosure should not be limited by any of the above-described exemplary embodiments.
What is claimed is:
1. A system for measuring an anterior surface topography of an eye, the system comprising:
a projection system configured to emit light towards an ocular surface of the eye, wherein the projection system may contain one or more projectors, each configured to project a single pattern or a sequence of patterns onto the ocular surface;
one or more image sensors configured to record one or more images of the patterns resulting from the projected pattern or sequence,
an analysis system comprising computing hardware configured to determine a topographic map of the ocular surface from the one or more images of the patterns.
2. The system of claim 1, wherein the patterns in the sequence are projected in a single wavelength band.
3. The system of claim 1, wherein the patterns in the sequence are projected in two or more wavelength bands.
4. The system of claim 1, further comprising a fixation target system configured to permit a gaze of the eye to be sequentially fixed at a plurality of gaze directions.
5. The system of claim 4, wherein the fixation target system comprises one or both of: a plurality of targets to be illuminated in sequence or an emissive screen configured to display stationary or moving gaze fixation targets.
6. The system of claim 1, further comprising a Placido disk corneal topographer system.
7. The system of claim 1 wherein at least one of the one or more image sensors is configured to simultaneously record at least one of the one or more images in a plurality of wavelengths. 8. The system of claim 1, wherein the pattern projection system is configured to illuminate the ocular surface with a flat field in a first wavelength and one or more structured light patterns in second wavelength.
9. The system of claim 1, further comprising a scleral contact lens manufacturing system, wherein the system is configured to communicate information related to the topographic map of the ocular surface to the scleral contact lens manufacturing system.
10. A method for calculating a three-dimensional topographical model of an anterior ocular surface of an eye, the method comprising:
under control of an ocular topographic mapping system comprising computer hardware:
receiving images of projected structured light patterns that are reflected or emitted from the anterior ocular surface, the images obtained from a system comprising a single or plurality of imaging sensors configured to record images projected on the anterior ocular surface by a pattern projection system; analyzing the received image or images using one or more of the following techniques:
coded light triangulation between any one of the plurality of imaging sensors and the pattern projection system, or
rastersterographic triangulation between any two of the plurality of imaging sensors,
slit-spline surface reconstruction, or
phase-shift surface reconstruction, or
Moire pattern analysis and;
determining, based at least in part on the analyzed images, a composite measurement of topography of at least a portion of the anterior ocular surface.
11. The method of claim 10, wherein determining the composite measurement of topography of at least a portion of the anterior ocular surface comprises combining a plurality of individual topography measurements of portions of the anterior surface of the eye taken at a plurality of orientations of an optical axis of the eye.
12. The method of claim 11, wherein each of the plurality of individual topography measurements is created from one or series of images of projected structured light pattern sequences reflected from the anterior ocular surface, wherein each pattern sequence comprises:
at least one image of the eye where the eye is illuminated in a wavelength range that overlaps a fluorescence wavelength of a fluorescent dye used to prepare the ocular surface but does not overlap an excitation wavelength of the fluorescent dye, and
at least one projected structured light pattern where the projected pattern is illuminated in a wavelength range that overlaps the excitation wavelength of the fluorescent dye used to prepare the ocular surface but does not overlap the fluorescence wavelength of the fluorescent dye.
13. The method of claim 12, wherein determining the composite measurement of topography of at least a portion of the anterior ocular surface comprises:
analyzing each of the plurality of individual topography measurements to provide a respective individual three-dimensional topographical model of a segment of the ocular surface.
14. The method of claim 13, wherein the individual three-dimensional topographical model comprises three dimensional coordinate data, color intensity data, and a feature-based description of the individual three-dimensional topographical model produced from analyzing captured images of a reflected or a fluorescent pattern sequence.
15. The method of claim 14, wherein each of the plurality of individual topography measurements is taken in a Discrete Station Mode wherein an individual topography measurement is acquired with an optical axis of the eye directed at one of a plurality of fixed location fixation targets, wherein each individual topography measurement is created from a series of images of projected structured light pattern sequences reflected from the ocular surface, where each pattern sequence comprises:
at least two flat-field images of the eye where the eye is illuminated in a wavelength range that overlaps a fluorescence wavelength of a fluorescent dye used to prepare the ocular surface but does not overlap an excitation wavelength of the fluorescent dye, and
a sequence of at least one projected structured light pattern where the projected pattern is illuminated in wavelength range that overlaps the excitation wavelength of the fluorescent dye used to prepare the ocular surface but does not overlap the fluorescence wavelength of the fluorescent dye, and wherein at least one of the at least two flat-field images of the eye precedes the sequence of structured light patterns, and at least one of the at least two flat-field images of the eye follows the sequence of structured light patterns; and
wherein the method further comprises processing each of the plurality of individual topography measurements into a respective individual three- dimensional topography model.
16. The method of claim 15, further comprising analyzing at least one of the flat-field images which preceded the sequence of structured light patterns and at least one of the flat-field images which followed the structured light patterns to compute a metric describing apparent motion of the eye during a measurement period.
17. The method of claim 11, wherein the projected structured light patterns comprise grids of parallel lines or square wave patterns.
18. The method of claim 11, wherein the one or more of the following techniques comprise at least two of the techniques, and the at least two techniques are applied to constrain or refine the composite measurement of topography.
19. The method of claim 11, further comprising:
analyzing received images of reflected flat-field illumination to compute a feature-based description of individual three-dimensional topographical models of the measured portion of the ocular surface for registering the individual three-dimensional topographical models in Cartesian space; and
analyzing, based at least in part on the feature -based description, rotation and translation of an individual three-dimensional topographical model with respect to another individual three-dimensional topographical model or with respect to that same individual three-dimensional topographical model over the course of the measurement.
20. The method of claim 18, wherein the feature -based description contains points which correspond to the corneal limbus of the eye, the method further comprising creating a masking region which excludes non-topographical features in the corneal region from the feature -based description of the individual three- dimensional topographical models to prevent optical properties of the cornea from skewing the analysis of the rotation and translation of an individual three-dimensional topographical model, and wherein a plane-fit to the corneal limbus points is used to determine an approximate orientation for the optical axis of the eye.

Claims

What is claimed is:
1. A system for measuring an anterior surface topography of an eye, the system comprising:
a projection system configured to emit light towards an ocular surface of the eye, wherein the projection system may contain one or more projectors, each configured to project a single pattern or a sequence of patterns onto the ocular surface;
one or more image sensors configured to record one or more images of the patterns resulting from the projected pattern or sequence,
an analysis system comprising computing hardware configured to determine a topographic map of the ocular surface from the one or more images of the patterns.
2. The system of claim 1 , wherein the patterns in the sequence are projected in a single wavelength band.
3. The system of claim 1 , wherein the patterns in the sequence are proj ected in two or more wavelength bands.
4. The system of claim 1, further comprising a fixation target system configured to permit a gaze of the eye to be sequentially fixed at a plurality of gaze directions.
5. The system of claim 4, wherein the fixation target system comprises one or both of: a plurality of targets to be illuminated in sequence or an emissive screen configured to display stationary or moving gaze fixation targets.
6. The system of claim 1, further comprising a Placido disk corneal topographer system.
7. The system of claim 1 wherein at least one of the one or more image sensors is configured to simultaneously record at least one of the one or more images in a plurality of wavelengths.
8. The system of claim 1, wherein the pattern projection system is configured to illuminate the ocular surface with a flat field in a first wavelength and one or more structured light patterns in second wavelength.
9. The system of claim 1, further comprising a scleral contact lens manufacturing system, wherein the system is configured to communicate information related to the topographic map of the ocular surface to the scleral contact lens manufacturing system.
10. A method for calculating a three-dimensional topographical model of an anterior ocular surface of an eye, the method comprising:
under control of an ocular topographic mapping system comprising computer hardware:
receiving images of projected structured light patterns that are reflected or emitted from the anterior ocular surface, the images obtained from a system comprising a single or plurality of imaging sensors configured to record images projected on the anterior ocular surface by a pattern projection system;
analyzing the received image or images using one or more of the following techniques:
coded light triangulation between any one of the plurality of imaging sensors and the pattern projection system, or
rastersterographic triangulation between any two of the plurality of imaging sensors,
slit-spline surface reconstruction, or
phase-shift surface reconstruction, or
Moire pattern analysis and;
determining, based at least in part on the analyzed images, a composite measurement of topography of at least a portion of the anterior ocular surface.
11. The method of claim 10, wherein determining the composite measurement of topography of at least a portion of the anterior ocular surface comprises combining a plurality of individual topography measurements of portions of the anterior surface of the eye taken at a plurality of orientations of an optical axis of the eye.
12. The method of claim 11, wherein each of the plurality of individual topography measurements is created from one or series of images of projected structured light pattern sequences reflected from the anterior ocular surface, wherein each pattern sequence comprises:
at least one image of the eye where the eye is illuminated in a wavelength range that overlaps a fluorescence wavelength of a fluorescent dye used to prepare the ocular surface but does not overlap an excitation wavelength of the fluorescent dye, and
at least one projected structured light pattern where the projected pattern is illuminated in a wavelength range that overlaps the excitation wavelength of the fluorescent dye used to prepare the ocular surface but does not overlap the fluorescence wavelength of the fluorescent dye.
13. The method of claim 12, wherein determining the composite measurement of topography of at least a portion of the anterior ocular surface comprises:
analyzing each of the plurality of individual topography measurements to provide a respective individual three-dimensional topographical model of a segment of the ocular surface.
14. The method of claim 13, wherein the individual three-dimensional topographical model comprises three dimensional coordinate data, color intensity data, and a feature-based description of the individual three-dimensional topographical model produced from analyzing captured images of a reflected or a fluorescent pattern sequence.
15. The method of claim 14, wherein each of the plurality of individual topography measurements is taken in a Discrete Station Mode wherein an individual topography measurement is acquired with an optical axis of the eye directed at one of a plurality of fixed location fixation targets, wherein each individual topography measurement is created from a series of images of projected structured light pattern sequences reflected from the ocular surface, where each pattern sequence comprises:
at least two flat-field images of the eye where the eye is illuminated in a wavelength range that overlaps a fluorescence wavelength of a fluorescent dye used to prepare the ocular surface but does not overlap an excitation wavelength of the fluorescent dye, and
a sequence of at least one projected structured light pattern where the projected pattern is illuminated in wavelength range that overlaps the excitation wavelength of the fluorescent dye used to prepare the ocular surface but does not overlap the fluorescence wavelength of the fluorescent dye, and
wherein at least one of the at least two flat-field images of the eye precedes the sequence of structured light patterns, and at least one of the at least two flat-field images of the eye follows the sequence of structured light patterns; and
wherein the method further comprises processing each of the plurality of individual topography measurements into a respective individual three- dimensional topography model.
16. The method of claim 15, further comprising analyzing at least one of the flat-field images which preceded the sequence of structured light patterns and at least one of the flat-field images which followed the structured light patterns to compute a metric describing apparent motion of the eye during a measurement period.
17. The method of claim 11, wherein the projected structured light patterns comprise grids of parallel lines or square wave patterns.
18. The method of claim 11, wherein the one or more of the following techniques comprise at least two of the techniques, and the at least two techniques are applied to constrain or refine the composite measurement of topography.
19. The method of claim 11, further comprising:
analyzing received images of reflected flat-field illumination to compute a feature-based description of individual three-dimensional topographical models of the measured portion of the ocular surface for registering the individual three-dimensional topographical models in Cartesian space; and analyzing, based at least in part on the feature-based description, rotation and translation of an individual three-dimensional topographical model with respect to another individual three-dimensional topographical model or with respect to that same individual three-dimensional topographical model over the course of the measurement.
20. The method of claim 18, wherein the feature-based description contains points which correspond to the corneal limbus of the eye, the method further comprising creating a masking region which excludes non-topographical features in the corneal region from the feature-based description of the individual three-dimensional topographical models to prevent optical properties of the cornea from skewing the analysis of the rotation and translation of an individual three-dimensional topographical model, and wherein a plane-fit to the corneal limbus points is used to determine an approximate orientation for the optical axis of the eye.
PCT/US2016/015579 2015-01-30 2016-01-29 Systems and method for mapping the ocular surface usually obstructed by the eyelids WO2016123448A2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201562110362P 2015-01-30 2015-01-30
US62/110,362 2015-01-30

Publications (2)

Publication Number Publication Date
WO2016123448A2 true WO2016123448A2 (en) 2016-08-04
WO2016123448A3 WO2016123448A3 (en) 2016-10-27

Family

ID=56544533

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2016/015579 WO2016123448A2 (en) 2015-01-30 2016-01-29 Systems and method for mapping the ocular surface usually obstructed by the eyelids

Country Status (1)

Country Link
WO (1) WO2016123448A2 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108564626A (en) * 2018-03-19 2018-09-21 百度在线网络技术(北京)有限公司 Method and apparatus for determining the relative attitude angle being installed between the camera of acquisition entity
CN113784655A (en) * 2019-03-14 2021-12-10 梅德蒙特国际有限公司 Ophthalmic imaging method, apparatus and system
WO2023118782A1 (en) * 2021-12-21 2023-06-29 University Of Strathclyde Optical measurement apparatus

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
NL9200071A (en) * 1992-01-15 1993-08-02 Stichting Science Park Maastri DEVICE FOR DETERMINING THE TOPOGRAPHY OF A CURVED SURFACE.
WO2002082980A2 (en) * 2001-04-16 2002-10-24 Lasersight Technologies, Inc. Stereoscopic measurement of cornea and illumination patterns
US7111938B2 (en) * 2001-04-27 2006-09-26 Novartis Ag Automatic lens design and manufacturing system
EP1391176A1 (en) * 2002-08-16 2004-02-25 Universiteit Maastricht Method and arrangement for performing measurements of the topography of a corneal surface
AU2002952772A0 (en) * 2002-11-20 2002-12-05 Clearmark Technologies Pty Ltd A corneal topographer
US8226235B2 (en) * 2006-02-14 2012-07-24 Vision Optimization, Llc Method and apparatus for determining dynamic deformation characteristics of an object
US20090275929A1 (en) * 2008-04-30 2009-11-05 Amo Development, Llc System and method for controlling measurement in an eye during ophthalmic procedure
US8888286B2 (en) * 2009-04-01 2014-11-18 Tearscience, Inc. Full-eye illumination ocular surface imaging of an ocular tear film for determining tear film thickness and/or providing ocular topography
CA2852671C (en) * 2011-10-17 2021-03-30 Eyedeal Scanning, Llc Method and apparatus for determining eye topography
US9545200B2 (en) * 2013-11-08 2017-01-17 Precision Ocular Metrology, L.L.C. Systems and methods for mapping the ocular surface

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108564626A (en) * 2018-03-19 2018-09-21 百度在线网络技术(北京)有限公司 Method and apparatus for determining the relative attitude angle being installed between the camera of acquisition entity
CN108564626B (en) * 2018-03-19 2021-08-31 百度在线网络技术(北京)有限公司 Method and apparatus for determining relative pose angle between cameras mounted to an acquisition entity
CN113784655A (en) * 2019-03-14 2021-12-10 梅德蒙特国际有限公司 Ophthalmic imaging method, apparatus and system
EP3937754A4 (en) * 2019-03-14 2023-04-19 Medmont International Pty Ltd. Ophthalmological imaging method, device and system
WO2023118782A1 (en) * 2021-12-21 2023-06-29 University Of Strathclyde Optical measurement apparatus

Also Published As

Publication number Publication date
WO2016123448A3 (en) 2016-10-27

Similar Documents

Publication Publication Date Title
US9545200B2 (en) Systems and methods for mapping the ocular surface
JP6586523B2 (en) Eye tracking using structured light
JP6429772B2 (en) 3D scanning and positioning system
JP5571864B1 (en) System and method for telemetry of optical focus
KR101094766B1 (en) Apparatus and mehtod for tracking eye
JP2016508799A (en) Apparatus for optical coherence tomography of the eye and method for optical coherence tomography of the eye
JP2001512241A (en) Scanning apparatus and method
JP7279711B2 (en) IMAGE PROCESSING METHOD, PROGRAM, IMAGE PROCESSING APPARATUS, AND OPHTHALMIC SYSTEM
JP6841863B2 (en) OCT retinal image data rendering method, storage medium storing OCT retinal image data rendering program, OCT retinal image data rendering program and OCT retinal image data rendering device
US10561311B2 (en) Ophthalmic imaging apparatus and ophthalmic information processing apparatus
JP2023009530A (en) Image processing method, image processing device, and program
JP2022523306A (en) Eye tracking devices and methods
CN105395163A (en) Ophthalmologic Apparatus And Ophthalmologic Apparatus Control Method
WO2021106987A1 (en) Medical image processing device, optical coherence tomography device, medical image processing method, and program
WO2016123448A2 (en) Systems and method for mapping the ocular surface usually obstructed by the eyelids
JP6900994B2 (en) Line-of-sight detection device and line-of-sight detection method
JP2023060069A (en) Image processing method, image processing device, image processing program and blood vessel diameter calculation device
JP3711053B2 (en) Line-of-sight measurement device and method, line-of-sight measurement program, and recording medium recording the program
CN114189623B (en) Light field-based refraction pattern generation method, device, equipment and storage medium
US10674902B2 (en) Information processing apparatus, operation method thereof, and computer program
JP6558161B2 (en) Ophthalmic apparatus and image processing program
WO2022024104A1 (en) Eye tracking systems and methods
JP2021058285A (en) Image processing device, image processing method and program
Chaudhuri et al. Depth recovery from defocused images

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16744166

Country of ref document: EP

Kind code of ref document: A2

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 16744166

Country of ref document: EP

Kind code of ref document: A2