US8240852B2 - Image processing method and image processing device - Google Patents

Image processing method and image processing device Download PDF

Info

Publication number
US8240852B2
US8240852B2 US12/930,739 US93073911A US8240852B2 US 8240852 B2 US8240852 B2 US 8240852B2 US 93073911 A US93073911 A US 93073911A US 8240852 B2 US8240852 B2 US 8240852B2
Authority
US
United States
Prior art keywords
region
disc
image processing
cup
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US12/930,739
Other versions
US20110176108A1 (en
Inventor
Toshiaki Nakagawa
Shigeru Takimoto
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Kowa Co Ltd
Original Assignee
Kowa Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Kowa Co Ltd filed Critical Kowa Co Ltd
Assigned to KOWA COMPANY LTD. reassignment KOWA COMPANY LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: NAKAGAWA, TOSHIAKI, TAKIMOTO, SHIGERU
Publication of US20110176108A1 publication Critical patent/US20110176108A1/en
Application granted granted Critical
Publication of US8240852B2 publication Critical patent/US8240852B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/12Edge-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • G06T2207/10012Stereo images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20112Image segmentation details
    • G06T2207/20168Radial search
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30041Eye; Retina; Ophthalmic

Definitions

  • the present invention relates to an image processing method and an image processing device for processing images of the optic disc region of a photographed ocular fundus under examination.
  • image processing which evaluates the shape and size of the optic disc (hereinafter referred to simply as the disc or disc region) in an image of the ocular fundus.
  • the magnitude of optic disc cupping When the magnitude of optic disc cupping is to be evaluated via image analysis, they define the diameter ratio C/D of the diameter of the optic disc cupping, called the “cup”, and the diameter of the optic disc, called the “disc” (see Japanese Patent No. 3594468).
  • the medial region lying between the disc region and the cup region is termed the “rim” and observation of the width of this rim is one method for the examiner to diagnose glaucoma.
  • an image of the ocular fundus containing the disc region is displayed, and the disc region is specified manually in the displayed image using a mouse, keyboard, input stylus or the like.
  • the cup region is then computed automatically from the height thereof.
  • Japanese Laid-Open Patent Application 2008-73188 also proposes a method of automatically computing the cup line (cup contour) from the degree of bending of the vessels in the ocular fundus image.
  • Japanese Laid-Open Patent Application 2006-280411 discloses a method in which the disc line (disc contour) and cup line (cup contour) are computed using a spline approach to derive therefrom the C/D ratio, which is compared with the C/D ratio from the measurements with a table of C/D ratios for normal eyes created in advance in order to diagnose ophthalmic disorders.
  • the present invention provides an image processing method for processing an image of an optic disc region in a photographed ocular fundus, comprising extracting a disc region; dividing the disc region into a plurality of regions by a plurality of dividing lines that extend radially from the center of gravity of the disc region; radially scanning the divided regions from the center of gravity while angles are shifted in order to detect in each individual region a point at which luminance variation reaches a maximum; and determining a contour line of the disc region based on the detected points at which luminance variation reaches a maximum.
  • the present invention further provides an image processing method for processing an image of an optic disc region in a photographed ocular fundus, comprising dividing a disc region into a plurality of regions by a plurality of dividing lines that extend radially from a point that is set in the disc region; radially scanning the divided regions from the set point while angles are shifted in order to detect in each individual region a point at which depth variation reaches a maximum; and determining a contour line of the cup region based on the detected points at which depth variation reaches a maximum.
  • a disc region is divided into a plurality of regions by a plurality of dividing lines that extend radially from a prescribed point.
  • the divided regions are then scanned radially from the established point while angles are shifted in order to detect in each region a point at which luminance variation or depth variation reaches a maximum, and the contour line of the disc region or cup region is determined from these points. This allows the disc contours or cup contours to be determined automatically, and assisted diagnosis of glaucoma to be carried out more accurately and efficiently.
  • FIG. 1 is a block diagram showing a configuration of an image processing system that the present invention employs
  • FIG. 2 is an illustrative view showing operation buttons on the display of the image processing system
  • FIG. 3 is a flowchart showing the flow of the entire image processing process of the present invention.
  • FIG. 4 is a flowchart showing the flow of extraction of disc contours
  • FIG. 5 is a flowchart showing the flow of extraction of cup contours
  • FIG. 6 a is an illustrative view showing a method in which scan lines are used to search a disc region having a standard disc area
  • FIG. 6 b is an illustrative view showing a luminance distribution of image-processed regions along the scan lines
  • FIG. 7 is an illustrative view showing a method in which a divided disc detection region is angularly scanned to derive therefrom maximum luminance variation
  • FIG. 8 is an illustrative view showing a method of deriving luminance variation
  • FIG. 9 a is an illustrative view showing a method in which scan lines are used to search a cup region having a standard cup area
  • FIG. 9 b is an illustrative view showing a depth distribution of image-processed regions along the scan lines
  • FIG. 10 is an illustrative view showing a method of angular scanning of a divided cup detection region, and deriving maximum depth variation
  • FIG. 11 is an illustrative view showing a method of deriving depth variation
  • FIG. 12 is an illustrative view showing an extracted disc region and cup region
  • FIG. 13 is an ocular fundus view showing a disc detection region together with blood vessel images
  • FIG. 14 is an ocular fundus view showing extracted points of maximum luminance variation, shown together with blood vessel images.
  • FIG. 15 is an ocular fundus view showing extracted points of maximum depth variation, shown together with blood vessel images.
  • the present invention will now be described in detail with reference to the embodiment, which shows an ophthalmic measurement apparatus in which the ocular fundus under examination is stereographically photographed using a stereographical imaging optical system, and the photographed image undergoes three-dimensional measurement processing.
  • FIG. 1 shows a configuration of an ophthalmic measurement apparatus showing one embodiment of the present invention.
  • symbol 101 denotes a fundus camera.
  • the fundus camera In order to photograph the ocular fundus under examination (not shown) under prescribed photographic conditions, the fundus camera is equipped with mechanisms such as, for example, an alignment mechanism, a monocular photography mechanism, and a stereographic photography mechanism.
  • the fundus camera 101 has an image pickup device for color photography, such as a three-plate CCD or CMOS sensor for example, and outputs ocular fundus color image data of the eye under examination as digital data to an image processing device 100 .
  • the picture signal output by the fundus camera 101 has a format such as YUV (YPbPr or YCbCr)
  • a color separation process is carried out to convert the image data to that of a different spectrum such as RGB image data.
  • Such a color separation process would be necessary when the fundus camera 101 outputs images in a format such as JPEG (or MPEG), for example.
  • the image processing device 100 is constituted by hardware such as a PC, for example.
  • the image processing device 100 carries out control of the system as a whole, and also includes a CPU 102 that constitutes the principal image processing means for carrying out image processing, which will be discussed later.
  • the image processing device 100 could instead be constituted by dedicated hardware integrated in the fundus camera 101 .
  • Image processing is executed using a VRAM (image memory) 104 serving as the work area.
  • An additional memory for use in system control apart from image processing is a memory constituted by dynamic RAM or the like.
  • a program for execution by the CPU 102 to carry out the image processing is stored in a ROM 103 or a hard disk (HDD) 105 .
  • the hard disk 105 is used for storage of captured image data of an eye under examination, numeric data such as measurement results, output image data generated by the image processing, and the like.
  • a display 107 composed of an LCD or EL panel, a CRT, or the like is connected to the image processing device 100 , and a user interface screen or output screen is displayed on the display 107 for the purpose of controlling image processing by the image processing device 100 .
  • the image processing device 100 is provided with user interface means composed of a keyboard 108 and a mouse or other pointing device 109 .
  • the image processing device 100 performs image processing, generates image data enabling the examiner to easily determine the ocular fundus under examination, particularly the cup region and disc region, and outputs the data to the display 107 .
  • a network 106 is connected to the image processing device 100 via a network interface (not shown).
  • the image processing device 100 outputs the aforementioned captured image data of the eye under examination, numeric data such as measurement results, output image data generated by the image processing, and the like to an external computer, or to a separate image processing device or ophthalmic measurement apparatus, or the like.
  • FIGS. 3 to 5 The flow of the image processing procedure carried out by the image processing device 100 is illustrated in FIGS. 3 to 5 .
  • the program by which the CPU 102 carries out image processing in the image processing device 100 is stored in the ROM 103 or on the hard disk 105 , and the CPU 102 loads and executes this program to carry out image processing according to the present embodiment.
  • FIG. 3 shows the overall flow.
  • the fundus camera 101 has taken a monocular color image of the eye under examination, or taken color stereo images of a pair of parallax images which allows stereoscopic viewing of the ocular fundus; and that the captured ocular fundus images are stored in the VRAM 104 or the hard disk 105 .
  • Step S 1 of FIG. 3 an automatic extraction button is pushed to automatically extract the disc region or cup region of the optic disc from the ocular fundus image. This is accomplished by pushing the “Auto” button 208 of the “Region Selection” of the screen 200 on the display 107 in FIG. 2 .
  • Step S 2 if the disc region has not yet been determined, automatic extraction of the disc region is performed. The user then checks the extracted region, performs correction if necessary, and pushes the “Determine Contour” button 211 .
  • Step S 3 if only the cup region has not yet been determined, automatic extraction of the cup region is performed. The user checks the extracted region, performs correction if necessary, and pushes the “Determine Contour” button 211 .
  • Automatic extraction of the disc region or the cup region may be accomplished by respectively pushing the “Disc” button 201 or the “Cup” button 202 shown in FIG. 2 , then pushing the “Auto” button 208 .
  • the user may check the region, perform correction if necessary, and then push the “Determine Contour” button 211 to determine these contours independently.
  • the determined disc region or cup region can be saved to the hard disk 105 by pushing the “Save” button 210 . Also, by pushing the “Back” button 207 , it is possible to load previously saved contour data, and to call the data up on the display 107 in order to make modifications thereto. By pressing the “Delete” button 209 it is possible to delete editing data etc. To cancel some or all image processing, the “Cancel” button 212 is pushed. In the event that multiple displays are connected, the “Select Screen” button can be used to select the display on which the image is to be displayed. The “User” button can be used to select a user. A “Depth computation” screen is provided as well, and the parameters such as the patient's diopter, corneal curvature, and the like can be selected during depth computation.
  • FIG. 4 shows the flow of automatic extraction of a disc contour, i.e. the periphery of the disc region.
  • An ocular fundus image obtained by monocular photography, or one parallax image, for example, the left eye image of a pair of parallax ocular fundus images obtained by stereographic photography is read from the hard disk 105 as the ocular fundus image for disc contour extraction and is saved in the VRAM 104 .
  • Step S 11 a process such as morphology is used to erase blood vessel images from the ocular fundus image in the VRAM 104 , and color separation to RGB images is performed.
  • the R image mostly contains information from relatively deep portions of the retina, e.g. from the choroid membrane; the G image mostly contains information from the retinal pigment epithelium; and the B image mostly contains information from the retinal surface.
  • Step S 12 the RGB images are scanned for a region of high luminance in proximity to the disc region.
  • FIG. 6 a shows scanning along scan lines 302 within a window 301 established in proximity to the disc region of an R image 300 for example.
  • FIG. 6 b shows detection of pixels having greater luminance values than a prescribed luminance threshold value 304 . If pixels having greater luminance values than the luminance threshold value 304 are detected, it is checked whether the area of the region defined by those pixels is equivalent to or approximately equivalent to a predetermined standard disc area. In the example of FIG. 6 a , the pixels of a region 303 indicated by halftone dots have been detected as pixels that have greater luminance values than the luminance threshold value.
  • the region 303 defined by the detected pixels and indicated by halftone dots constitutes a region of luminance greater than a prescribed luminance threshold value.
  • the area of the luminance region increases as the luminance threshold value 304 indicated by the solid line in FIG. 6 b decreases to the luminance threshold value 304 ′ indicated by the dot and dash line.
  • the image is scanned while varying the luminance threshold value, and a luminance region which has luminance greater than a prescribed luminance threshold value and whose area is equivalent to a predetermined standard disc area is extracted as a region having the standard disc area.
  • the region having the standard disc area extracted in this manner is a region close to the disc region that is ultimately desired. Therefore, the region thus extracted is stored in the VRAM 104 as the disc detection region or detected disc region.
  • Step S 12 the disc region having the standard disc area was extracted from each of the RGB images, and the region of overlap of these detected images was designated as the disc detection region.
  • Step S 13 the disc detection region extracted in Step S 13 is indicated by symbol 11 and shown together with a color ocular fundus image 10 .
  • the erased blood vessel images 12 are restored for reference purposes.
  • Step S 14 the center of gravity of the disc detection region 11 , is calculated, and the disc detection region 11 is divided into eight regions by line segments that extend radially at equiangular intervals (of 45 degrees) from this center.
  • FIG. 14 shows the center of gravity 13 of the disc detection region 11 calculated in Step S 14 and the eight regions S 1 to S 8 divided by eight dividing lines 14 a to 14 h which extend from the center of gravity 13 .
  • each of the eight divided regions S 1 to S 8 is scanned along scan lines extending radially from the center of gravity 13 in order to detect a point at which luminance varies maximally, i.e. luminance variation reaches a maximum.
  • FIG. 7 shows scanning of one divided region S 1 along equiangular scan lines 17 extending radially from the center of gravity 13 . Scanning is performed radially from the center of gravity 13 at resolution of from 1 to 2 degrees for example.
  • FIG. 8 shows luminance variation observed when the disc detection region 11 is scanned along a single angular scan line.
  • the vertical axis in FIG. 8 indicates luminance, and the horizontal axis indicates distance from the center of gravity 13 along the scan line 17 .
  • This angular scanning is performed on the region S 1 to extract the maximum luminance variation point on each scan line. For example, where angular scanning takes place at resolution of 1 degree, 45 maximum luminance variation points will be extracted in the region S 1 . The one with the highest value of these 45 points is extracted as the disc contour point in the region S 1 , and is saved to the VRAM 104 .
  • a maximum luminance variation point 15 a detected on angular scan line 17 a is the point having the highest value of the maximum luminance variation points in the region S 1 . Consequently, the point 15 a is extracted as the disc contour point in the region S 1 .
  • This angular scanning is carried out for each of the regions S 1 to S 8 .
  • Disc contour points, namely maximum luminance variation points extracted from the regions S 1 to S 8 are shown as 15 a to 15 h in FIG. 14 .
  • Step S 15 The process described above is the process of Step S 15 .
  • Step S 16 the eight extracted points 15 a to 15 h are joined by a third- or fourth-order spline curve, for example, to derive a disc contour line.
  • the region enclosed by this contour line is designated as the disc region.
  • FIG. 12 shows a disc region 500 and a disc contour line 501 determined in this manner.
  • FIG. 5 shows the flow of a procedure for automatically extracting the cup contour, i.e., the periphery of the cup region.
  • the ocular fundus image from which the cup contour is extracted is the same image as that used for disc contour extraction.
  • Step S 21 a process such as morphology is used as well to erase blood vessel images and color separation into RGB images is performed.
  • parallax is extracted from a pair of parallax images of each component of the RGB images, and on the basis of corresponding pixels in the pair of parallax images, depth data for pixels inside a region in proximity to the disc region is measured using a known method, and the measured depth data is saved to the VRAM 104 .
  • Step S 22 the RGB images are scanned for a region of great depth, i.e. a deep region in proximity to the disc region.
  • FIG. 9 a shows scanning along scan lines 302 within a window 301 established in proximity to the disc region of an R image 300 for example.
  • FIG. 9 b shows detection of pixels deeper than a prescribed depth threshold value 305 . Pixels deeper than the depth threshold value are detected while successively varying the depth threshold value 305 from larger values to smaller values (or the reverse), and it is checked whether the area of the region defined by those pixels is equivalent to or approximately equivalent to a predetermined standard cup area.
  • FIG. 9 a shows scanning along scan lines 302 within a window 301 established in proximity to the disc region of an R image 300 for example.
  • FIG. 9 b shows detection of pixels deeper than a prescribed depth threshold value 305 . Pixels deeper than the depth threshold value are detected while successively varying the depth threshold value 305 from larger values to smaller values (or the reverse), and it is checked whether the area of the region defined by
  • the pixels of a region 306 represented by halftone dots have been detected as pixels deeper than the luminance threshold value 305 , and the region 306 defined by the detected pixels and represented by halftone dots constitutes a region of greater depth than the depth threshold value.
  • the area of the region decreases as the depth threshold value 305 indicated by the solid line in FIG. 9 b decreases to the depth threshold value 305 ′ indicated by the dotted-dashed line.
  • the image is scanned while the depth threshold value is varied, and a depth region having a depth greater than a prescribed depth threshold value and having a area equivalent to a predetermined standard cup area is extracted as a region having the standard cup area.
  • the region having the standard cup area thus extracted is a region close to the cup region that is ultimately desired, so that the extracted region is stored in the VRAM 104 as the cup detection region or detected cup region.
  • FIG. 10 shows an extracted cup detection region 20 , represented by a broken line.
  • Step S 22 a cup region of standard cup area was extracted from each of the RGB images, and the region of overlap of these detected images was designated as the cup detection region.
  • Step S 24 the cup detection region 20 is divided into eight regions by line segments extending radially at equiangular intervals (of 45 degrees) from the center of gravity 13 that was calculated for the disc detection region.
  • each of the eight divided regions S 1 to S 8 is scanned along scan lines extending radially from the center of gravity 13 to detect a point at which depth variation reaches a maximum.
  • FIG. 10 shows scanning of one divided region S 1 along equiangular scan lines 18 extending radially from the center of gravity 13 . Scanning is performed radially from the center of gravity 13 at resolution of from 1 to 2 degrees for example.
  • FIG. 11 shows depth variation observed when the cup detection region 20 is scanned along a single angular scan line.
  • the vertical axis in FIG. 11 indicates depth from the retina, and the horizontal axis indicates distance from the center of gravity 13 along the scan line 18 .
  • This angular scanning is performed on the region S 1 to extract the maximum depth variation point on each scan line. For example, where angular scanning takes place at resolution of 1 degree, 45 maximum depth points will be extracted in the region S 1 . Therefore, the one with the highest value of these 45 points is extracted as the cup contour point in the region S 1 , and is saved to the VRAM 104 .
  • a maximum depth variation point 16 a detected on angular scan line 18 a is the maximum depth variation point in the region S 1 . Consequently, the point 16 a is extracted as the cup contour point in the region S 1 .
  • Cup contour points namely maximum depth variation points extracted from the regions S 1 to S 8 , are shown as 16 a to 16 h in FIG. 15 .
  • Step S 25 The process described above is the process of Step S 25 .
  • Step S 26 the eight extracted points 16 a to 16 h are joined by a third- or fourth-order spline curve for example, to derive a cup contour line.
  • the region enclosed by this contour line is designated as the cup region.
  • FIG. 12 shows a cup region 510 and a cup contour line 511 determined in this manner.
  • the disk region 500 and the contour line 501 thereof as well as the cup region 510 and the contour line 511 thereof, which are obtained by image processing as described above, may be saved as appended information to the ocular fundus image and recorded to a recording medium such as the hard disk 105 . Saved disc regions and cup regions may be displayed in time series on the display 107 . Also, saved disc contour lines and cup contour lines may be displayed on the display 107 to allow for correction of these contour lines if necessary.
  • the disc detection region 11 and the cup detection region 20 herein are divided into eight parts by equiangular dividing lines extending from the center, the number of divisions is not limited to 8; some other plural number of divisions, such as 6 divisions or 12 divisions, is also acceptable.
  • the angular resolution of the scan lines need not be set to a fixed resolution.
  • the angle interval for scanning may vary according to angle region, for example, using finer angular intervals for the scan lines 17 , 18 in angle regions in which the contour lines of the disc region or the cup region are more complex, and coarser ones in angle regions in which the contour lines of the disc region or the cup region are fairly simple.
  • the center of gravity of the disk region is selected when determining the contour line of a cup region, and the disk region and the cup region are divided into a plurality of regions by a plurality of dividing lines extending radially from the center of gravity.
  • the disk region and the cup region may then be divided into a plurality of regions by a plurality of dividing lines extending radially from this selected point or center.
  • the standard disc area means an area of the disc region serving as a reference, and may be the average value of measured area of a large number of disc regions
  • the standard cup area means an area of the cup region serving as a reference, and may be the average value of the measured area of a large number of cup regions.
  • a luminance region having a greater luminance value than a predetermined standard luminance value may be extracted as the disc region
  • Step S 23 a region of greater depth than a predetermined standard depth value may be extracted as the cup region.
  • a luminance region or depth region may be extracted in each of images obtained by color separation of an image into RGB images, and the region of overlap thereof may be selected as the disc region or cup region.
  • the luminance region or depth region may be detected from the image prior to color separation into RGB images, and the disc region or cup region may be extracted therefrom.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Eye Examination Apparatus (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

An ocular fundus image is captured by a fundus camera. Blood vessel regions are erased from the ocular fundus image, and a luminance region greater than a prescribed luminance threshold value and having an area equivalent to a predetermined standard disc area is extracted as a region having a standard disc area. The region having the standard disc area is divided into a plurality of regions by a plurality of dividing lines extending radially from the center of gravity thereof, and divided regions are scanned radially from the center of gravity while angles are shifted to detect in each individual region a point at which luminance variation reaches a maximum. The contour line of the disc region is derived from these points.

Description

BACKGROUND OF THE INVENTION
1. Field of the Invention
The present invention relates to an image processing method and an image processing device for processing images of the optic disc region of a photographed ocular fundus under examination.
2. Description of the Related Art
In the field of methods for diagnosing the ophthalmic disorder of glaucoma, image processing is known which evaluates the shape and size of the optic disc (hereinafter referred to simply as the disc or disc region) in an image of the ocular fundus.
When the magnitude of optic disc cupping is to be evaluated via image analysis, they define the diameter ratio C/D of the diameter of the optic disc cupping, called the “cup”, and the diameter of the optic disc, called the “disc” (see Japanese Patent No. 3594468).
The medial region lying between the disc region and the cup region is termed the “rim” and observation of the width of this rim is one method for the examiner to diagnose glaucoma.
According to Japanese Patent No. 3594468, an image of the ocular fundus containing the disc region is displayed, and the disc region is specified manually in the displayed image using a mouse, keyboard, input stylus or the like. The cup region is then computed automatically from the height thereof.
Japanese Laid-Open Patent Application 2008-73188 also proposes a method of automatically computing the cup line (cup contour) from the degree of bending of the vessels in the ocular fundus image.
Japanese Laid-Open Patent Application 2006-280411 discloses a method in which the disc line (disc contour) and cup line (cup contour) are computed using a spline approach to derive therefrom the C/D ratio, which is compared with the C/D ratio from the measurements with a table of C/D ratios for normal eyes created in advance in order to diagnose ophthalmic disorders.
Document, “Application in Auto-recognition and Pseudo-stereoscopic Image Generation of the Optic Disc Using Images with Blood Vessels Erased for a System of Assisted Diagnosis of Ocular Fundus Images”, IEICE Transactions, D, Vol. J89-D, No. 11, pp. 2491-2501, 2006, also proposes a method of erasing blood vessel regions from a captured ocular fundus image, automatically recognizing the optic disc using the image from which the blood vessel regions are erased, and generating a pseudo-stereoscopic image.
SUMMARY OF THE INVENTION
When the C/D ratio is employed for diagnosing the disorder of glaucoma, it is necessary to determine the cup contour and the disc contour. However, in the prior art, contour points of these were determined manually using an input device such as a mouse, and thus the determination involved some subjectivity, resulting in individual variability. Individual variability is particularly great in the cup region, which significantly affects diagnosis of glaucoma and poses a risk of misdiagnosis.
It is therefore an object of the present invention to provide an image processing method and an image processing device for automatically determining cup contour and disc contour and for minimizing individual variability (subjectivity) in computation of the C/D ratio so that assisted diagnosis of glaucoma may be carried out more accurately and efficiently.
The present invention provides an image processing method for processing an image of an optic disc region in a photographed ocular fundus, comprising extracting a disc region; dividing the disc region into a plurality of regions by a plurality of dividing lines that extend radially from the center of gravity of the disc region; radially scanning the divided regions from the center of gravity while angles are shifted in order to detect in each individual region a point at which luminance variation reaches a maximum; and determining a contour line of the disc region based on the detected points at which luminance variation reaches a maximum.
The present invention further provides an image processing method for processing an image of an optic disc region in a photographed ocular fundus, comprising dividing a disc region into a plurality of regions by a plurality of dividing lines that extend radially from a point that is set in the disc region; radially scanning the divided regions from the set point while angles are shifted in order to detect in each individual region a point at which depth variation reaches a maximum; and determining a contour line of the cup region based on the detected points at which depth variation reaches a maximum.
According to the present invention, a disc region is divided into a plurality of regions by a plurality of dividing lines that extend radially from a prescribed point.
The divided regions are then scanned radially from the established point while angles are shifted in order to detect in each region a point at which luminance variation or depth variation reaches a maximum, and the contour line of the disc region or cup region is determined from these points. This allows the disc contours or cup contours to be determined automatically, and assisted diagnosis of glaucoma to be carried out more accurately and efficiently.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a block diagram showing a configuration of an image processing system that the present invention employs;
FIG. 2 is an illustrative view showing operation buttons on the display of the image processing system;
FIG. 3 is a flowchart showing the flow of the entire image processing process of the present invention;
FIG. 4 is a flowchart showing the flow of extraction of disc contours;
FIG. 5 is a flowchart showing the flow of extraction of cup contours;
FIG. 6 a is an illustrative view showing a method in which scan lines are used to search a disc region having a standard disc area;
FIG. 6 b is an illustrative view showing a luminance distribution of image-processed regions along the scan lines;
FIG. 7 is an illustrative view showing a method in which a divided disc detection region is angularly scanned to derive therefrom maximum luminance variation;
FIG. 8 is an illustrative view showing a method of deriving luminance variation;
FIG. 9 a is an illustrative view showing a method in which scan lines are used to search a cup region having a standard cup area;
FIG. 9 b is an illustrative view showing a depth distribution of image-processed regions along the scan lines;
FIG. 10 is an illustrative view showing a method of angular scanning of a divided cup detection region, and deriving maximum depth variation;
FIG. 11 is an illustrative view showing a method of deriving depth variation;
FIG. 12 is an illustrative view showing an extracted disc region and cup region;
FIG. 13 is an ocular fundus view showing a disc detection region together with blood vessel images;
FIG. 14 is an ocular fundus view showing extracted points of maximum luminance variation, shown together with blood vessel images; and
FIG. 15 is an ocular fundus view showing extracted points of maximum depth variation, shown together with blood vessel images.
DESCRIPTION OF THE PREFERRED EMBODIMENTS
The present invention will now be described in detail with reference to the embodiment, which shows an ophthalmic measurement apparatus in which the ocular fundus under examination is stereographically photographed using a stereographical imaging optical system, and the photographed image undergoes three-dimensional measurement processing.
<System Configuration>
FIG. 1 shows a configuration of an ophthalmic measurement apparatus showing one embodiment of the present invention. In FIG. 1, symbol 101 denotes a fundus camera. In order to photograph the ocular fundus under examination (not shown) under prescribed photographic conditions, the fundus camera is equipped with mechanisms such as, for example, an alignment mechanism, a monocular photography mechanism, and a stereographic photography mechanism. The fundus camera 101 has an image pickup device for color photography, such as a three-plate CCD or CMOS sensor for example, and outputs ocular fundus color image data of the eye under examination as digital data to an image processing device 100. When the picture signal output by the fundus camera 101 has a format such as YUV (YPbPr or YCbCr), a color separation process is carried out to convert the image data to that of a different spectrum such as RGB image data.
Such a color separation process would be necessary when the fundus camera 101 outputs images in a format such as JPEG (or MPEG), for example.
The image processing device 100 is constituted by hardware such as a PC, for example. The image processing device 100 carries out control of the system as a whole, and also includes a CPU 102 that constitutes the principal image processing means for carrying out image processing, which will be discussed later. Of course, the image processing device 100 could instead be constituted by dedicated hardware integrated in the fundus camera 101.
Image processing is executed using a VRAM (image memory) 104 serving as the work area. An additional memory for use in system control apart from image processing is a memory constituted by dynamic RAM or the like. A program for execution by the CPU 102 to carry out the image processing is stored in a ROM 103 or a hard disk (HDD) 105.
The hard disk 105 is used for storage of captured image data of an eye under examination, numeric data such as measurement results, output image data generated by the image processing, and the like.
As display output means, a display 107 composed of an LCD or EL panel, a CRT, or the like is connected to the image processing device 100, and a user interface screen or output screen is displayed on the display 107 for the purpose of controlling image processing by the image processing device 100. For the purpose of carrying out these screen displays and overall control of the instrument, the image processing device 100 is provided with user interface means composed of a keyboard 108 and a mouse or other pointing device 109.
The image processing device 100 performs image processing, generates image data enabling the examiner to easily determine the ocular fundus under examination, particularly the cup region and disc region, and outputs the data to the display 107.
A network 106 is connected to the image processing device 100 via a network interface (not shown). The image processing device 100 outputs the aforementioned captured image data of the eye under examination, numeric data such as measurement results, output image data generated by the image processing, and the like to an external computer, or to a separate image processing device or ophthalmic measurement apparatus, or the like.
<Image Processing>
The flow of the image processing procedure carried out by the image processing device 100 is illustrated in FIGS. 3 to 5. The program by which the CPU 102 carries out image processing in the image processing device 100 is stored in the ROM 103 or on the hard disk 105, and the CPU 102 loads and executes this program to carry out image processing according to the present embodiment.
FIG. 3 shows the overall flow. As preconditions for image processing, it is assumed that the fundus camera 101 has taken a monocular color image of the eye under examination, or taken color stereo images of a pair of parallax images which allows stereoscopic viewing of the ocular fundus; and that the captured ocular fundus images are stored in the VRAM 104 or the hard disk 105.
In Step S1 of FIG. 3, an automatic extraction button is pushed to automatically extract the disc region or cup region of the optic disc from the ocular fundus image. This is accomplished by pushing the “Auto” button 208 of the “Region Selection” of the screen 200 on the display 107 in FIG. 2.
In Step S2, if the disc region has not yet been determined, automatic extraction of the disc region is performed. The user then checks the extracted region, performs correction if necessary, and pushes the “Determine Contour” button 211. In Step S3, if only the cup region has not yet been determined, automatic extraction of the cup region is performed. The user checks the extracted region, performs correction if necessary, and pushes the “Determine Contour” button 211.
Automatic extraction of the disc region or the cup region may be accomplished by respectively pushing the “Disc” button 201 or the “Cup” button 202 shown in FIG. 2, then pushing the “Auto” button 208. The user may check the region, perform correction if necessary, and then push the “Determine Contour” button 211 to determine these contours independently.
In either case, the determined disc region or cup region can be saved to the hard disk 105 by pushing the “Save” button 210. Also, by pushing the “Back” button 207, it is possible to load previously saved contour data, and to call the data up on the display 107 in order to make modifications thereto. By pressing the “Delete” button 209 it is possible to delete editing data etc. To cancel some or all image processing, the “Cancel” button 212 is pushed. In the event that multiple displays are connected, the “Select Screen” button can be used to select the display on which the image is to be displayed. The “User” button can be used to select a user. A “Depth computation” screen is provided as well, and the parameters such as the patient's diopter, corneal curvature, and the like can be selected during depth computation.
FIG. 4 shows the flow of automatic extraction of a disc contour, i.e. the periphery of the disc region.
An ocular fundus image obtained by monocular photography, or one parallax image, for example, the left eye image of a pair of parallax ocular fundus images obtained by stereographic photography is read from the hard disk 105 as the ocular fundus image for disc contour extraction and is saved in the VRAM 104.
In Step S11, a process such as morphology is used to erase blood vessel images from the ocular fundus image in the VRAM 104, and color separation to RGB images is performed. The R image mostly contains information from relatively deep portions of the retina, e.g. from the choroid membrane; the G image mostly contains information from the retinal pigment epithelium; and the B image mostly contains information from the retinal surface.
Next, in Step S12, the RGB images are scanned for a region of high luminance in proximity to the disc region. FIG. 6 a shows scanning along scan lines 302 within a window 301 established in proximity to the disc region of an R image 300 for example. FIG. 6 b shows detection of pixels having greater luminance values than a prescribed luminance threshold value 304. If pixels having greater luminance values than the luminance threshold value 304 are detected, it is checked whether the area of the region defined by those pixels is equivalent to or approximately equivalent to a predetermined standard disc area. In the example of FIG. 6 a, the pixels of a region 303 indicated by halftone dots have been detected as pixels that have greater luminance values than the luminance threshold value. The region 303 defined by the detected pixels and indicated by halftone dots constitutes a region of luminance greater than a prescribed luminance threshold value. The area of the luminance region increases as the luminance threshold value 304 indicated by the solid line in FIG. 6 b decreases to the luminance threshold value 304′ indicated by the dot and dash line.
In this way, the image is scanned while varying the luminance threshold value, and a luminance region which has luminance greater than a prescribed luminance threshold value and whose area is equivalent to a predetermined standard disc area is extracted as a region having the standard disc area. The region having the standard disc area extracted in this manner is a region close to the disc region that is ultimately desired. Therefore, the region thus extracted is stored in the VRAM 104 as the disc detection region or detected disc region.
Extraction of such a region is performed on the RGB images. The three regions, i.e. the region of standard disc area extracted from the R image, the region of standard disc area extracted from the G image, and the region of standard disc area extracted from the B image, are superimposed in order to extract the overlapping region, which is then saved as the final disc detection region in the VRAM 104. This is the process of Step S13.
In Step S12, the disc region having the standard disc area was extracted from each of the RGB images, and the region of overlap of these detected images was designated as the disc detection region. However, it is also acceptable to extract a disc region having the standard disc area from the color image prior to color separation to RGB, and to use this as the disc detection region.
In FIG. 13, the disc detection region extracted in Step S13 is indicated by symbol 11 and shown together with a color ocular fundus image 10. In the drawing, the erased blood vessel images 12 are restored for reference purposes.
Next, in Step S14, the center of gravity of the disc detection region 11, is calculated, and the disc detection region 11 is divided into eight regions by line segments that extend radially at equiangular intervals (of 45 degrees) from this center.
FIG. 14 shows the center of gravity 13 of the disc detection region 11 calculated in Step S14 and the eight regions S1 to S8 divided by eight dividing lines 14 a to 14 h which extend from the center of gravity 13.
Next, each of the eight divided regions S1 to S8 is scanned along scan lines extending radially from the center of gravity 13 in order to detect a point at which luminance varies maximally, i.e. luminance variation reaches a maximum.
FIG. 7 shows scanning of one divided region S1 along equiangular scan lines 17 extending radially from the center of gravity 13. Scanning is performed radially from the center of gravity 13 at resolution of from 1 to 2 degrees for example.
FIG. 8 shows luminance variation observed when the disc detection region 11 is scanned along a single angular scan line. The vertical axis in FIG. 8 indicates luminance, and the horizontal axis indicates distance from the center of gravity 13 along the scan line 17. Pn (n=1, 2, 3, . . . , N−2, N−1, N) denotes pixels of an ocular fundus image. Luminance variation V in a pixel of interest P, can be represented as V=Dn−w−Dn, where Dn is the luminance value in the pixel Pn and W is a calculation range determined by the image resolution (e.g. W=1 where image resolution is 0.0053 mm/pixel). This allows the pixel point at which V reaches maximum to be extracted. In the example shown in FIG. 8, luminance variation reaches a maximum at pixel P, and therefore this pixel is extracted as the maximum luminance variation point on the scan line concerned.
This angular scanning is performed on the region S1 to extract the maximum luminance variation point on each scan line. For example, where angular scanning takes place at resolution of 1 degree, 45 maximum luminance variation points will be extracted in the region S1. The one with the highest value of these 45 points is extracted as the disc contour point in the region S1, and is saved to the VRAM 104.
In the example shown in FIG. 7, a maximum luminance variation point 15 a detected on angular scan line 17 a is the point having the highest value of the maximum luminance variation points in the region S1. Consequently, the point 15 a is extracted as the disc contour point in the region S1.
This angular scanning is carried out for each of the regions S1 to S8. Disc contour points, namely maximum luminance variation points extracted from the regions S1 to S8 are shown as 15 a to 15 h in FIG. 14.
The process described above is the process of Step S15. In the following Step S16, the eight extracted points 15 a to 15 h are joined by a third- or fourth-order spline curve, for example, to derive a disc contour line. The region enclosed by this contour line is designated as the disc region. FIG. 12 shows a disc region 500 and a disc contour line 501 determined in this manner.
FIG. 5 shows the flow of a procedure for automatically extracting the cup contour, i.e., the periphery of the cup region.
The ocular fundus image from which the cup contour is extracted is the same image as that used for disc contour extraction. In Step S21, a process such as morphology is used as well to erase blood vessel images and color separation into RGB images is performed.
For cup contour extraction, parallax is extracted from a pair of parallax images of each component of the RGB images, and on the basis of corresponding pixels in the pair of parallax images, depth data for pixels inside a region in proximity to the disc region is measured using a known method, and the measured depth data is saved to the VRAM 104.
In Step S22, the RGB images are scanned for a region of great depth, i.e. a deep region in proximity to the disc region. FIG. 9 a shows scanning along scan lines 302 within a window 301 established in proximity to the disc region of an R image 300 for example. FIG. 9 b shows detection of pixels deeper than a prescribed depth threshold value 305. Pixels deeper than the depth threshold value are detected while successively varying the depth threshold value 305 from larger values to smaller values (or the reverse), and it is checked whether the area of the region defined by those pixels is equivalent to or approximately equivalent to a predetermined standard cup area. In the example of FIG. 9 a, the pixels of a region 306 represented by halftone dots have been detected as pixels deeper than the luminance threshold value 305, and the region 306 defined by the detected pixels and represented by halftone dots constitutes a region of greater depth than the depth threshold value. The area of the region decreases as the depth threshold value 305 indicated by the solid line in FIG. 9 b decreases to the depth threshold value 305′ indicated by the dotted-dashed line.
In this way, the image is scanned while the depth threshold value is varied, and a depth region having a depth greater than a prescribed depth threshold value and having a area equivalent to a predetermined standard cup area is extracted as a region having the standard cup area. The region having the standard cup area thus extracted is a region close to the cup region that is ultimately desired, so that the extracted region is stored in the VRAM 104 as the cup detection region or detected cup region.
Extraction of such a region is performed on the RGB images. The three regions, i.e. the region of standard cup area extracted from the R image, the region of standard cup area extracted from the G image, and the region of standard cup area extracted from the B image, are superimposed, and the region of overlap is extracted as the final cup detection region, which is then saved to the VRAM 104. This is the process of Step S23. FIG. 10 shows an extracted cup detection region 20, represented by a broken line.
In Step S22, a cup region of standard cup area was extracted from each of the RGB images, and the region of overlap of these detected images was designated as the cup detection region. However, it is also acceptable to extract a cup region of standard cup area from the color image prior to color separation to RGB, and to designate this as the cup detection region.
Next, in Step S24, the cup detection region 20 is divided into eight regions by line segments extending radially at equiangular intervals (of 45 degrees) from the center of gravity 13 that was calculated for the disc detection region.
Next, each of the eight divided regions S1 to S8 is scanned along scan lines extending radially from the center of gravity 13 to detect a point at which depth variation reaches a maximum.
FIG. 10 shows scanning of one divided region S1 along equiangular scan lines 18 extending radially from the center of gravity 13. Scanning is performed radially from the center of gravity 13 at resolution of from 1 to 2 degrees for example.
FIG. 11 shows depth variation observed when the cup detection region 20 is scanned along a single angular scan line. The vertical axis in FIG. 11 indicates depth from the retina, and the horizontal axis indicates distance from the center of gravity 13 along the scan line 18. Pn (n=1, 2, 3, . . . , N−2, N−1, N) denotes pixels of an ocular fundus image. Depth variation U in a pixel of interest Pn′ can be represented as U=Dn+w−Dn−w, where Dn is depth data in the pixel Pn and W is a calculation range determined by the image resolution (e.g. W=2 where image resolution is 0.0053 mm/pixel). This allows the pixel point at which U reaches maximum to be extracted. In the example shown in FIG. 11, depth variation reaches maximum at pixel P′, and therefore this pixel is extracted as the maximum depth variation point on the scan line concerned.
This angular scanning is performed on the region S1 to extract the maximum depth variation point on each scan line. For example, where angular scanning takes place at resolution of 1 degree, 45 maximum depth points will be extracted in the region S1. Therefore, the one with the highest value of these 45 points is extracted as the cup contour point in the region S1, and is saved to the VRAM 104.
In the example shown in FIG. 10, a maximum depth variation point 16 a detected on angular scan line 18 a is the maximum depth variation point in the region S1. Consequently, the point 16 a is extracted as the cup contour point in the region S1.
This angular scanning is carried out for each of the regions S1 to S8. Cup contour points, namely maximum depth variation points extracted from the regions S1 to S8, are shown as 16 a to 16 h in FIG. 15.
The process described above is the process of Step S25. In the following Step S26, the eight extracted points 16 a to 16 h are joined by a third- or fourth-order spline curve for example, to derive a cup contour line. The region enclosed by this contour line is designated as the cup region. FIG. 12 shows a cup region 510 and a cup contour line 511 determined in this manner.
The disk region 500 and the contour line 501 thereof as well as the cup region 510 and the contour line 511 thereof, which are obtained by image processing as described above, may be saved as appended information to the ocular fundus image and recorded to a recording medium such as the hard disk 105. Saved disc regions and cup regions may be displayed in time series on the display 107. Also, saved disc contour lines and cup contour lines may be displayed on the display 107 to allow for correction of these contour lines if necessary.
While the disc detection region 11 and the cup detection region 20 herein are divided into eight parts by equiangular dividing lines extending from the center, the number of divisions is not limited to 8; some other plural number of divisions, such as 6 divisions or 12 divisions, is also acceptable. The angular resolution of the scan lines need not be set to a fixed resolution. The angle interval for scanning may vary according to angle region, for example, using finer angular intervals for the scan lines 17, 18 in angle regions in which the contour lines of the disc region or the cup region are more complex, and coarser ones in angle regions in which the contour lines of the disc region or the cup region are fairly simple.
According to the embodiment described above, the center of gravity of the disk region is selected when determining the contour line of a cup region, and the disk region and the cup region are divided into a plurality of regions by a plurality of dividing lines extending radially from the center of gravity. However, one may instead select, as the point set within the disc region for the purpose of dividing the disk region and the cup region, the deepest point inside the disc region, or the center of a region such that the area deeper than a prescribed depth threshold value inside the disc region is equivalent to a predetermined standard cup area. The disk region and the cup region may then be divided into a plurality of regions by a plurality of dividing lines extending radially from this selected point or center.
In the embodiment described above, the standard disc area means an area of the disc region serving as a reference, and may be the average value of measured area of a large number of disc regions, and the standard cup area means an area of the cup region serving as a reference, and may be the average value of the measured area of a large number of cup regions.
In Step S13, a luminance region having a greater luminance value than a predetermined standard luminance value may be extracted as the disc region, and in Step S23, a region of greater depth than a predetermined standard depth value may be extracted as the cup region. In this case, a luminance region or depth region may be extracted in each of images obtained by color separation of an image into RGB images, and the region of overlap thereof may be selected as the disc region or cup region. Alternatively, the luminance region or depth region may be detected from the image prior to color separation into RGB images, and the disc region or cup region may be extracted therefrom.

Claims (14)

1. An image processing method for processing an image of an optic disc region in a photographed ocular fundus, comprising:
extracting a disc region;
dividing the disc region into a plurality of regions by a plurality of dividing lines that extend radially from the center of gravity of the disc region;
radially scanning the divided regions from the center of gravity while angles are shifted in order to detect in each individual region a point at which luminance variation reaches a maximum; and
determining a contour line of the disc region based on the detected points at which luminance variation reaches a maximum.
2. An image processing method according to claim 1, wherein a luminance region having a greater luminance value than a prescribed luminance threshold value and having an area equivalent to a predetermined standard disc area is extracted as the disc region.
3. An image processing method according to claim 1, wherein a spline curve is used to connect the points of maximum luminance variation, thereby creating the disc contour line and determining the disc region.
4. An image processing method according to claim 1, wherein the created disc contour line is displayed on a display so that it can be corrected.
5. An image processing method according to claim 1, wherein the extracted disc region is saved as appended information for an ocular fundus image, and disc regions can be displayed in a time series.
6. An image processing method for processing an image of an optic disc region in a photographed ocular fundus, comprising:
dividing a disc region into a plurality of regions by a plurality of dividing lines that extend radially from a point that is set in the disc region;
radially scanning the divided regions from the set point while angles are shifted in order to detect in each individual region a point at which depth variation reaches a maximum; and
determining a contour line of the cup region based on the detected points at which depth variation reaches a maximum.
7. An image processing method according to claim 6, wherein the point set within the disc region is a center of gravity of the disc region.
8. An image processing method according to claim 6, wherein the point set within the disc region is the deepest point within the disc region.
9. An image processing method according to claim 6, wherein the point set within the disc region is the center of gravity of a region that is deeper than a prescribed depth threshold value inside the disc region, and whose area is equivalent to a predetermined standard cup area.
10. An image processing method according to claim 6, wherein a spline curve is used to connect the points of maximum depth variation, thereby creating the cup contour line and determining the cup region.
11. An image processing method according to claim 6, wherein the cup contour line is displayed on a display so that it can be corrected.
12. An image processing method according to claim 6, wherein the extracted cup region is saved as appended information for an ocular fundus image, and the cup regions can be displayed in a time series.
13. An image processing device comprising an ocular fundus imaging device for carrying out the image processing method according to claim 1, and image processing means for processing an image of a photographed ocular fundus.
14. An image processing device comprising an ocular fundus imaging device for carrying out the image processing method according to claim 6, and image processing means for processing an image of a photographed ocular fundus.
US12/930,739 2010-01-20 2011-01-14 Image processing method and image processing device Expired - Fee Related US8240852B2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2010010292A JP5583980B2 (en) 2010-01-20 2010-01-20 Image processing method and image processing apparatus
JP2010-010292 2010-01-20

Publications (2)

Publication Number Publication Date
US20110176108A1 US20110176108A1 (en) 2011-07-21
US8240852B2 true US8240852B2 (en) 2012-08-14

Family

ID=43909633

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/930,739 Expired - Fee Related US8240852B2 (en) 2010-01-20 2011-01-14 Image processing method and image processing device

Country Status (3)

Country Link
US (1) US8240852B2 (en)
EP (1) EP2360641B1 (en)
JP (1) JP5583980B2 (en)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9370328B2 (en) * 2012-11-29 2016-06-21 University Of Washington Through Its Center For Commercialization Methods and systems for determining tumor boundary characteristics
JP2014113422A (en) * 2012-12-12 2014-06-26 Canon Inc Ophthalmological photographing apparatus, and control method and program of ophthalmological photographing apparatus
CN104434027B (en) * 2013-09-12 2016-03-02 广东福地新视野光电技术有限公司 The method and apparatus of optic disc area is obtained by ophthalmic optical coherence tomoscan image
DE202014106268U1 (en) 2014-12-23 2015-02-25 Toll Collect Gmbh Vehicle equipment for DSRC communication
US10554956B2 (en) 2015-10-29 2020-02-04 Dell Products, Lp Depth masks for image segmentation for depth-based computational photography
US10021371B2 (en) 2015-11-24 2018-07-10 Dell Products, Lp Method and apparatus for gross-level user and input detection using similar or dissimilar camera pair
CN109993765B (en) * 2019-04-09 2020-10-30 东莞理工学院 Method for detecting retinal vein cross compression angle
CN110443256B (en) * 2019-07-03 2022-04-12 大连理工大学 Method for extracting multi-target regions of image
CN112001920B (en) * 2020-10-28 2021-02-05 北京至真互联网技术有限公司 Fundus image recognition method, device and equipment

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7524061B2 (en) * 2005-10-12 2009-04-28 Siemens Corporate Research, Inc. System and method for robust optic disk detection in retinal images using vessel structure and radon transform
US7854510B2 (en) * 2008-10-16 2010-12-21 Steven Roger Verdooner Apparatus and method for imaging the eye

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3508112B2 (en) * 1993-09-21 2004-03-22 株式会社トプコン Fundus disease analyzer
JP3585331B2 (en) * 1996-12-03 2004-11-04 株式会社ニデック Analysis method of fundus stereoscopic image
JP3594468B2 (en) * 1997-11-21 2004-12-02 株式会社ニデック Fundus image analysis method
JP4790295B2 (en) 2005-03-31 2011-10-12 株式会社ニデック Method of operating the analyzer
JP5007420B2 (en) 2006-09-21 2012-08-22 タック株式会社 Image analysis system and image analysis program

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7524061B2 (en) * 2005-10-12 2009-04-28 Siemens Corporate Research, Inc. System and method for robust optic disk detection in retinal images using vessel structure and radon transform
US7854510B2 (en) * 2008-10-16 2010-12-21 Steven Roger Verdooner Apparatus and method for imaging the eye

Also Published As

Publication number Publication date
EP2360641A1 (en) 2011-08-24
EP2360641B1 (en) 2019-07-24
JP2011147547A (en) 2011-08-04
JP5583980B2 (en) 2014-09-03
US20110176108A1 (en) 2011-07-21

Similar Documents

Publication Publication Date Title
US8240852B2 (en) Image processing method and image processing device
CN101404923B (en) Image processing device for medical use
US6714672B1 (en) Automated stereo fundus evaluation
US10945637B2 (en) Image based jaundice diagnosing method and apparatus and image based jaundice diagnosis assisting apparatus
KR101998595B1 (en) Method and Apparatus for jaundice diagnosis based on an image
US20110243415A1 (en) Image processing apparatus, control method thereof, and program
US20070188705A1 (en) Ocular fundus portion analyzer and ocular fundus portion analyzing method
CN102697446B (en) Image processing apparatus and image processing method
JP4915737B2 (en) Image analysis system and image analysis program
JP5478103B2 (en) Image processing method
US10796418B2 (en) Image processing apparatus, image processing method, and program
JP5111845B2 (en) Fundus image processing apparatus, fundus photographing apparatus, and program
JP2014198223A (en) Information processor and information processing method
JP2020201823A (en) Image processing device, image processing method, and program
JP2008237840A (en) Image analysis system and image analysis program
WO2015037316A1 (en) Organ-imaging device and organ-imaging method
JP2000245700A (en) Instrument for measuring eyeground and recording medium for recording measurement program
JP2014102713A (en) Face component extraction device, face component extraction method, and program
EP4185184B1 (en) Method for determining a coronal position of an eye relative to the head
JP2006074498A (en) Image processor and imaging apparatus
JP2006324727A (en) Imaging apparatus and image processing method thereof
JP2012186821A (en) Face image processing device, face image processing method, electronic still camera, digital image processing device and digital image processing method
JP4949007B2 (en) Fundus image processing apparatus, fundus photographing apparatus, and program
CN110584588A (en) Method and device for detecting visual field defect
EP4349268A1 (en) Determination device, determination method, and program

Legal Events

Date Code Title Description
AS Assignment

Owner name: KOWA COMPANY LTD., JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:NAKAGAWA, TOSHIAKI;TAKIMOTO, SHIGERU;REEL/FRAME:025703/0383

Effective date: 20110111

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

ZAAA Notice of allowance and fees due

Free format text: ORIGINAL CODE: NOA

ZAAB Notice of allowance mailed

Free format text: ORIGINAL CODE: MN/=.

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Free format text: PAYER NUMBER DE-ASSIGNED (ORIGINAL EVENT CODE: RMPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20240814