US8240852B2 - Image processing method and image processing device - Google Patents
Image processing method and image processing device Download PDFInfo
- Publication number
- US8240852B2 US8240852B2 US12/930,739 US93073911A US8240852B2 US 8240852 B2 US8240852 B2 US 8240852B2 US 93073911 A US93073911 A US 93073911A US 8240852 B2 US8240852 B2 US 8240852B2
- Authority
- US
- United States
- Prior art keywords
- region
- disc
- image processing
- cup
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related, expires
Links
- 238000012545 processing Methods 0.000 title claims description 48
- 238000003672 processing method Methods 0.000 title claims description 19
- 230000005484 gravity Effects 0.000 claims abstract description 22
- 238000003384 imaging method Methods 0.000 claims description 3
- 210000004204 blood vessel Anatomy 0.000 abstract description 10
- 238000001514 detection method Methods 0.000 description 24
- 238000000034 method Methods 0.000 description 23
- 238000000605 extraction Methods 0.000 description 12
- 238000000926 separation method Methods 0.000 description 8
- 238000005259 measurement Methods 0.000 description 7
- 208000010412 Glaucoma Diseases 0.000 description 6
- 238000012937 correction Methods 0.000 description 4
- 238000003745 diagnosis Methods 0.000 description 4
- 230000007246 mechanism Effects 0.000 description 4
- 230000007423 decrease Effects 0.000 description 3
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 3
- 210000001525 retina Anatomy 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 238000005452 bending Methods 0.000 description 1
- 210000003161 choroid Anatomy 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 239000012528 membrane Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
- 230000002207 retinal effect Effects 0.000 description 1
- 210000003583 retinal pigment epithelium Anatomy 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/12—Edge-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/136—Segmentation; Edge detection involving thresholding
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
- G06T2207/10012—Stereo images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20112—Image segmentation details
- G06T2207/20168—Radial search
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30041—Eye; Retina; Ophthalmic
Definitions
- the present invention relates to an image processing method and an image processing device for processing images of the optic disc region of a photographed ocular fundus under examination.
- image processing which evaluates the shape and size of the optic disc (hereinafter referred to simply as the disc or disc region) in an image of the ocular fundus.
- the magnitude of optic disc cupping When the magnitude of optic disc cupping is to be evaluated via image analysis, they define the diameter ratio C/D of the diameter of the optic disc cupping, called the “cup”, and the diameter of the optic disc, called the “disc” (see Japanese Patent No. 3594468).
- the medial region lying between the disc region and the cup region is termed the “rim” and observation of the width of this rim is one method for the examiner to diagnose glaucoma.
- an image of the ocular fundus containing the disc region is displayed, and the disc region is specified manually in the displayed image using a mouse, keyboard, input stylus or the like.
- the cup region is then computed automatically from the height thereof.
- Japanese Laid-Open Patent Application 2008-73188 also proposes a method of automatically computing the cup line (cup contour) from the degree of bending of the vessels in the ocular fundus image.
- Japanese Laid-Open Patent Application 2006-280411 discloses a method in which the disc line (disc contour) and cup line (cup contour) are computed using a spline approach to derive therefrom the C/D ratio, which is compared with the C/D ratio from the measurements with a table of C/D ratios for normal eyes created in advance in order to diagnose ophthalmic disorders.
- the present invention provides an image processing method for processing an image of an optic disc region in a photographed ocular fundus, comprising extracting a disc region; dividing the disc region into a plurality of regions by a plurality of dividing lines that extend radially from the center of gravity of the disc region; radially scanning the divided regions from the center of gravity while angles are shifted in order to detect in each individual region a point at which luminance variation reaches a maximum; and determining a contour line of the disc region based on the detected points at which luminance variation reaches a maximum.
- the present invention further provides an image processing method for processing an image of an optic disc region in a photographed ocular fundus, comprising dividing a disc region into a plurality of regions by a plurality of dividing lines that extend radially from a point that is set in the disc region; radially scanning the divided regions from the set point while angles are shifted in order to detect in each individual region a point at which depth variation reaches a maximum; and determining a contour line of the cup region based on the detected points at which depth variation reaches a maximum.
- a disc region is divided into a plurality of regions by a plurality of dividing lines that extend radially from a prescribed point.
- the divided regions are then scanned radially from the established point while angles are shifted in order to detect in each region a point at which luminance variation or depth variation reaches a maximum, and the contour line of the disc region or cup region is determined from these points. This allows the disc contours or cup contours to be determined automatically, and assisted diagnosis of glaucoma to be carried out more accurately and efficiently.
- FIG. 1 is a block diagram showing a configuration of an image processing system that the present invention employs
- FIG. 2 is an illustrative view showing operation buttons on the display of the image processing system
- FIG. 3 is a flowchart showing the flow of the entire image processing process of the present invention.
- FIG. 4 is a flowchart showing the flow of extraction of disc contours
- FIG. 5 is a flowchart showing the flow of extraction of cup contours
- FIG. 6 a is an illustrative view showing a method in which scan lines are used to search a disc region having a standard disc area
- FIG. 6 b is an illustrative view showing a luminance distribution of image-processed regions along the scan lines
- FIG. 7 is an illustrative view showing a method in which a divided disc detection region is angularly scanned to derive therefrom maximum luminance variation
- FIG. 8 is an illustrative view showing a method of deriving luminance variation
- FIG. 9 a is an illustrative view showing a method in which scan lines are used to search a cup region having a standard cup area
- FIG. 9 b is an illustrative view showing a depth distribution of image-processed regions along the scan lines
- FIG. 10 is an illustrative view showing a method of angular scanning of a divided cup detection region, and deriving maximum depth variation
- FIG. 11 is an illustrative view showing a method of deriving depth variation
- FIG. 12 is an illustrative view showing an extracted disc region and cup region
- FIG. 13 is an ocular fundus view showing a disc detection region together with blood vessel images
- FIG. 14 is an ocular fundus view showing extracted points of maximum luminance variation, shown together with blood vessel images.
- FIG. 15 is an ocular fundus view showing extracted points of maximum depth variation, shown together with blood vessel images.
- the present invention will now be described in detail with reference to the embodiment, which shows an ophthalmic measurement apparatus in which the ocular fundus under examination is stereographically photographed using a stereographical imaging optical system, and the photographed image undergoes three-dimensional measurement processing.
- FIG. 1 shows a configuration of an ophthalmic measurement apparatus showing one embodiment of the present invention.
- symbol 101 denotes a fundus camera.
- the fundus camera In order to photograph the ocular fundus under examination (not shown) under prescribed photographic conditions, the fundus camera is equipped with mechanisms such as, for example, an alignment mechanism, a monocular photography mechanism, and a stereographic photography mechanism.
- the fundus camera 101 has an image pickup device for color photography, such as a three-plate CCD or CMOS sensor for example, and outputs ocular fundus color image data of the eye under examination as digital data to an image processing device 100 .
- the picture signal output by the fundus camera 101 has a format such as YUV (YPbPr or YCbCr)
- a color separation process is carried out to convert the image data to that of a different spectrum such as RGB image data.
- Such a color separation process would be necessary when the fundus camera 101 outputs images in a format such as JPEG (or MPEG), for example.
- the image processing device 100 is constituted by hardware such as a PC, for example.
- the image processing device 100 carries out control of the system as a whole, and also includes a CPU 102 that constitutes the principal image processing means for carrying out image processing, which will be discussed later.
- the image processing device 100 could instead be constituted by dedicated hardware integrated in the fundus camera 101 .
- Image processing is executed using a VRAM (image memory) 104 serving as the work area.
- An additional memory for use in system control apart from image processing is a memory constituted by dynamic RAM or the like.
- a program for execution by the CPU 102 to carry out the image processing is stored in a ROM 103 or a hard disk (HDD) 105 .
- the hard disk 105 is used for storage of captured image data of an eye under examination, numeric data such as measurement results, output image data generated by the image processing, and the like.
- a display 107 composed of an LCD or EL panel, a CRT, or the like is connected to the image processing device 100 , and a user interface screen or output screen is displayed on the display 107 for the purpose of controlling image processing by the image processing device 100 .
- the image processing device 100 is provided with user interface means composed of a keyboard 108 and a mouse or other pointing device 109 .
- the image processing device 100 performs image processing, generates image data enabling the examiner to easily determine the ocular fundus under examination, particularly the cup region and disc region, and outputs the data to the display 107 .
- a network 106 is connected to the image processing device 100 via a network interface (not shown).
- the image processing device 100 outputs the aforementioned captured image data of the eye under examination, numeric data such as measurement results, output image data generated by the image processing, and the like to an external computer, or to a separate image processing device or ophthalmic measurement apparatus, or the like.
- FIGS. 3 to 5 The flow of the image processing procedure carried out by the image processing device 100 is illustrated in FIGS. 3 to 5 .
- the program by which the CPU 102 carries out image processing in the image processing device 100 is stored in the ROM 103 or on the hard disk 105 , and the CPU 102 loads and executes this program to carry out image processing according to the present embodiment.
- FIG. 3 shows the overall flow.
- the fundus camera 101 has taken a monocular color image of the eye under examination, or taken color stereo images of a pair of parallax images which allows stereoscopic viewing of the ocular fundus; and that the captured ocular fundus images are stored in the VRAM 104 or the hard disk 105 .
- Step S 1 of FIG. 3 an automatic extraction button is pushed to automatically extract the disc region or cup region of the optic disc from the ocular fundus image. This is accomplished by pushing the “Auto” button 208 of the “Region Selection” of the screen 200 on the display 107 in FIG. 2 .
- Step S 2 if the disc region has not yet been determined, automatic extraction of the disc region is performed. The user then checks the extracted region, performs correction if necessary, and pushes the “Determine Contour” button 211 .
- Step S 3 if only the cup region has not yet been determined, automatic extraction of the cup region is performed. The user checks the extracted region, performs correction if necessary, and pushes the “Determine Contour” button 211 .
- Automatic extraction of the disc region or the cup region may be accomplished by respectively pushing the “Disc” button 201 or the “Cup” button 202 shown in FIG. 2 , then pushing the “Auto” button 208 .
- the user may check the region, perform correction if necessary, and then push the “Determine Contour” button 211 to determine these contours independently.
- the determined disc region or cup region can be saved to the hard disk 105 by pushing the “Save” button 210 . Also, by pushing the “Back” button 207 , it is possible to load previously saved contour data, and to call the data up on the display 107 in order to make modifications thereto. By pressing the “Delete” button 209 it is possible to delete editing data etc. To cancel some or all image processing, the “Cancel” button 212 is pushed. In the event that multiple displays are connected, the “Select Screen” button can be used to select the display on which the image is to be displayed. The “User” button can be used to select a user. A “Depth computation” screen is provided as well, and the parameters such as the patient's diopter, corneal curvature, and the like can be selected during depth computation.
- FIG. 4 shows the flow of automatic extraction of a disc contour, i.e. the periphery of the disc region.
- An ocular fundus image obtained by monocular photography, or one parallax image, for example, the left eye image of a pair of parallax ocular fundus images obtained by stereographic photography is read from the hard disk 105 as the ocular fundus image for disc contour extraction and is saved in the VRAM 104 .
- Step S 11 a process such as morphology is used to erase blood vessel images from the ocular fundus image in the VRAM 104 , and color separation to RGB images is performed.
- the R image mostly contains information from relatively deep portions of the retina, e.g. from the choroid membrane; the G image mostly contains information from the retinal pigment epithelium; and the B image mostly contains information from the retinal surface.
- Step S 12 the RGB images are scanned for a region of high luminance in proximity to the disc region.
- FIG. 6 a shows scanning along scan lines 302 within a window 301 established in proximity to the disc region of an R image 300 for example.
- FIG. 6 b shows detection of pixels having greater luminance values than a prescribed luminance threshold value 304 . If pixels having greater luminance values than the luminance threshold value 304 are detected, it is checked whether the area of the region defined by those pixels is equivalent to or approximately equivalent to a predetermined standard disc area. In the example of FIG. 6 a , the pixels of a region 303 indicated by halftone dots have been detected as pixels that have greater luminance values than the luminance threshold value.
- the region 303 defined by the detected pixels and indicated by halftone dots constitutes a region of luminance greater than a prescribed luminance threshold value.
- the area of the luminance region increases as the luminance threshold value 304 indicated by the solid line in FIG. 6 b decreases to the luminance threshold value 304 ′ indicated by the dot and dash line.
- the image is scanned while varying the luminance threshold value, and a luminance region which has luminance greater than a prescribed luminance threshold value and whose area is equivalent to a predetermined standard disc area is extracted as a region having the standard disc area.
- the region having the standard disc area extracted in this manner is a region close to the disc region that is ultimately desired. Therefore, the region thus extracted is stored in the VRAM 104 as the disc detection region or detected disc region.
- Step S 12 the disc region having the standard disc area was extracted from each of the RGB images, and the region of overlap of these detected images was designated as the disc detection region.
- Step S 13 the disc detection region extracted in Step S 13 is indicated by symbol 11 and shown together with a color ocular fundus image 10 .
- the erased blood vessel images 12 are restored for reference purposes.
- Step S 14 the center of gravity of the disc detection region 11 , is calculated, and the disc detection region 11 is divided into eight regions by line segments that extend radially at equiangular intervals (of 45 degrees) from this center.
- FIG. 14 shows the center of gravity 13 of the disc detection region 11 calculated in Step S 14 and the eight regions S 1 to S 8 divided by eight dividing lines 14 a to 14 h which extend from the center of gravity 13 .
- each of the eight divided regions S 1 to S 8 is scanned along scan lines extending radially from the center of gravity 13 in order to detect a point at which luminance varies maximally, i.e. luminance variation reaches a maximum.
- FIG. 7 shows scanning of one divided region S 1 along equiangular scan lines 17 extending radially from the center of gravity 13 . Scanning is performed radially from the center of gravity 13 at resolution of from 1 to 2 degrees for example.
- FIG. 8 shows luminance variation observed when the disc detection region 11 is scanned along a single angular scan line.
- the vertical axis in FIG. 8 indicates luminance, and the horizontal axis indicates distance from the center of gravity 13 along the scan line 17 .
- This angular scanning is performed on the region S 1 to extract the maximum luminance variation point on each scan line. For example, where angular scanning takes place at resolution of 1 degree, 45 maximum luminance variation points will be extracted in the region S 1 . The one with the highest value of these 45 points is extracted as the disc contour point in the region S 1 , and is saved to the VRAM 104 .
- a maximum luminance variation point 15 a detected on angular scan line 17 a is the point having the highest value of the maximum luminance variation points in the region S 1 . Consequently, the point 15 a is extracted as the disc contour point in the region S 1 .
- This angular scanning is carried out for each of the regions S 1 to S 8 .
- Disc contour points, namely maximum luminance variation points extracted from the regions S 1 to S 8 are shown as 15 a to 15 h in FIG. 14 .
- Step S 15 The process described above is the process of Step S 15 .
- Step S 16 the eight extracted points 15 a to 15 h are joined by a third- or fourth-order spline curve, for example, to derive a disc contour line.
- the region enclosed by this contour line is designated as the disc region.
- FIG. 12 shows a disc region 500 and a disc contour line 501 determined in this manner.
- FIG. 5 shows the flow of a procedure for automatically extracting the cup contour, i.e., the periphery of the cup region.
- the ocular fundus image from which the cup contour is extracted is the same image as that used for disc contour extraction.
- Step S 21 a process such as morphology is used as well to erase blood vessel images and color separation into RGB images is performed.
- parallax is extracted from a pair of parallax images of each component of the RGB images, and on the basis of corresponding pixels in the pair of parallax images, depth data for pixels inside a region in proximity to the disc region is measured using a known method, and the measured depth data is saved to the VRAM 104 .
- Step S 22 the RGB images are scanned for a region of great depth, i.e. a deep region in proximity to the disc region.
- FIG. 9 a shows scanning along scan lines 302 within a window 301 established in proximity to the disc region of an R image 300 for example.
- FIG. 9 b shows detection of pixels deeper than a prescribed depth threshold value 305 . Pixels deeper than the depth threshold value are detected while successively varying the depth threshold value 305 from larger values to smaller values (or the reverse), and it is checked whether the area of the region defined by those pixels is equivalent to or approximately equivalent to a predetermined standard cup area.
- FIG. 9 a shows scanning along scan lines 302 within a window 301 established in proximity to the disc region of an R image 300 for example.
- FIG. 9 b shows detection of pixels deeper than a prescribed depth threshold value 305 . Pixels deeper than the depth threshold value are detected while successively varying the depth threshold value 305 from larger values to smaller values (or the reverse), and it is checked whether the area of the region defined by
- the pixels of a region 306 represented by halftone dots have been detected as pixels deeper than the luminance threshold value 305 , and the region 306 defined by the detected pixels and represented by halftone dots constitutes a region of greater depth than the depth threshold value.
- the area of the region decreases as the depth threshold value 305 indicated by the solid line in FIG. 9 b decreases to the depth threshold value 305 ′ indicated by the dotted-dashed line.
- the image is scanned while the depth threshold value is varied, and a depth region having a depth greater than a prescribed depth threshold value and having a area equivalent to a predetermined standard cup area is extracted as a region having the standard cup area.
- the region having the standard cup area thus extracted is a region close to the cup region that is ultimately desired, so that the extracted region is stored in the VRAM 104 as the cup detection region or detected cup region.
- FIG. 10 shows an extracted cup detection region 20 , represented by a broken line.
- Step S 22 a cup region of standard cup area was extracted from each of the RGB images, and the region of overlap of these detected images was designated as the cup detection region.
- Step S 24 the cup detection region 20 is divided into eight regions by line segments extending radially at equiangular intervals (of 45 degrees) from the center of gravity 13 that was calculated for the disc detection region.
- each of the eight divided regions S 1 to S 8 is scanned along scan lines extending radially from the center of gravity 13 to detect a point at which depth variation reaches a maximum.
- FIG. 10 shows scanning of one divided region S 1 along equiangular scan lines 18 extending radially from the center of gravity 13 . Scanning is performed radially from the center of gravity 13 at resolution of from 1 to 2 degrees for example.
- FIG. 11 shows depth variation observed when the cup detection region 20 is scanned along a single angular scan line.
- the vertical axis in FIG. 11 indicates depth from the retina, and the horizontal axis indicates distance from the center of gravity 13 along the scan line 18 .
- This angular scanning is performed on the region S 1 to extract the maximum depth variation point on each scan line. For example, where angular scanning takes place at resolution of 1 degree, 45 maximum depth points will be extracted in the region S 1 . Therefore, the one with the highest value of these 45 points is extracted as the cup contour point in the region S 1 , and is saved to the VRAM 104 .
- a maximum depth variation point 16 a detected on angular scan line 18 a is the maximum depth variation point in the region S 1 . Consequently, the point 16 a is extracted as the cup contour point in the region S 1 .
- Cup contour points namely maximum depth variation points extracted from the regions S 1 to S 8 , are shown as 16 a to 16 h in FIG. 15 .
- Step S 25 The process described above is the process of Step S 25 .
- Step S 26 the eight extracted points 16 a to 16 h are joined by a third- or fourth-order spline curve for example, to derive a cup contour line.
- the region enclosed by this contour line is designated as the cup region.
- FIG. 12 shows a cup region 510 and a cup contour line 511 determined in this manner.
- the disk region 500 and the contour line 501 thereof as well as the cup region 510 and the contour line 511 thereof, which are obtained by image processing as described above, may be saved as appended information to the ocular fundus image and recorded to a recording medium such as the hard disk 105 . Saved disc regions and cup regions may be displayed in time series on the display 107 . Also, saved disc contour lines and cup contour lines may be displayed on the display 107 to allow for correction of these contour lines if necessary.
- the disc detection region 11 and the cup detection region 20 herein are divided into eight parts by equiangular dividing lines extending from the center, the number of divisions is not limited to 8; some other plural number of divisions, such as 6 divisions or 12 divisions, is also acceptable.
- the angular resolution of the scan lines need not be set to a fixed resolution.
- the angle interval for scanning may vary according to angle region, for example, using finer angular intervals for the scan lines 17 , 18 in angle regions in which the contour lines of the disc region or the cup region are more complex, and coarser ones in angle regions in which the contour lines of the disc region or the cup region are fairly simple.
- the center of gravity of the disk region is selected when determining the contour line of a cup region, and the disk region and the cup region are divided into a plurality of regions by a plurality of dividing lines extending radially from the center of gravity.
- the disk region and the cup region may then be divided into a plurality of regions by a plurality of dividing lines extending radially from this selected point or center.
- the standard disc area means an area of the disc region serving as a reference, and may be the average value of measured area of a large number of disc regions
- the standard cup area means an area of the cup region serving as a reference, and may be the average value of the measured area of a large number of cup regions.
- a luminance region having a greater luminance value than a predetermined standard luminance value may be extracted as the disc region
- Step S 23 a region of greater depth than a predetermined standard depth value may be extracted as the cup region.
- a luminance region or depth region may be extracted in each of images obtained by color separation of an image into RGB images, and the region of overlap thereof may be selected as the disc region or cup region.
- the luminance region or depth region may be detected from the image prior to color separation into RGB images, and the disc region or cup region may be extracted therefrom.
Landscapes
- Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Eye Examination Apparatus (AREA)
- Image Processing (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (14)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2010010292A JP5583980B2 (en) | 2010-01-20 | 2010-01-20 | Image processing method and image processing apparatus |
JP2010-010292 | 2010-01-20 |
Publications (2)
Publication Number | Publication Date |
---|---|
US20110176108A1 US20110176108A1 (en) | 2011-07-21 |
US8240852B2 true US8240852B2 (en) | 2012-08-14 |
Family
ID=43909633
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/930,739 Expired - Fee Related US8240852B2 (en) | 2010-01-20 | 2011-01-14 | Image processing method and image processing device |
Country Status (3)
Country | Link |
---|---|
US (1) | US8240852B2 (en) |
EP (1) | EP2360641B1 (en) |
JP (1) | JP5583980B2 (en) |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9370328B2 (en) * | 2012-11-29 | 2016-06-21 | University Of Washington Through Its Center For Commercialization | Methods and systems for determining tumor boundary characteristics |
JP2014113422A (en) * | 2012-12-12 | 2014-06-26 | Canon Inc | Ophthalmological photographing apparatus, and control method and program of ophthalmological photographing apparatus |
CN104434027B (en) * | 2013-09-12 | 2016-03-02 | 广东福地新视野光电技术有限公司 | The method and apparatus of optic disc area is obtained by ophthalmic optical coherence tomoscan image |
DE202014106268U1 (en) | 2014-12-23 | 2015-02-25 | Toll Collect Gmbh | Vehicle equipment for DSRC communication |
US10554956B2 (en) | 2015-10-29 | 2020-02-04 | Dell Products, Lp | Depth masks for image segmentation for depth-based computational photography |
US10021371B2 (en) | 2015-11-24 | 2018-07-10 | Dell Products, Lp | Method and apparatus for gross-level user and input detection using similar or dissimilar camera pair |
CN109993765B (en) * | 2019-04-09 | 2020-10-30 | 东莞理工学院 | Method for detecting retinal vein cross compression angle |
CN110443256B (en) * | 2019-07-03 | 2022-04-12 | 大连理工大学 | Method for extracting multi-target regions of image |
CN112001920B (en) * | 2020-10-28 | 2021-02-05 | 北京至真互联网技术有限公司 | Fundus image recognition method, device and equipment |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7524061B2 (en) * | 2005-10-12 | 2009-04-28 | Siemens Corporate Research, Inc. | System and method for robust optic disk detection in retinal images using vessel structure and radon transform |
US7854510B2 (en) * | 2008-10-16 | 2010-12-21 | Steven Roger Verdooner | Apparatus and method for imaging the eye |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3508112B2 (en) * | 1993-09-21 | 2004-03-22 | 株式会社トプコン | Fundus disease analyzer |
JP3585331B2 (en) * | 1996-12-03 | 2004-11-04 | 株式会社ニデック | Analysis method of fundus stereoscopic image |
JP3594468B2 (en) * | 1997-11-21 | 2004-12-02 | 株式会社ニデック | Fundus image analysis method |
JP4790295B2 (en) | 2005-03-31 | 2011-10-12 | 株式会社ニデック | Method of operating the analyzer |
JP5007420B2 (en) | 2006-09-21 | 2012-08-22 | タック株式会社 | Image analysis system and image analysis program |
-
2010
- 2010-01-20 JP JP2010010292A patent/JP5583980B2/en not_active Expired - Fee Related
-
2011
- 2011-01-14 US US12/930,739 patent/US8240852B2/en not_active Expired - Fee Related
- 2011-01-20 EP EP11151553.2A patent/EP2360641B1/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7524061B2 (en) * | 2005-10-12 | 2009-04-28 | Siemens Corporate Research, Inc. | System and method for robust optic disk detection in retinal images using vessel structure and radon transform |
US7854510B2 (en) * | 2008-10-16 | 2010-12-21 | Steven Roger Verdooner | Apparatus and method for imaging the eye |
Also Published As
Publication number | Publication date |
---|---|
EP2360641A1 (en) | 2011-08-24 |
EP2360641B1 (en) | 2019-07-24 |
JP2011147547A (en) | 2011-08-04 |
JP5583980B2 (en) | 2014-09-03 |
US20110176108A1 (en) | 2011-07-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8240852B2 (en) | Image processing method and image processing device | |
CN101404923B (en) | Image processing device for medical use | |
US6714672B1 (en) | Automated stereo fundus evaluation | |
US10945637B2 (en) | Image based jaundice diagnosing method and apparatus and image based jaundice diagnosis assisting apparatus | |
KR101998595B1 (en) | Method and Apparatus for jaundice diagnosis based on an image | |
US20110243415A1 (en) | Image processing apparatus, control method thereof, and program | |
US20070188705A1 (en) | Ocular fundus portion analyzer and ocular fundus portion analyzing method | |
CN102697446B (en) | Image processing apparatus and image processing method | |
JP4915737B2 (en) | Image analysis system and image analysis program | |
JP5478103B2 (en) | Image processing method | |
US10796418B2 (en) | Image processing apparatus, image processing method, and program | |
JP5111845B2 (en) | Fundus image processing apparatus, fundus photographing apparatus, and program | |
JP2014198223A (en) | Information processor and information processing method | |
JP2020201823A (en) | Image processing device, image processing method, and program | |
JP2008237840A (en) | Image analysis system and image analysis program | |
WO2015037316A1 (en) | Organ-imaging device and organ-imaging method | |
JP2000245700A (en) | Instrument for measuring eyeground and recording medium for recording measurement program | |
JP2014102713A (en) | Face component extraction device, face component extraction method, and program | |
EP4185184B1 (en) | Method for determining a coronal position of an eye relative to the head | |
JP2006074498A (en) | Image processor and imaging apparatus | |
JP2006324727A (en) | Imaging apparatus and image processing method thereof | |
JP2012186821A (en) | Face image processing device, face image processing method, electronic still camera, digital image processing device and digital image processing method | |
JP4949007B2 (en) | Fundus image processing apparatus, fundus photographing apparatus, and program | |
CN110584588A (en) | Method and device for detecting visual field defect | |
EP4349268A1 (en) | Determination device, determination method, and program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: KOWA COMPANY LTD., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:NAKAGAWA, TOSHIAKI;TAKIMOTO, SHIGERU;REEL/FRAME:025703/0383 Effective date: 20110111 |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
ZAAA | Notice of allowance and fees due |
Free format text: ORIGINAL CODE: NOA |
|
ZAAB | Notice of allowance mailed |
Free format text: ORIGINAL CODE: MN/=. |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Free format text: PAYER NUMBER DE-ASSIGNED (ORIGINAL EVENT CODE: RMPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20240814 |