CN114897831A - Ultra-wide-angle eyeground image optic disk extraction method combining target positioning and semantic segmentation - Google Patents

Ultra-wide-angle eyeground image optic disk extraction method combining target positioning and semantic segmentation Download PDF

Info

Publication number
CN114897831A
CN114897831A CN202210521293.2A CN202210521293A CN114897831A CN 114897831 A CN114897831 A CN 114897831A CN 202210521293 A CN202210521293 A CN 202210521293A CN 114897831 A CN114897831 A CN 114897831A
Authority
CN
China
Prior art keywords
optic disc
image
ultra
optic
model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210521293.2A
Other languages
Chinese (zh)
Inventor
徐光柱
林文杰
陈莎
刘鸣
雷帮军
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Three Gorges University CTGU
Original Assignee
China Three Gorges University CTGU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Three Gorges University CTGU filed Critical China Three Gorges University CTGU
Priority to CN202210521293.2A priority Critical patent/CN114897831A/en
Publication of CN114897831A publication Critical patent/CN114897831A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20092Interactive image processing based on input by user
    • G06T2207/20104Interactive definition of region of interest [ROI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30041Eye; Retina; Ophthalmic

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Quality & Reliability (AREA)
  • Medical Informatics (AREA)
  • Eye Examination Apparatus (AREA)

Abstract

The method for extracting the ultra-wide-angle fundus image optic disk by combining target positioning and semantic segmentation comprises the following steps: step 1: establishing a YOLOv4 model for roughly positioning the optic disc area; step 2: extracting the optic disc area according to the coarse positioning result of the optic disc area, and removing the eye periphery area; and step 3: interactively segmenting the optic disc region through a Snake model of the active contour model to construct a U 2 -a dataset with disc labels required for Net model training;and 4, step 4: using U 2 -Net model for optic disc region extraction. The invention not only effectively improves the optic disc extraction precision, but also can obtain optic disc labeling data required by the supervised learning segmentation method. The accuracy rate of the optic disc positioning of the whole method reaches 99.7 percent, and stable and reliable input is provided for the subsequent segmentation links; the segmentation precision is higher than that of widely used PCNN, U-Net, DeepLabV3 and SegNet models, and the method has good application value.

Description

Ultra-wide-angle eyeground image optic disk extraction method combining target positioning and semantic segmentation
Technical Field
The invention relates to the technical field of ultra-wide-angle fundus imaging, in particular to an ultra-wide-angle fundus image optic disc extraction method combining target positioning and semantic segmentation.
Background
Glaucoma is a serious hazard and can irreversibly damage a person's vision, even leading to blindness. The change of structures such as the size and the shape of the optic disc is closely related to the pathological change degree of glaucoma of a patient, and the glaucoma can be found in time by detecting the eye structures of the optic disc and the optic cup. See document [1] Amsa Shabbir, Aqsa Rasheed, Huma Shehraz, et al.Detection of glaucoma using recombinant enzymes A comprehensive review. physical Biosciences and Engineering,2021,18(3): 2033. 2076.
The number of glaucoma patients in China is huge, but the number of professional ophthalmologists is seriously deficient. Moreover, manual extraction of the fundus optic disc region takes a long time, has low efficiency, and is difficult to meet clinical requirements. Therefore, the automatic extraction of the fundus optic disc region by utilizing the AI technology is an effective solution for solving the problem of the lack of ophthalmologists and improving the early screening participation of glaucoma patients, and is widely concerned by domestic and foreign researchers. In the document [2] chinese patent, "a retinal optic disc segmentation method based on deep learning" (CN113240677A), in order to segment the optic disc, a hole convolution is introduced into a U-Net model, and a subtraction operation element by element is introduced into a jump connection, and features obtained by sampling each layer are subtracted from corresponding features of a lower layer and from a feature of an upper layer which has the same size after a maximum pooling operation of 4x 4.
Document [3] Fu Huazhu, Cheng Jun, Xu Yanwu, et al. Joint optical disc and cup segmentation based on multi-label deep learning, 2018,37(7): 1597) 1605, proposes a deep learning model named M-Net, which uses multi-labels to solve the problem of optic disc and cup segmentation. The network mainly comprises a multi-scale input layer, a U-shaped convolution network, an edge output layer and a multi-label loss function. The multi-scale input layer constructs an image pyramid to achieve multi-scale receptive field size. A U-Net model is used as a main network structure to learn multi-scale blood vessel characteristic information, and a side output layer is used as an early classifier to generate corresponding local prediction images for different scale layers. Finally, a multi-label penalty function is used to generate the final segmentation map.
In the document [4] chinese patent "a cup optic disc segmentation method based on fundus map data set migration learning" (CN112541923B), a cup optic disc segmentation method based on migration learning is proposed for the problem of low generalization of model segmented optic discs caused by large feature differences between different fundus map data sets. According to the method, through the countermeasure training of the backbone segmentation network, the feature field discriminator and the attention field discriminator, the general features among the eye fundus image data sets are extracted, and the features are weighted by the attention module, so that the problem of blurred vision cup and optic disc boundaries is effectively relieved.
Compared with the traditional fundus image imaging, as shown in fig. 1(a), the ultra-wide angle fundus imaging technology has the advantages of large visual field, fast imaging, no mydriasis, good patient compliance and the like, as shown in fig. 1 (b). However, the brightness and definition of the formed image are inferior to those of the traditional fundus image and the resolution is high, so that the research of automatic interpretation of the fundus image is still in the initial research stage.
In the aspect of disk extraction of super wide-angle fundus images, current special research reports are rare, and documents [5] Li Zhongwen, Guo Chong, Nie Danyao, et al. A deep learning system for identifying molecular differentiation and regeneration research using ultra-wide-field images, 2019,7(22):618. Then the video disc is sent to a U-Net model trained based on a traditional fundus image data set after certain pretreatment to realize the extraction of the video disc. The reason for this is because the ultra-wide fundus image optic disk region labeling data is lacking, and the accuracy is also affected.
Disclosure of Invention
The method aims at the problems that the optic disc region extraction precision is low and the optic disc segmentation model cannot be directly used for training due to the small optic disc target, the lack of label data, the uneven illumination, the serious noise interference and the like in the ultra-wide angle eye ground image. The invention provides an ultra-wide angle fundus image optic disc extraction method combining target positioning and semantic segmentation, which comprises the steps of firstly training a visual target detection network YOLOv4 to realize coarse positioning of an optic disc area, and converting a small target positioning problem of optic disc detection in an ultra-wide angle fundus image into a larger target detection problem of a candidate area containing an optic disc. The problem difficulty is simplified, and the detection precision is improved; secondly, segmenting a video disc region approved by human eyes in an interactive mode by using a Snake algorithm in the active contour curve model to obtain a label used for subsequent neural network training; finally train U 2 The Net model performs accurate extraction of the optic discs in the candidate region.
The technical scheme adopted by the invention is as follows:
the method for extracting the ultra-wide-angle fundus image optic disk by combining target positioning and semantic segmentation comprises the following steps:
step 1: establishing a YOLOv4 model for roughly positioning the optic disc area;
step 2: extracting the optic disc area according to the coarse positioning result of the optic disc area, and removing the eye periphery area;
and 3, step 3: interactively segmenting the optic disc region through a Snake model of the active contour model to construct a U 2 -a dataset with disc labels required for Net model training;
and 4, step 4: using U 2 -Net model for optic disc region extraction.
In the step 2, according to the positioning result of the model optodisc of YOLOv4, extracting an interested region containing the optodisc, removing the periocular region extracted from the interference optodisc region, positioning the center of the optodisc at the left or right position of the intercepted image, intercepting two interested regions with the size of 256 × 256, and covering with a mask.
The step 3 comprises the following steps:
firstly, carrying out gray scale normalization and Gaussian filtering on an ultra-wide angle fundus image to remove noise; then, manually selecting coordinate points of the initial contour curve, and obtaining a closed contour curve by utilizing an interpolation method; then, calculating an energy function E, obtaining the gradient F and the five-diagonal strip matrix A of the energy function E by using the energy function E, and further updating each point position on the curve; finally, filling a blank area in the closed curve to obtain a video disc label;
as shown in formulas (1) to (6):
E snake =∫E(v(s))ds=E int (v(s))+E ext (v(s)) (1);
wherein: v(s) represents a curve function, E snake Representing the energy of the Snake curve, E int Represents internal energy, E ext Represents external energy;
E int (v(s))=∫(|v′(s)| 2 +|v″(s)| 2 )ds (2);
wherein: v'(s) is the first derivative of the curve function, v "(s) is the second derivative of the curve function,
E ext (v(s))=E img (v(s))+E constaint (v(s)) (3);
wherein: e img (v(s))、E constaint (v (s)) represent image energy and binding force energy, respectively;
Figure BDA0003643530300000031
wherein: i is an input image, I x And I y Respectively representing the partial derivatives of the image pair on an x axis and a y axis, wherein x and y are coordinate values;
Figure BDA0003643530300000033
representing the reciprocal of the image, I xx 、I xy 、I yy Respectively represent: second partial derivatives of the image on the x-axis, second partial derivatives of the first derivatives on the x-and y-axes, and second partial derivatives on the y-axisThe second partial derivative of (c).
Figure BDA0003643530300000032
Wherein: x is the number of ss And x ssss Respectively representing v(s) first and second order partial derivatives, y, of the x-axis ss And y ssss Respectively representing the first and second order partial derivatives of v(s) to the y-axis; alpha and beta are weights.
Figure BDA0003643530300000041
Wherein: f. of x And f y Each represents E snake For the partial derivatives of the x axis and the y axis, t represents the iteration times;
x t 、y t coordinates of each point of the iteration curve are obtained;
x t-1 、y t-1 coordinates of each point of a last iteration curve are shown, A is a five-diagonal strip matrix, and gamma is a weight coefficient;
in said step 3, in U 2 Before the training of the Net model, the contrast between the optic disc and the background is enhanced through a contrast limited histogram equalization technology CLAHE so as to improve the training effect of the model.
The step 4 comprises the following steps:
s4.1: inputting enhanced pictures into U 2 -Net model, extracting optic disc regions;
s4.2: to U 2 And (4) performing binarization on the output result of the Net model to obtain a binary optic disc region image.
The invention relates to an ultra-wide angle fundus image optic disk extraction method combining target positioning and semantic segmentation, which has the following technical effects:
1) according to the method, the YOLOv4 network model with both speed and detection precision is used for positioning the region of interest containing the optic disc, so that the difficulty of subsequent segmentation caused by small proportion of the optic disc region in the whole ultra-wide-angle fundus image can be effectively reduced.
2) The invention then cuts out the positioned interested region, and utilizes Snake algorithm in the active contour motion curve model to realize, cuts out the optic disk region recognized by human eyes in an interactive mode, and uses the optic disk region as a label for subsequent neural network training. The problem of super wide angle eye ground image concentrated video disk region label lack is solved.
3) After the invention obtains the label data of the video disc area, the invention carries out the U-pair 2 -Net training, before training, enhancing the contrast between the optic disc and the background by a contrast-limited histogram equalization technique (CLAHE) to improve the model training effect, and finally, carrying out U training 2 And carrying out binarization on the Net result to obtain a segmented optic disc region.
Drawings
Fig. 1(a) is a conventional fundus image;
fig. 1(b) is an ultra-wide-angle fundus image.
Fig. 2 is a schematic diagram of a super wide-angle image optic disk extraction process.
Fig. 3 is a diagram of the YOLOv4 model structure.
FIG. 4(a) shows a first positioning of the optic disc for YOLOv 4;
FIG. 4(b) shows the positioning of disc result two for YOLOv 4.
Fig. 5 is a schematic diagram of extracting a region of interest containing an optic disc.
Fig. 6 is a flow chart of Snake model segmentation optic disc.
Fig. 7(a) is an original image before image enhancement by the CLAHE method;
fig. 7(b) is a diagram after image enhancement by the CLAHE method.
FIG. 8 is U 2 -Net network architecture diagram.
Fig. 9 is a flowchart of disc region extraction.
Fig. 10 is a comparison graph of the extraction result of the ultra-wide angle fundus image optic disk region, in which:
a, the unit pictures in the column represent the original pictures;
b the unit figures contained in column represent label diagrams;
the unit graph contained in the c column represents a PCNN model optic disc region extraction result graph;
the unit graph contained in the d column represents a U-Net model optic disk region extraction result graph;
the unit diagram contained in the e column represents a drawing result diagram of a DeepLabv3 model optic disc region;
cell graphs contained in f columns represent a graph for extracting results of SegNet model optic disk regions;
the cell diagram included in the g column represents the disc region extraction result diagram of the method of the invention.
FIG. 11 is a diagram illustrating the quantitative comparison of the disk region extraction results of the method of the present invention with other methods.
Fig. 12 is a Snake evolution diagram.
Detailed Description
The method aims at the problems that the optic disc region extraction precision is low and the optic disc segmentation model cannot be directly used for training in ultra-wide angle eye ground images due to the fact that the optic disc region is small and lacks of labeling information, illumination is uneven, noise interference is serious and the like. The invention provides an ultra-wide angle eyeground image optic disk region extraction algorithm combining a target detection model and a semantic segmentation model. Because the optic disc region is a small target relative to the whole ultra-wide angle fundus image and cannot be directly sent to a semantic segmentation network for extraction, the method firstly uses the YOLOv4 network to roughly position the optic disc region, the positioning accuracy reaches 99.7 percent, and the optic disc region of interest is extracted; then, aiming at the problem that the ultra-wide angle fundus image lacks of optic disc standard information, an optic disc region is segmented interactively by using a Snake algorithm of an active contour model, and a data set with an optic disc label required by subsequent neural network training is constructed; finally, aiming at the problem of low precision of the extraction of the ultra-wide angle eye ground image optic disk region, the U is used 2 Net model extracts optic disc regions accurately. The extraction performance of the optic disc area of the ultra-wide angle fundus image is superior to that of other optic disc extraction methods at present, such as:
the scheme is described in the literature [8] Xuguan column, Wang Yao, Hu Song, etc.. PCNN and form matching enhancement combined retina blood vessel segmentation, photoelectric engineering, 2019,46(04): 74-85;
the protocol described in the literature [9] Ronneberger Olaf, Fischer Philipp, Brox Thomas. U-net: volumetric networks for biological image segmentation. International Conference on Medical image computing and computer-assisted interaction 2015: 234-241;
the protocol described in the document [10] Chen Liang-Chieh, Papandrou George, Schroff Florian, et al.
The protocol described in the document [11] Badrinarayanan Vijay, Kendall Alex, Cipola Robert.Segnet: A deep connected encoder-decoder architecture for image segmentation. IEEE transactions on pattern analysis and machine understanding, 2017,39(12):2481 and 2495.
The method for extracting the ultra-wide-angle fundus image optic disk by combining target positioning and semantic segmentation comprises the following steps:
the method comprises the following steps: because the optic disc area is a small target relative to the whole ultra-wide angle fundus image, the optic disc area cannot be directly sent to a semantic segmentation network for extraction, and in addition, the ultra-wide angle fundus image has poor quality and comprises eye peripheral areas such as eyelids and eyelashes which are not beneficial to optic disc area extraction. The invention aims at the problem of coarse positioning of the optic disc area by using a single-stage target detection model YOLOv4, and the structure of the invention is shown in FIG. 3
The rough positioning of the optic disc area by YOLOv4 specifically includes:
the Yolov4 model takes a CSPDarknet53 module as a backbone network, is used for extracting deep semantic features of an optic disc in an ultra-wide angle fundus image, then fuses multi-scale features of the optic disc by using a space pyramid pooling and path aggregation network, and finally obtains optic disc position information by using a multi-scale output module. It can be seen from fig. 4(a) and 4(b) that the YOLOv4 model can accurately locate the optic disc position with a confidence of over 90%.
Step two: according to the positioning result of the disc of the YOLOv4 model, extracting the region of interest containing the disc, and removing the periocular region which interferes with the extraction of the disc region. As shown in fig. 5, the disk is centered to the left or right of the truncated image, and two 256 x 256 regions of interest are truncated and covered with a mask.
Step three: aiming at the problem that the ultra-wide angle fundus image lacks of optic disc region labels, the optic disc region is segmented interactively by using an active contour model Snake model to serve as a pixel level label required by subsequent deep learning model training, and the process is shown in FIG. 6. Firstly, carrying out gray scale normalization and Gaussian filtering on the ultra-wide angle fundus image to remove noise, then manually selecting a coordinate point of an initial contour curve, and obtaining a closed contour curve by using an interpolation method; and calculating an energy function E, obtaining the gradient F and the five-diagonal ribbon matrix A of the video label by using the energy function, further updating each point position on the curve, and finally filling a blank area in the closed curve to obtain the video label.
As shown in formulas (1) to (6):
E snake =∫E(v(s))ds=E int (v(s))+E ext (v(s)) (1);
wherein: v(s) represents a curve function, E snake Representing the energy of the Snake curve, E int Represents internal energy, E ext Represents external energy;
E int (v(s))=∫(|v(s)| 2 +|v″(s)| 2 ds (2);
wherein: v'(s) is the first derivative of the curve function, v "(s) is the second derivative of the curve function,
E ext (v(s))=E img (v(s))+E constaint (v(s)) (3);
wherein: e img (v(s))、E constaint (v (s)) represent image energy and binding force energy, respectively;
Figure BDA0003643530300000071
wherein: i is an input image, I x And I y Respectively representing the partial derivatives of the image pair on an x axis and a y axis, wherein x and y are coordinate values;
Figure BDA0003643530300000072
representing the reciprocal of the image;
I xx 、I xy 、I yy respectively represent: second order offset of image on x-axisDerivatives, second partial derivatives of each first derivative on the x and y axes, and second partial derivatives on the y axis;
Figure BDA0003643530300000073
wherein: x is the number of ss And x ssss Respectively representing v(s) first and second order partial derivatives, y, of the x-axis ss And y ssss Respectively representing the first and second order partial derivatives of v(s) to the y-axis; alpha and beta are weights.
Figure BDA0003643530300000074
Wherein: f. of x And f y Each represents E snake For the partial derivatives of the x axis and the y axis, t represents the iteration times;
x t 、y t coordinates of each point of the iteration curve are obtained;
x t-1 、y t-1 coordinates of each point of the last iteration curve;
a is a five diagonal banded matrix and gamma is a weight coefficient.
Step four: aiming at the problems of relatively deficient data volume and poor quality of ultra-wide-angle fundus images, the invention uses image augmentation and image contrast enhancement to increase the data volume and enhance the images. The invention increases the number of pictures in the data set to 4 times of the original number by rotating and turning. Then, the grey normalization operation is carried out on the video disc image, and then the CLAHE method is used for weakening noise and increasing the contrast ratio of the video disc and the background. As shown in fig. 7(a) and 7 (b).
Step five: aiming at the problem of low precision of ultra-wide angle eyeground image optic disk region extraction, the invention uses U 2 Net model performs optic disc region extraction. As shown in FIG. 8, U 2 The Net model obtains strong global and local feature extraction capability through an RSU module capable of capturing context information of different scales and a jump connection capable of fusing shallow spatial position information and deep semantic information. The invention inputs the enhanced image directly to the U 2 The Net model performs the training of optic disc extraction.
The invention provides an ultra-wide angle eyeground image optic disk region extraction algorithm combining visual target positioning and semantic segmentation. In order to solve the problems that the area proportion of the optic disc area in the ultra-wide angle fundus image is small, the optic disc area cannot be directly and effectively extracted, and the optic disc labeling information is lacked, the invention firstly divides the precise extraction of the optic disc area into two stages of optic disc target detection and optic disc area segmentation. In the first stage, the optic disc positioning small target detection problem is converted into the larger target detection problem by positioning the larger image area containing the optic disc, and the precise positioning is realized by training the Yolov3 network model. And then, interactively segmenting the optic disk in the detected region by using a Snake algorithm in the main profile active curve model, and constructing a training image set with optic disk labeling information required by the training of a subsequent segmentation model. In the second stage, the data set with the video disc labeling information obtained in the previous stage is used for training U 2 Net, enabling accurate extraction of optic disc regions. The two-stage optic disc extraction method and the interactive optic disc labeling method not only effectively improve optic disc extraction precision, but also can obtain optic disc labeling data required by a supervised learning segmentation method. The accuracy rate of the optic disc positioning of the whole method reaches 99.7 percent, and stable and reliable input is provided for the subsequent segmentation links; the segmentation precision is higher than that of widely used PCNN, U-Net, DeepLabV3 and SegNet models, and the method has good application value.
As can be seen from FIG. 10, compared to the PCNN, U-Net, DeepLabV3 and SegNet models widely used for optic disc extraction, the algorithm of the present invention can extract a more complete optic disc region, which is especially evident on the focus image with blurred optic disc boundaries.
As can be seen from FIG. 11, in addition to the sensitivity, the algorithm of the present invention has the best results in four quantitative indicators of specificity, accuracy and Dice coefficient, compared with the PCNN, U-Net, DeepLabV3 and SegNet models, and has a powerful optic disc region extraction capability.

Claims (5)

1. The method for extracting the ultra-wide-angle fundus image optic disk by combining target positioning and semantic segmentation is characterized by comprising the following steps of:
step 1: establishing a YOLOv4 model for roughly positioning the optic disc area;
step 2: extracting the optic disc area according to the coarse positioning result of the optic disc area, and removing the eye periphery area;
and 3, step 3: interactively segmenting the optic disc area through an active contour model Snake model to construct U 2 -a dataset with disc labels required for Net model training;
and 4, step 4: using U 2 Net model performs optic disc region extraction.
2. The method for extracting optic disk of ultra-wide angle fundus image by combining target location and semantic segmentation as claimed in claim 1, wherein: in the step 2, according to the positioning result of the model optodisc of YOLOv4, extracting an interested region containing the optodisc, removing the periocular region extracted from the interference optodisc region, positioning the center of the optodisc at the left or right position of the intercepted image, intercepting two interested regions with the size of 256 × 256, and covering with a mask.
3. The method for extracting optic disk of ultra-wide angle fundus image by combining target location and semantic segmentation as claimed in claim 1, wherein: the step 3 comprises the following steps:
firstly, carrying out gray scale normalization and Gaussian filtering on an ultra-wide angle fundus image to remove noise; then, manually selecting coordinate points of the initial contour curve, and obtaining a closed contour curve by utilizing an interpolation method; then, calculating an energy function E, obtaining the gradient F and the five-diagonal strip matrix A of the energy function E by using the energy function E, and further updating each point position on the curve; finally, filling a blank area in the closed curve to obtain a video disc label;
as shown in formulas (1) to (6):
E snake =∫E(v(s))ds=E int (v(s))+E ext (v(s)) (1);
wherein: v(s) represents a curve function, E snake Representing the energy of the Snake curve, E int Represents internal energy, E ext Represents external energy;
E int (v(s))=∫(|v′(s)| 2 +|v″(s)| 2 )ds (2);
wherein: v'(s) is the first derivative of the curve function, v "(s) is the second derivative of the curve function;
E ext (v(s))=E img (v(s))+E constaint (v(s)) (3);
wherein: e img (v(s))、E constaint (v (s)) represent image energy and binding force energy, respectively;
Figure FDA0003643530290000011
wherein: i is an input image, I x And I y Respectively representing the partial derivatives of the image pair on an x axis and a y axis, wherein x and y are coordinate values;
Figure FDA0003643530290000021
representing the reciprocal of the image, I xx 、I xy 、I yy Respectively represent: second-order partial derivatives of the image on an x axis, second-order partial derivatives of first-order derivatives on the x axis and a y axis and second-order partial derivatives on the y axis;
Figure FDA0003643530290000022
wherein: x is the number of ss And x ssss Respectively representing v(s) first and second order partial derivatives, y, of the x-axis ss And y ssss Respectively representing the first and second order partial derivatives of v(s) to the y-axis; alpha and beta are weights;
Figure FDA0003643530290000023
wherein: f. of x And f y Each represents E snake For the partial derivatives of the x axis and the y axis, t represents the iteration times;
x t 、y t coordinates of each point of the iteration curve are obtained;
x t-1 、y t-1 the coordinates of each point of the last iteration curve are A, a five-diagonal strip matrix and gamma is a weight coefficient.
4. The method for extracting optic disk of ultra-wide angle fundus image by combining target location and semantic segmentation as claimed in claim 1, wherein: in said step 3, in U 2 Before the Net model training, the contrast between the optic disc and the background is enhanced through a contrast limited histogram equalization technology CLAHE, so that the model training effect is improved.
5. The method for extracting optic disk of ultra-wide angle fundus image by combining target location and semantic segmentation as claimed in claim 1, wherein: the step 4 comprises the following steps:
s4.1: inputting enhanced pictures into U 2 -Net model, extracting optic disc regions;
s4.2: to U 2 And (4) performing binarization on the output result of the Net model to obtain a binary optic disc region image.
CN202210521293.2A 2022-05-13 2022-05-13 Ultra-wide-angle eyeground image optic disk extraction method combining target positioning and semantic segmentation Pending CN114897831A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210521293.2A CN114897831A (en) 2022-05-13 2022-05-13 Ultra-wide-angle eyeground image optic disk extraction method combining target positioning and semantic segmentation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210521293.2A CN114897831A (en) 2022-05-13 2022-05-13 Ultra-wide-angle eyeground image optic disk extraction method combining target positioning and semantic segmentation

Publications (1)

Publication Number Publication Date
CN114897831A true CN114897831A (en) 2022-08-12

Family

ID=82722473

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210521293.2A Pending CN114897831A (en) 2022-05-13 2022-05-13 Ultra-wide-angle eyeground image optic disk extraction method combining target positioning and semantic segmentation

Country Status (1)

Country Link
CN (1) CN114897831A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116309549A (en) * 2023-05-11 2023-06-23 爱尔眼科医院集团股份有限公司 Fundus region detection method, fundus region detection device, fundus region detection equipment and readable storage medium

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116309549A (en) * 2023-05-11 2023-06-23 爱尔眼科医院集团股份有限公司 Fundus region detection method, fundus region detection device, fundus region detection equipment and readable storage medium
CN116309549B (en) * 2023-05-11 2023-10-03 爱尔眼科医院集团股份有限公司 Fundus region detection method, fundus region detection device, fundus region detection equipment and readable storage medium

Similar Documents

Publication Publication Date Title
CN110992382B (en) Fundus image optic cup optic disc segmentation method and system for assisting glaucoma screening
CN110428432B (en) Deep neural network algorithm for automatically segmenting colon gland image
CN112132817B (en) Retina blood vessel segmentation method for fundus image based on mixed attention mechanism
CN110276356A (en) Eye fundus image aneurysms recognition methods based on R-CNN
CN111046835A (en) Eyeground illumination multiple disease detection system based on regional feature set neural network
Xu et al. Ffu-net: Feature fusion u-net for lesion segmentation of diabetic retinopathy
CN109886946B (en) Deep learning-based early senile maculopathy weakening supervision and classification method
Lin et al. A high resolution representation network with multi-path scale for retinal vessel segmentation
CN108305241B (en) SD-OCT image GA lesion segmentation method based on depth voting model
CN112288720A (en) Deep learning-based color fundus image glaucoma screening method and system
CN110689526B (en) Retinal blood vessel segmentation method and system based on retinal fundus image
Xu et al. Dual-channel asymmetric convolutional neural network for an efficient retinal blood vessel segmentation in eye fundus images
CN109670489B (en) Weak supervision type early senile macular degeneration classification method based on multi-instance learning
CN112884788B (en) Cup optic disk segmentation method and imaging method based on rich context network
Bhatkalkar et al. Automated fundus image quality assessment and segmentation of optic disc using convolutional neural networks
CN114648806A (en) Multi-mechanism self-adaptive fundus image segmentation method
Jian et al. Triple-DRNet: A triple-cascade convolution neural network for diabetic retinopathy grading using fundus images
CN110889846A (en) Diabetes retina image optic disk segmentation method based on FCM
CN114897831A (en) Ultra-wide-angle eyeground image optic disk extraction method combining target positioning and semantic segmentation
Wang et al. Towards an extended EfficientNet-based U-Net framework for joint optic disc and cup segmentation in the fundus image
CN110610480A (en) MCASPP neural network eyeground image optic cup optic disc segmentation model based on Attention mechanism
CN114334124A (en) Pathological myopia detection system based on deep neural network
CN113191352A (en) Water meter pointer reading identification method based on target detection and binary image detection
Huang et al. DBFU-Net: Double branch fusion U-Net with hard example weighting train strategy to segment retinal vessel
CN116597424A (en) Fatigue driving detection system based on face recognition

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination