CN117078697B - Fundus disease seed detection method based on cascade model fusion - Google Patents
Fundus disease seed detection method based on cascade model fusion Download PDFInfo
- Publication number
- CN117078697B CN117078697B CN202311053124.1A CN202311053124A CN117078697B CN 117078697 B CN117078697 B CN 117078697B CN 202311053124 A CN202311053124 A CN 202311053124A CN 117078697 B CN117078697 B CN 117078697B
- Authority
- CN
- China
- Prior art keywords
- fundus
- image
- module
- segmentation
- attention
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 201000010099 disease Diseases 0.000 title claims abstract description 57
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 title claims abstract description 57
- 238000001514 detection method Methods 0.000 title claims abstract description 15
- 230000004927 fusion Effects 0.000 title claims abstract description 13
- 230000011218 segmentation Effects 0.000 claims abstract description 55
- 238000003745 diagnosis Methods 0.000 claims abstract description 19
- 238000012545 processing Methods 0.000 claims abstract description 16
- 210000004204 blood vessel Anatomy 0.000 claims abstract description 10
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 8
- 238000007781 pre-processing Methods 0.000 claims abstract description 6
- 238000000034 method Methods 0.000 claims description 22
- 238000010586 diagram Methods 0.000 claims description 20
- 238000011176 pooling Methods 0.000 claims description 16
- 238000000605 extraction Methods 0.000 claims description 12
- 230000007246 mechanism Effects 0.000 claims description 7
- 230000008569 process Effects 0.000 claims description 7
- 230000004044 response Effects 0.000 claims description 6
- 238000007635 classification algorithm Methods 0.000 claims description 5
- 230000004913 activation Effects 0.000 claims description 4
- 206010012689 Diabetic retinopathy Diseases 0.000 claims description 3
- 208000010412 Glaucoma Diseases 0.000 claims description 3
- 208000035719 Maculopathy Diseases 0.000 claims description 3
- 206010058990 Venous occlusion Diseases 0.000 claims description 3
- 239000003086 colorant Substances 0.000 claims description 3
- 230000006835 compression Effects 0.000 claims description 3
- 238000007906 compression Methods 0.000 claims description 3
- 238000013461 design Methods 0.000 claims description 3
- 230000004402 high myopia Effects 0.000 claims description 3
- 230000003902 lesion Effects 0.000 claims description 3
- 208000002780 macular degeneration Diseases 0.000 claims description 3
- 238000005070 sampling Methods 0.000 claims description 3
- 206010025421 Macule Diseases 0.000 claims 1
- 238000012360 testing method Methods 0.000 abstract description 7
- 230000000007 visual effect Effects 0.000 abstract description 5
- 238000013145 classification model Methods 0.000 abstract description 3
- 238000003759 clinical diagnosis Methods 0.000 abstract description 3
- 238000013473 artificial intelligence Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 2
- 230000007547 defect Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 230000002708 enhancing effect Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 210000004220 fundus oculi Anatomy 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000002203 pretreatment Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000012549 training Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/136—Segmentation; Edge detection involving thresholding
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/90—Determination of colour characteristics
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
- G06V10/765—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects using rules for classification or partitioning the feature space
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30041—Eye; Retina; Ophthalmic
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02A—TECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
- Y02A90/00—Technologies having an indirect contribution to adaptation to climate change
- Y02A90/10—Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Multimedia (AREA)
- Evolutionary Computation (AREA)
- Computing Systems (AREA)
- Artificial Intelligence (AREA)
- Databases & Information Systems (AREA)
- Software Systems (AREA)
- Quality & Reliability (AREA)
- Radiology & Medical Imaging (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Eye Examination Apparatus (AREA)
- Image Processing (AREA)
Abstract
The invention provides a fundus disease detection method based on cascade model fusion, which comprises the following steps: collecting an original fundus image; preprocessing an image and inputting the image into a segmentation network; extracting the fundus visual cup test disc, blood vessels and macular area part images by adopting a segmentation network; inputting the preprocessed image and the segmentation result into a healthy/unhealthy fundus image classifier, and judging whether the image is healthy fundus; the preprocessed image and the segmentation result are input into a fundus disease seed classifier to judge fundus diseases. The invention performs accurate part segmentation based on the fundus image automatic segmentation network, combines the fundus image automatic segmentation network with the original image for diagnosis, improves the algorithm interpretability, and the segmentation result is helpful for clinical diagnosis of doctors and can rapidly position the focus. According to the invention, the single-model processing single disease type combining mode precision can be achieved only by using a single-segmentation model and double-classification model cascading mode, the problem that the number of models is continuously increased along with the continuous increase of the number of disease types is avoided, the diagnosis efficiency is improved, and the operation cost is reduced.
Description
Technical Field
The invention belongs to the technical field of image detection, and particularly relates to a fundus disease detection method based on cascade model fusion.
Background
In recent years, the incidence rate of various diseases of ocular fundus is high, and the damage is serious. Fundus colour photography is the simplest and most effective method for finding fundus diseases, and a doctor generally collects fundus images by using a professional fundus camera and then performs manual diagnosis. Fundus examination requires a fundus professional doctor and corresponding examination equipment, so that the doctor is difficult to cultivate, the equipment investment is high, and a great number of people in fundus disease examination demands exist in primary hospitals and physical examination centers, so that the problem is very serious. In addition, even if the number of examination instruments and doctors is sufficient, the diagnosis of fundus diseases is very dependent on the personal status and experience of the doctor. Therefore, with the continuous development of deep learning, artificial intelligence is gradually introduced to assist medical diagnosis, so that diagnosis suggestions of fundus oculi multiple diseases can be provided for doctors, and diagnosis efficiency and accuracy are improved.
However, the existing artificial intelligence disease diagnosis method has the following defects due to the great difference between fundus color images and natural images: (1) In the current fundus disease diagnosis method based on the deep learning algorithm, attention to important parts is lacked, interpretability and pertinence are lacked, and practical help is difficult to be provided for clinical diagnosis; (2) At present, a single model is basically adopted in the multi-disease diagnosis method, but the single model cannot well cope with the extremely unbalanced fundus disease label distribution in the actual situation, so that the detection precision is low; (3) The mode that single model that commonly uses in the prior art handles single disease and makes up can improve detection accuracy, but with high costs, process time is long, and the flow is loaded down with trivial details, is unfavorable for using widely.
Disclosure of Invention
Aiming at the defects in the prior art, the invention provides a fundus disease detection method based on cascade model fusion, which can automatically extract fundus important part information such as blood vessels, visual cup test discs and the like, can provide the segmented parts for operators such as doctors to improve diagnosis efficiency, and can also provide features for further disease intelligent diagnosis.
The present invention achieves the above technical object by the following means.
A fundus disease detection method based on cascade model fusion comprises the following steps:
step 1: collecting an original fundus image and inputting the original fundus image into image processing equipment;
step 2: the image processing device preprocesses the original fundus image through a preprocessing algorithm, and inputs the preprocessed image into a segmentation network;
step 3: extracting the fundus visual cup test disc, blood vessel and macular region part images by adopting a segmentation network, and inputting segmentation results into data receiving equipment; the segmentation network comprises an encoder module, a hole convolution module, an attention module and a decoder module;
step 4: inputting the preprocessed image and the segmentation result received by the data receiving device into a healthy/unhealthy fundus image classifier together, judging whether the image is a healthy fundus, if so, entering a step 6, otherwise, entering a step 5;
step 5: the preprocessed images and the segmentation result received by the data receiving equipment are input into a fundus disease type classifier together, and fundus disease types corresponding to the fundus images are judged;
step 6: and storing the specific fundus disease diagnosis conclusion corresponding to the fundus image into a data server, and returning the specific fundus disease diagnosis conclusion to an interface calling party for viewing.
Further, the pretreatment method in the step 2 is as follows:
step 2.1: converting the original fundus image into a gray scale image, wherein the gray scale value of each pixel point at the moment is 0-255;
step 2.2: creating a mask image with the same size as the gray image, setting a threshold value, marking the threshold value as X, setting a pixel point with a gray value lower than the threshold value in the gray image as false, and setting a pixel point with a gray value higher than the threshold value as true;
step 2.3: respectively carrying out OR operation on the mask graph according to the rows (columns), if the result is false, indicating that the gray values of the pixel points of the whole row (column) are lower than the threshold value, and cutting off the row (column); if the result is true, indicating that the gray values of the pixel points in the whole row (column) are higher than or equal to the threshold value, and reserving the row (column);
step 2.4: merging the original fundus image into the mask image after the operation to obtain a region of interest with a black region removed;
step 2.5: normalizing the image colors, performing Gaussian blur processing on the combined image, reversely superposing the combined image with the original fundus image, and moving the pixel color mean value to 128.
Further, in the step 3, the specific method for extracting the fundus cup test disc, the blood vessel and the macular region part image by the segmentation network is as follows:
step 3.1: the encoder module adopts a physical sign extraction part of an Efficient Net-V2 convolution network to perform feature extraction on the preprocessed image to obtain core coding features and multi-stage features, and the core coding features and the multi-stage features are respectively input into the cavity convolution module and the attention module;
step 3.2: the cavity convolution module is built in the core part of the segmentation network, adopts three cavity convolution layers with cavity coefficients to extract and combine the characteristics of the input characteristic graphs, finally obtains multi-receptive field characteristics, and inputs the multi-receptive field characteristics to the decoder module;
the attention module comprises a spatial attention module and a channel attention module, wherein the spatial attention module weights through different pixel points of the same feature map so as to improve the feature response of a focus part in the feature map, and the channel attention module weights each feature map by gathering all the feature maps so as to selectively emphasize certain feature maps; finally, fusing the output characteristics of the spatial attention module and the channel attention module to obtain a final weight redistribution characteristic, and inputting the final weight redistribution characteristic to a decoder module;
step 3.3: the decoder module generates a corresponding size segmentation map through multi-stage up-sampling of the extracted features, wherein the segmentation map is a binarization result, 1 represents a target area, and 0 represents a background area; the decoder module outputs a three-channel segmentation chart finally, and the segmentation chart represents segmentation results of three eye fundus key positions, namely a visual cup test disc, blood vessels and a macular region, and the segmentation results are further input into the data receiving equipment.
Further, the encoder module has compression and activation on channels at each node to improve the response of key channel characteristics, the characteristic extraction of the encoder module is split into 5 stages, each stage adopts 3×3 convolution on the output characteristic diagram, the step length is 2, and the same size mode as the original diagram is adopted for filling, so that the size of the final output characteristic diagram of each stage is half of that of the previous stage; the module shallow layer of the cavity convolution module obtains adjacent information of the feature map so as to improve the identification capability of small targets, and the module deep layer obtains receptive fields similar to the feature map so as to improve the identification capability of large targets.
Further, in the step 4, the healthy/unhealthy fundus image classifier adopts an afflicientnetv 2-M network structure with an attention module added, a core module MBConv in the afflicientnet has a channel attention mechanism and lacks a spatial attention mechanism, a reference model integrated attention module is designed at the end of the model, a pooling layer and a full-connection layer of the model are removed, the attention module is replaced by an attention net, the extracted features of the model are directly input into the attention module, and a mask map generated by the attention module is multiplied by an original feature map, so that spatial attention is obtained, and expression of spatially useless features is inhibited.
Further, the classification algorithm process of the healthy/unhealthy fundus image classifier is as follows:
step 4.1: the preprocessed image and the segmentation map of the three fundus important parts are input into a transducer model together for feature extraction;
step 4.2: the space weight is redistributed to the characteristics extracted by the transducer by adopting a space attention module as a characteristic diagram so as to improve the characteristic expression of the important attention part;
step 4.3: carrying out pooling treatment on the characteristics extracted by the Transformer in a generalized mean pooling mode, wherein an index factor p is set as Y, the range of Y is dynamically adjusted according to judgment precision, the value range is 2-4, and the initialization is 3;
step 4.4: the pooled results are output through the full-connection layer, and the full-connection layer of the healthy/unhealthy fundus image classifier outputs 2 nodes which respectively represent the healthy or unhealthy probabilities of fundus images.
Further, the step 5, the fundus disease seed classifier adopts EfficientNetV2-S added with an attention module, and the design of the attention module and the method for adding the attention module are the same as those of the healthy/unhealthy fundus image classifier;
the fundus disease type classifier outputs six classification results for judging which specific fundus disease the fundus image belongs to, namely the fully connected layer of the fundus disease type classifier outputs 6 nodes which respectively represent the probability of the fundus image suffering from high myopia, maculopathy, glaucoma, venous obstruction, diabetic retinopathy or other lesions.
The invention has the following beneficial effects:
(1) According to the fundus disease detection method provided by the invention, the accurate part segmentation result is carried out based on the fundus image automatic segmentation network, the fundus image automatic segmentation network and the original image are diagnosed together, so that the algorithm interpretability is improved, the segmentation result is beneficial to clinical diagnosis of doctors, and the focus is positioned rapidly.
(2) According to the invention, the actual distribution of the fundus data set in reality is obtained through long-time data acquisition and analysis of a plurality of hospitals, wherein the healthy fundus accounts for nearly half of the proportion of the sample, and in order to fully utilize the characteristic that the neural network has better performance on the balanced data set, a double-classification model cascading mode is adopted, so that the training and sample distribution balance prediction of the healthy/unhealthy fundus image classifier and the fundus disease classifier are improved, and the accuracy of classifying specific disease types is effectively improved.
(2) The invention only uses a single segmentation model and double classification model cascade mode, can achieve the precision of single model processing single disease and combination mode, avoids the problem of continuously increasing model quantity along with the continuously increasing disease quantity, effectively improves diagnosis efficiency and reduces operation cost.
Drawings
FIG. 1 is a flow chart of a fundus disease detection method based on cascading model fusion;
FIG. 2 is a diagram of a split network framework;
FIG. 3 is an algorithmic schematic of a split network;
FIG. 4 is a block diagram of a hole convolution module;
FIG. 5 is a block diagram of an attention module;
FIG. 6 is a block diagram of a spatial attention module;
FIG. 7 is a block diagram of a channel attention module;
FIG. 8 is a flowchart of a preprocessing algorithm;
FIG. 9 is a flowchart of a classification algorithm;
FIG. 10 is a schematic diagram of a reference model in combination with an attention module;
fig. 11 is a schematic diagram of an attritionnet structure.
Detailed Description
The invention will be further described with reference to the drawings and the specific embodiments, but the scope of the invention is not limited thereto.
A fundus disease detection method based on cascade model fusion is shown in figure 1, and comprises the following steps:
step 1: acquiring an original fundus image by using fundus photographing equipment, and inputting the original fundus image into image processing equipment, wherein the image processing equipment is generally a desktop computer;
step 2: the image processing device preprocesses the original fundus image through a preprocessing algorithm, and inputs the preprocessed image into a segmentation network;
referring to fig. 8, a specific method of pretreatment is as follows:
step 2.1: converting the original fundus image into a gray scale image, wherein the gray scale value of each pixel point at the moment is 0-255;
step 2.2: creating a mask image with the same size as the gray image, setting a threshold value, wherein the threshold value is marked as X, the X in the embodiment takes 7, and the pixel point with the gray value lower than the threshold value in the gray image is set as false, and the pixel point with the gray value higher than the threshold value is set as true;
step 2.3: performing OR operation on the mask image according to the rows (columns) respectively, if the result is false, indicating that the gray values of the pixel points of the whole row (column) are lower than the threshold value, and cutting off the row (column); in order to fully preserve the details of the boundary, if the result is true, the gray values of the pixel points of the whole row (column) are higher than or equal to a threshold value, and the row (column) is preserved;
step 2.4: merging the original fundus image into the mask image after the operation to obtain a region of interest with a black region removed;
step 2.5: normalizing the image colors, performing Gaussian blur processing on the combined image, reversely superposing the combined image with the original fundus image, and moving the pixel color mean value to 128.
Step 3: referring to fig. 2 and 3, a segmentation network is used for extracting fundus cup test disc, blood vessel and macular region part images, and comprises an encoder module, a cavity convolution module, an attention module and a decoder module;
step 3.1: the encoder module adopts a physical sign extraction part of an Efficient Net-V2 convolution network to perform feature extraction on the preprocessed image to obtain core coding features and multi-stage features, and the core coding features and the multi-stage features are respectively input into the cavity convolution module and the attention module;
the encoder module has compression and activation on channels at each node so as to improve the response of key channel characteristics, and effectively solves the problem of complex characteristics of the color image of the eye bottom; the characteristic extraction of the encoder module is split into 5 stages, each stage adopts 3×3 convolution on the output characteristic diagram, the step length is 2, and the characteristic diagram is filled in the same size mode as the original diagram, so that the size of the final output characteristic diagram of each stage is half of that of the previous stage;
step 3.2: in order to improve the recognition precision of a segmentation network to different targets and enrich the network receptive field and obtain different receptive field features, a cavity convolution module shown in fig. 4 is built in the core part of the segmentation network, and the cavity convolution module extracts and combines features of input feature images by adopting cavity convolution layers with three cavity coefficients, finally obtains multi-receptive field features and inputs the multi-receptive field features to a decoder module; the module shallow layer of the cavity convolution module obtains adjacent information of the feature map, so that the recognition capability of a small target can be improved, the module deep layer obtains a receptive field similar to the feature map, and the recognition capability of a large target can be improved;
the attention module is used for enhancing key feature expression to be segmented, and carrying out weighting processing on a space and a channel aiming at an input feature map; 5, 6 and 7, the attention module comprises a spatial attention module and a channel attention module, wherein the spatial attention module weights different pixels of the same feature map so as to improve the feature response of the effective position (focus part) in the feature map, and the channel attention module selectively emphasizes certain feature maps by gathering all the feature maps and weighting each feature map respectively; finally, fusing the output characteristics of the spatial attention module and the channel attention module to obtain better characteristic representation, namely obtaining a final weight redistribution characteristic, and inputting the final weight redistribution characteristic to a decoder module;
step 3.3: the decoder module generates a corresponding size segmentation map through multi-stage up-sampling of the extracted features, wherein the segmentation map is a binarization result, 1 represents a target area, and 0 represents a background area; in this embodiment, the decoder module outputs a three-channel segmentation map, which represents segmentation results of three fundus important parts, namely, a visual cup test disc, a blood vessel and a macular region, respectively, and the segmentation results are further input into the data receiving device.
Step 4: the method comprises the steps that an image preprocessed by an image processing device and a segmentation result received by a data receiving device are input into a healthy/unhealthy fundus image classifier together, for selection of the classifier, the healthy/unhealthy fundus image classifier adopts an EfficientNetV2-M network structure with an attention module, a core module MBConv in the EfficientNet is provided with a channel attention mechanism and lacks a spatial attention mechanism, in order to further improve classification performance of the network, a basic model integrated attention module is designed at the tail end of the model, a pooling layer and a full connection layer of the model are removed, the characteristics extracted by the model are replaced by Attennet, the characteristics extracted by the model are directly input into the attention module, and a mask image generated by the attention module is multiplied with an original characteristic image, so that spatial attention is obtained, and expression of unnecessary characteristics in space is inhibited; the attention module can be applied to any model without redesigning;
the reference Model and attention module are combined as shown in FIG. 10, where Baseline Model is any Model that removes pooling and fully connected layers, F E Feature map generated for the model, F A For attention gates generated by the attention net, geM Pooling can be regarded as a more generalized Pooling operation, an exponential factor p is set for the Pooling operation, and when p=1, the Pooling operation is an average Pooling operation; p= infinity, the operation is the maximum pooling operation; the effect of stronger comprehensiveness can be obtained by taking the value between p, and experiments prove that the effect of p=3 in the data set is better; wherein, the AttenionNet structure can be expressed as C256-C128-C64-C1-C256, ck represents that k convolution cores exist in the convolution layer, the size of each convolution core is 1 multiplied by 1, the step sizes of five convolution layers are all 1, and after each convolution layer, reLU is used as an activation function, the number of final output channels is 1, and the attention gate F with the same size as the original feature map A The process is shown in fig. 11;
outputting a classification result after the processing of the healthy/unhealthy fundus image classifier, wherein the classification result indicates whether the fundus image is judged to be a healthy fundus, if so, the step 6 is entered, and if not, the step 5 is entered;
wherein, referring to fig. 9, the classification algorithm process of the healthy/unhealthy fundus image classifier is as follows:
step 4.1: the preprocessed image and the segmentation map of the three fundus important parts are input into a transducer model together for feature extraction;
step 4.2: the space weight is redistributed to the characteristics extracted by the transducer by adopting a space attention module as a characteristic diagram so as to improve the characteristic expression of the important attention part;
step 4.3: carrying out pooling treatment on the characteristics extracted by the Transformer in a generalized mean pooling mode, wherein an index factor p is set as Y, the range of Y is dynamically adjusted according to judgment precision, the value range is 2-4, and the initialization is 3;
step 4.4: the pooled result is output through a full connection layer, and the full connection layer of the healthy/unhealthy fundus image classifier outputs 2 nodes which respectively represent the healthy or unhealthy probability of fundus images;
in the process, the global self-attention mechanism of the transducer is good in capturing global features, can effectively cope with complex relations among the global features in the medical image, and captures relations among different features after the sequence.
Step 5: the image preprocessed by the image processing equipment and the segmentation result received by the data receiving equipment are input into a fundus disease type classifier together, wherein the fundus disease type classifier adopts EfficientNetV2-S with an attention module added, and the design of the attention module and the method for adding the attention module are the same as those of a healthy/unhealthy fundus image classifier. The fundus disease type classifier outputs six classification results for judging which specific fundus disease the fundus image belongs to; the classification algorithm of the fundus disease seed classifier is similar to that of the healthy/unhealthy fundus image classifier, except that the full connection layer of the fundus disease seed classifier outputs 6 nodes, which respectively represent the probability of the fundus image having high myopia, maculopathy, glaucoma, venous obstruction, diabetic retinopathy or other lesions, so that specific algorithm processes are not repeated here.
Step 6: and storing the specific fundus disease diagnosis conclusion corresponding to the fundus image into a data server, and returning the specific fundus disease diagnosis conclusion to an interface calling party for viewing.
The examples are preferred embodiments of the present invention, but the present invention is not limited to the above-described embodiments, and any obvious modifications, substitutions or variations that can be made by one skilled in the art without departing from the spirit of the present invention are within the scope of the present invention.
Claims (6)
1. The fundus disease detection method based on cascade model fusion is characterized by comprising the following steps of:
step 1: collecting an original fundus image and inputting the original fundus image into image processing equipment;
step 2: the image processing device preprocesses the original fundus image through a preprocessing algorithm, and inputs the preprocessed image into a segmentation network;
step 3: extracting the eye cup optic disc, blood vessel and macula area part images of the fundus by adopting a segmentation network, and inputting the segmentation result into data receiving equipment; the segmentation network comprises an encoder module, a hole convolution module, an attention module and a decoder module;
step 4: inputting the preprocessed image and the segmentation result received by the data receiving device into a healthy/unhealthy fundus image classifier together, judging whether the image is a healthy fundus, if so, entering a step 6, otherwise, entering a step 5;
step 5: the preprocessed images and the segmentation result received by the data receiving equipment are input into a fundus disease type classifier together, and fundus disease types corresponding to the fundus images are judged;
step 6: storing the specific fundus disease diagnosis conclusion corresponding to the fundus image into a data server, and returning the specific fundus disease diagnosis conclusion to an interface calling party for viewing;
in the step 3, the specific method for extracting the fundus cup optic disc, blood vessel and macular region part images by the segmentation network comprises the following steps:
step 3.1: the encoder module adopts a physical sign extraction part of an Efficient Net-V2 convolution network to perform feature extraction on the preprocessed image to obtain core coding features and multi-stage features, and the core coding features and the multi-stage features are respectively input into the cavity convolution module and the attention module;
step 3.2: the cavity convolution module is built in the core part of the segmentation network, adopts three cavity convolution layers with cavity coefficients to extract and combine the characteristics of the input characteristic graphs, finally obtains multi-receptive field characteristics, and inputs the multi-receptive field characteristics to the decoder module;
the attention module comprises a spatial attention module and a channel attention module, wherein the spatial attention module weights through different pixel points of the same feature map so as to improve the feature response of a focus part in the feature map, and the channel attention module weights each feature map by gathering all the feature maps so as to selectively emphasize certain feature maps; finally, fusing the output characteristics of the spatial attention module and the channel attention module to obtain a final weight redistribution characteristic, and inputting the final weight redistribution characteristic to a decoder module;
step 3.3: the decoder module generates a corresponding size segmentation map through multi-stage up-sampling of the extracted features, wherein the segmentation map is a binarization result, 1 represents a target area, and 0 represents a background area; the decoder module outputs a three-channel segmentation chart finally, and the segmentation chart represents segmentation results of three eye fundus important parts, namely the optic disc of the eye cup, the blood vessel and the macular area, and the segmentation results are further input into the data receiving equipment.
2. The fundus disease detection method based on cascading model fusion according to claim 1, wherein the preprocessing method in step 2 is as follows:
step 2.1: converting the original fundus image into a gray scale image, wherein the gray scale value of each pixel point at the moment is 0-255;
step 2.2: creating a mask image with the same size as the gray image, setting a threshold value, marking the threshold value as X, setting a pixel point with a gray value lower than the threshold value in the gray image as false, and setting a pixel point with a gray value higher than the threshold value as true;
step 2.3: performing or operation on the mask image according to the row/column respectively, if the result is false, indicating that the gray values of the pixel points of the row/column are lower than a threshold value, and cutting off the row/column; if the result is true, indicating that the gray values of the pixel points in the row/column are higher than or equal to the threshold value, and reserving the row/column;
step 2.4: merging the original fundus image into the mask image after the operation to obtain a region of interest with a black region removed;
step 2.5: normalizing the image colors, performing Gaussian blur processing on the combined image, reversely superposing the combined image with the original fundus image, and moving the pixel color mean value to 128.
3. The method for detecting fundus disease based on cascade model fusion according to claim 1, wherein the encoder module has compression and activation on channels at each node to improve response of key channel characteristics, the characteristic extraction of the encoder module is split into 5 stages, each stage convolves the output characteristic diagram with 3×3, the step length is 2, and the same size as the original diagram is adopted for filling, so that the size of the final output characteristic diagram of each stage is half of that of the previous stage; the module shallow layer of the cavity convolution module obtains adjacent information of the feature map so as to improve the identification capability of small targets, and the module deep layer obtains receptive fields similar to the feature map so as to improve the identification capability of large targets.
4. The method according to claim 1, wherein in the step 4, the healthy/unhealthy fundus image classifier adopts an affliction netv2-M network structure with an attention module, wherein the kernel module MBConv in affliction nett has a channel attention mechanism and lacks a spatial attention mechanism, the reference model integrated attention module is designed at the end of the model, the pooling layer and the full connection layer of the model are removed, the attention module is replaced by the attainmet, the extracted features of the model are directly input into the attention module, and the mask map generated by the attention module is multiplied with the original feature map, so that spatial attention is obtained, and expression of the spatially useless features is inhibited.
5. The fundus disease detection method based on cascading model fusion according to claim 1, wherein the classification algorithm process of the healthy/unhealthy fundus image classifier is as follows:
step 4.1: the preprocessed image and the segmentation map of the three fundus important parts are input into a transducer model together for feature extraction;
step 4.2: the space weight is redistributed to the characteristics extracted by the transducer by adopting a space attention module as a characteristic diagram so as to improve the characteristic expression of the important attention part;
step 4.3: carrying out pooling treatment on the characteristics extracted by the Transformer in a generalized mean pooling mode, wherein an index factor p is set as Y, the range of Y is dynamically adjusted according to judgment precision, the value range is 2-4, and the initialization is 3;
step 4.4: the pooled results are output through the full-connection layer, and the full-connection layer of the healthy/unhealthy fundus image classifier outputs 2 nodes which respectively represent the healthy or unhealthy probabilities of fundus images.
6. The method for detecting fundus disease based on cascade model fusion according to claim 4, wherein in the step 5, the fundus disease classifier adopts the affliction netv2-S with an attention module added, and the design of the attention module and the method for adding the attention module are the same as those of the healthy/unhealthy fundus image classifier;
the fundus disease type classifier outputs six classification results for judging which specific fundus disease the fundus image belongs to, namely the fully connected layer of the fundus disease type classifier outputs 6 nodes which respectively represent the probability of the fundus image suffering from high myopia, maculopathy, glaucoma, venous obstruction, diabetic retinopathy or other lesions.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311053124.1A CN117078697B (en) | 2023-08-21 | 2023-08-21 | Fundus disease seed detection method based on cascade model fusion |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311053124.1A CN117078697B (en) | 2023-08-21 | 2023-08-21 | Fundus disease seed detection method based on cascade model fusion |
Publications (2)
Publication Number | Publication Date |
---|---|
CN117078697A CN117078697A (en) | 2023-11-17 |
CN117078697B true CN117078697B (en) | 2024-04-09 |
Family
ID=88707578
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202311053124.1A Active CN117078697B (en) | 2023-08-21 | 2023-08-21 | Fundus disease seed detection method based on cascade model fusion |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117078697B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN118195924B (en) * | 2024-05-17 | 2024-07-26 | 南昌大学第二附属医院 | Premature infant retinopathy analysis system based on image recognition |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018222755A1 (en) * | 2017-05-30 | 2018-12-06 | Arterys Inc. | Automated lesion detection, segmentation, and longitudinal identification |
CN111046835A (en) * | 2019-12-24 | 2020-04-21 | 杭州求是创新健康科技有限公司 | Eyeground illumination multiple disease detection system based on regional feature set neural network |
CN111938569A (en) * | 2020-09-17 | 2020-11-17 | 南京航空航天大学 | Eye ground multi-disease classification detection method based on deep learning |
CN113256641A (en) * | 2021-07-08 | 2021-08-13 | 湖南大学 | Skin lesion image segmentation method based on deep learning |
CN114287878A (en) * | 2021-10-18 | 2022-04-08 | 江西财经大学 | Diabetic retinopathy focus image identification method based on attention model |
CN115205300A (en) * | 2022-09-19 | 2022-10-18 | 华东交通大学 | Fundus blood vessel image segmentation method and system based on cavity convolution and semantic fusion |
CN115661066A (en) * | 2022-10-21 | 2023-01-31 | 东北林业大学 | Diabetic retinopathy detection method based on segmentation and classification fusion |
CN116309631A (en) * | 2023-02-08 | 2023-06-23 | 大连交通大学 | Automatic segmentation method for fundus image of sugar net disease based on deep learning |
CN116579982A (en) * | 2023-03-30 | 2023-08-11 | 苏州大学 | Pneumonia CT image segmentation method, device and equipment |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20210374955A1 (en) * | 2020-06-02 | 2021-12-02 | Zasti Inc. | Retinal color fundus image analysis for detection of age-related macular degeneration |
-
2023
- 2023-08-21 CN CN202311053124.1A patent/CN117078697B/en active Active
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018222755A1 (en) * | 2017-05-30 | 2018-12-06 | Arterys Inc. | Automated lesion detection, segmentation, and longitudinal identification |
CN111046835A (en) * | 2019-12-24 | 2020-04-21 | 杭州求是创新健康科技有限公司 | Eyeground illumination multiple disease detection system based on regional feature set neural network |
CN111938569A (en) * | 2020-09-17 | 2020-11-17 | 南京航空航天大学 | Eye ground multi-disease classification detection method based on deep learning |
CN113256641A (en) * | 2021-07-08 | 2021-08-13 | 湖南大学 | Skin lesion image segmentation method based on deep learning |
CN114287878A (en) * | 2021-10-18 | 2022-04-08 | 江西财经大学 | Diabetic retinopathy focus image identification method based on attention model |
CN115205300A (en) * | 2022-09-19 | 2022-10-18 | 华东交通大学 | Fundus blood vessel image segmentation method and system based on cavity convolution and semantic fusion |
CN115661066A (en) * | 2022-10-21 | 2023-01-31 | 东北林业大学 | Diabetic retinopathy detection method based on segmentation and classification fusion |
CN116309631A (en) * | 2023-02-08 | 2023-06-23 | 大连交通大学 | Automatic segmentation method for fundus image of sugar net disease based on deep learning |
CN116579982A (en) * | 2023-03-30 | 2023-08-11 | 苏州大学 | Pneumonia CT image segmentation method, device and equipment |
Non-Patent Citations (5)
Title |
---|
ADD-Net:Attention U-Net with Dilated Skip Connection and Dense Connected Decoder for Retinal Vessel Segmentation;Dongjin Huang 等;《CGI 2021: Advances in Computer Graphics》;20211011;全文 * |
Automatic grading of Diabetic macular edema based on end-to-end network;Yinghua Fu 等;《Expert Systems with Applications》;20220916;第213卷;全文 * |
EAD-Net: A Novel Lesion Segmentation Method in Diabetic Retinopathy Using Neural Networks;Cheng Wan 等;《Disease Markers》;20210902;全文 * |
基于人工智能ResNeXt的高度近视诊断方法;万程 等;《实用老年医学》;20220331;第36卷(第3期);全文 * |
基于超像素分类及径向射线扫描动态优化的视盘分割方法;陈瑞 等;《国外电子测量技术》;20230228;第42卷(第2期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN117078697A (en) | 2023-11-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108021916B (en) | Deep learning diabetic retinopathy sorting technique based on attention mechanism | |
CN111815574B (en) | Fundus retina blood vessel image segmentation method based on rough set neural network | |
CN108806792B (en) | Deep learning face diagnosis system | |
CN110930418B (en) | Retina blood vessel segmentation method fusing W-net and conditional generation confrontation network | |
Claro et al. | An hybrid feature space from texture information and transfer learning for glaucoma classification | |
CN110766643A (en) | Microaneurysm detection method facing fundus images | |
CN110197493A (en) | Eye fundus image blood vessel segmentation method | |
CN111046835A (en) | Eyeground illumination multiple disease detection system based on regional feature set neural network | |
CN112508864B (en) | Retinal vessel image segmentation method based on improved UNet + | |
CN109300121A (en) | A kind of construction method of cardiovascular disease diagnosis model, system and the diagnostic model | |
Liu et al. | A framework of wound segmentation based on deep convolutional networks | |
CN108095683A (en) | The method and apparatus of processing eye fundus image based on deep learning | |
CN110751636B (en) | Fundus image retinal arteriosclerosis detection method based on improved coding and decoding network | |
CN108764342B (en) | Semantic segmentation method for optic discs and optic cups in fundus image | |
CN113012163A (en) | Retina blood vessel segmentation method, equipment and storage medium based on multi-scale attention network | |
CN117078697B (en) | Fundus disease seed detection method based on cascade model fusion | |
CN114565620B (en) | Fundus image blood vessel segmentation method based on skeleton prior and contrast loss | |
CN115035127A (en) | Retinal vessel segmentation method based on generative confrontation network | |
CN115294075A (en) | OCTA image retinal vessel segmentation method based on attention mechanism | |
CN112884788A (en) | Cup optic disk segmentation method and imaging method based on rich context network | |
CN110992309B (en) | Fundus image segmentation method based on deep information transfer network | |
CN110610480A (en) | MCASPP neural network eyeground image optic cup optic disc segmentation model based on Attention mechanism | |
Lwin et al. | Image Classification for Rice Leaf Disease Using AlexNet Model | |
Saranya et al. | Detection of exudates from retinal images for non-proliferative diabetic retinopathy detection using deep learning model | |
CN116092667A (en) | Disease detection method, system, device and storage medium based on multi-mode images |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |