CN112132782B - Method and terminal for processing DME (DME) typing based on deep neural network - Google Patents

Method and terminal for processing DME (DME) typing based on deep neural network Download PDF

Info

Publication number
CN112132782B
CN112132782B CN202010851829.8A CN202010851829A CN112132782B CN 112132782 B CN112132782 B CN 112132782B CN 202010851829 A CN202010851829 A CN 202010851829A CN 112132782 B CN112132782 B CN 112132782B
Authority
CN
China
Prior art keywords
dme
model
deep
neural network
oct image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010851829.8A
Other languages
Chinese (zh)
Other versions
CN112132782A (en
Inventor
余洪华
蔡宏民
吴乔伟
张滨
刘宝怡
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong General Hospital
Original Assignee
Guangdong General Hospital
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong General Hospital filed Critical Guangdong General Hospital
Priority to CN202010851829.8A priority Critical patent/CN112132782B/en
Publication of CN112132782A publication Critical patent/CN112132782A/en
Application granted granted Critical
Publication of CN112132782B publication Critical patent/CN112132782B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10101Optical tomography; Optical coherence tomography [OCT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30041Eye; Retina; Ophthalmic
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/10Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computing Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Medical Informatics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Quality & Reliability (AREA)
  • Radiology & Medical Imaging (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a method and a terminal for processing DME typing based on a deep neural network, wherein the method comprises the following steps: preprocessing an OCT image to be identified; performing image feature extraction on the OCT image subjected to pretreatment through a trained DME feature extraction model; the DME feature extraction model is obtained based on deep learning network training; based on the extracted image features, obtaining a binary classification function value corresponding to whether preset DME appears in the OCT image; the preset DME includes: DRT, CME and SRD; and obtaining a binary classification task result based on the binary classification function value and a preset threshold value. The scheme realizes the rapid and accurate identification of different types of DMEs.

Description

Method and terminal for processing DME (DME) typing based on deep neural network
Technical Field
The invention relates to the field of clinical medicine ophthalmology and computer engineering, in particular to a method and a terminal for processing DME typing based on a deep neural network.
Background
DME is the leading cause of vision loss in diabetics and can be classified into diffuse retinal thickening (DRT, diffuse retinal thickening), macular cystoid edema (CME, cystoid macular edema), serous retinal detachment (SRD, serous retinal detachment) and mixed DME (Mixed DME) of two or three types depending on its morphology in OCT examination. On OCT images, DRT appears as an increase in retinal nerve epithelial layer thickness with a decrease in inter-nerve epithelial layer reflection, and an enlargement of the low reflection region; CME is manifested as a retinal interlayer cyst-like cavity formation, while the lighter one shows several small cyst cavities in honeycomb form, which can merge into a larger cyst cavity when edema is evident, even retaining only a thin inner limiting membrane in the fovea; SRD is manifested as detachment of the nerve epithelium layer in the macular region, underlying the liquid dark space, where the highly reflective bands of the pigment epithelium layer are clearly visible.
Studies have shown that the systemic risk factors for different types of DME patients are different, indicating that the pathogenesis of different types of DME may vary. Previous studies have further demonstrated that DRT and CME are mainly associated with intracellular edema and necrosis of Muller cells caused by disruption of the blood-retinal barrier, respectively. In contrast, SRD occurs in association with blood-subretinal barrier dysfunction causing accumulation of subretinal fluid. Intravitreal injection of anti-VEGF drugs is currently the most prominent means of treating DME, but different morphological types of DME on OCT images may respond differently to anti-VEGF treatment due to different pathogenesis.
According to the above conventional predictive methods, ophthalmologists are generally presented in the field of accurately classifying DME from OCT images and formulating personalized treatment strategies for DME patients. The main reasons for this are that traditional prediction methods are subjective, have many and complex factors to consider comprehensively, and depend largely on the clinical experience and knowledge level of the ophthalmologist. And with the increasing prevalence of diabetes mellitus worldwide, it may place a great burden on the clinical diagnosis and management of DME patients. Thus, there is a great difficulty in making accurate DME classifications for young doctors lacking clinical experience and doctors in communities or primary hospitals where medical levels are relatively low. At present, no means for automatically and accurately classifying the DME is available for clinic, and the means are rapid, accurate, high in universality and wide in application range.
Disclosure of Invention
Aiming at the defects in the prior art, the invention provides a method and a terminal for processing DME typing based on a deep neural network,
specifically, the present invention proposes the following specific embodiments:
the embodiment of the invention provides a method for processing DME typing based on a deep neural network, which comprises the following steps:
preprocessing an OCT image to be identified;
performing image feature extraction on the OCT image subjected to pretreatment through a trained DME feature extraction model; the DME feature extraction model is obtained based on deep learning network training;
based on the extracted image features, obtaining a binary classification function value corresponding to whether preset DME appears in the OCT image; the preset DME includes: DRT, CME and SRD;
and obtaining a binary classification task result based on the binary classification function value and a preset threshold value.
In a specific embodiment, before "preprocessing the OCT image to be identified", it further includes:
acquiring OCT images of a DME-confirmed patient;
setting the OCT image of the patient as the OCT image to be identified.
In a specific embodiment, the "preprocessing the OCT image to be identified" includes:
removing saturated pixel values, reducing noise and cutting edges of the OCT image to be identified;
and reading the processed OCT images, and uniformly adjusting the read OCT images to a preset size.
In a specific embodiment, the DME feature extraction model is constructed on the basis of a VGG16 convolutional neural network.
In a specific embodiment, the DME feature extraction model is constructed on the basis of a VGG16 convolutional neural network, and the DME feature extraction model comprises the following steps:
constructing a deep DME model on the basis of a VGG16 convolutional neural network;
training the deep DME model based on a training set in the preprocessed preset OCT image dataset, and performing internal verification;
and carrying out external verification on the trained deep DME model based on a test set in a preset OCT image data set so as to obtain a DME feature extraction model.
In a specific embodiment, the "constructing the deep dme model on the basis of the VGG16 convolutional neural network" includes:
estimating the degree of difference between the VGG16 convolutional neural network output value and the true value by adopting a cross entropy loss function;
the ReLU activation function is adopted, so that the nonlinearity of the VGG16 convolutional neural network is improved, and the sparsity of the VGG16 convolutional neural network is increased;
the Dropout layer is adopted to act on a full connection layer of the VGG16 convolutional neural network;
after the Batchnorm layer is adopted to act on each convolution layer in the VGG16 convolution neural network, the distribution of the neuron input values of each layer of the neural network layer is pulled back to the standard normal distribution;
adjusting the VGG16 convolutional neural network which is pre-trained on the ImageNet data set, and constructing an initial model;
replacing the last full-connection layer, the SoftMax layer and the classification layer of the VGG16 convolutional neural network with a full-connection layer and a regression layer containing 3 neurons, so that the VGG16 convolutional neural network is converted into a regression network;
mapping the OCT image into OCT typing of DME based on the initial model, and respectively adjusting the output of the last fully-connected layer into binary classification function values of DRT, CME and SRD to obtain a deep DME model.
In a specific embodiment, the "training the deepdem model based on the training set in the pre-processed preset OCT image dataset" and performing internal verification "includes:
based on a training set in the preprocessed preset OCT image data set, performing internal inspection on the deep DME model by using a five-fold cross validation method; the training set is randomly divided into five parts, wherein 4 parts are used for training the deep DME model, the remaining 1 part is used for evaluating the performance of the deep DME model, and each part participates in the training of the deep DME model;
determining consistency between a predicted label and a real label obtained each time in internal verification of the deep DME model by using a confusion matrix, and calculating accuracy, sensitivity, specificity and area under a test subject working characteristic curve of each time of running of the model;
the performance of the deepdem model was verified by recording the average results of five runs of the deepdem model.
In a specific embodiment, the "externally validating the deep DME model after training based on a test set in a preset OCT image dataset to obtain a DME feature extraction model" includes:
externally verifying the trained deep DME model by using a test set in a preset OCT image data set;
determining consistency between a predicted tag and a real tag obtained by the deep DME model by using a confusion matrix;
based on the confusion matrix, calculating the accuracy, sensitivity, specificity and area under the working characteristic curve of the subject for detecting the DME typing by the deep DME model.
In a specific embodiment, the "obtaining the DME-typed result based on the binary classification function value and a preset threshold" includes:
for OCT images of the hybrid DME, binary classification function values of the DRT, the CME and the SRD and a preset threshold value are converted into the proportion of 3 types of DME in the hybrid DME by a normalization method.
The embodiment of the invention also provides a terminal which comprises a memory and a processor, wherein the processor executes the method when running the program stored in the memory.
In this way, the embodiment of the invention provides a method and a terminal for processing DME typing based on a deep neural network, wherein the method comprises the following steps: preprocessing an OCT image to be identified; performing image feature extraction on the OCT image subjected to pretreatment through a trained DME feature extraction model; the DME feature extraction model is obtained based on deep learning network training; based on the extracted image features, obtaining a binary classification function value corresponding to whether preset DME appears in the OCT image; the preset DME includes: DRT, CME and SRD; and obtaining a binary classification task result based on the binary classification function value and a preset threshold value. The scheme realizes the rapid and accurate identification of different types of DMEs.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings that are needed in the embodiments will be briefly described below, it being understood that the following drawings only illustrate some embodiments of the present invention and therefore should not be considered as limiting the scope, and other related drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a schematic flow chart of a method for processing DME typing based on a deep neural network according to an embodiment of the present invention;
FIG. 2 is a schematic diagram of a system corresponding to a method for processing DME typing based on a deep neural network according to an embodiment of the present invention;
FIG. 3 is a schematic flow chart of a method for processing DME typing based on a deep neural network according to an embodiment of the present invention;
fig. 4 is a schematic flow chart of a specific implementation process of a method for processing DME typing based on a deep neural network according to an embodiment of the present invention.
Detailed Description
Hereinafter, various embodiments of the present disclosure will be more fully described. The present disclosure is capable of various embodiments and of modifications and variations therein. However, it should be understood that: there is no intention to limit the various embodiments of the disclosure to the specific embodiments disclosed herein, but rather the disclosure is to be interpreted to cover all modifications, equivalents, and/or alternatives falling within the spirit and scope of the various embodiments of the disclosure.
The terminology used in the various embodiments of the disclosure is for the purpose of describing particular embodiments only and is not intended to be limiting of the various embodiments of the disclosure. As used herein, the singular is intended to include the plural as well, unless the context clearly indicates otherwise. Unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which various embodiments of this disclosure belong. The terms (such as those defined in commonly used dictionaries) will be interpreted as having a meaning that is the same as the context of the relevant art and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein in the various embodiments of the disclosure.
Example 1
The embodiment 1 of the invention discloses a method for processing DME typing based on a deep neural network, which comprises the following steps as shown in figure 1:
step 101, preprocessing an OCT image to be identified;
specifically, as shown in fig. 2, the step 101 corresponds to a preprocessing module, where the preprocessing module performs preprocessing on the labeled OCT image, and sends the result to a feature extraction module;
specifically, the OCT image may be in TIFF format.
102, extracting image features of the OCT image subjected to pretreatment through a trained DME feature extraction model; the DME feature extraction model is obtained based on deep learning network training;
specifically, step 102 corresponds to the feature extraction module in fig. 2, and the feature extraction module performs image feature extraction on the OCT image after preprocessing by using the deep neural network.
Step 103, based on processing the extracted image features, obtaining a binary classification function value corresponding to whether preset DME appears in the OCT image; the preset DME includes: DRT, CME and SRD;
step 103 corresponds to the data processing module in fig. 2, where the data processing module processes the image features obtained by the feature extraction module, and outputs whether the respective binary classification function values of the DRT, CME or SRD appear in the image.
And 104, obtaining a DME typing result based on the binary classification function value and a preset threshold value.
Specifically, step 104 corresponds to the decision module in fig. 2, and the decision module may divide the result into the binary classification task 0 or 1 according to the binary classification function value obtained by the data processing module, for example, a threshold of 0.5 (as shown in fig. 3 and 4, specifically, if less than 0.5, it is considered to be 0, and if not less than 0.5, it is considered to be 1).
Further, before the "preprocessing the OCT image to be identified", it further includes:
acquiring OCT images of a DME-confirmed patient;
setting the OCT image of the patient as the OCT image to be identified.
Specifically, the OCT image to be identified is an image obtained from a patient diagnosed with DME.
In a specific embodiment, the "preprocessing the OCT image to be identified" includes:
removing saturated pixel values, reducing noise and cutting edges of the OCT image to be identified;
and reading the processed OCT images, and uniformly adjusting the read OCT images to a preset size.
Specifically, for example, the OCT images are all adjusted to 224×224, or other sizes, and the specific size can be set and flexibly adjusted according to the actual situation.
In a specific embodiment, the DME feature extraction model is constructed on the basis of a VGG16 convolutional neural network.
Further, constructing and obtaining the DME characteristic extraction model on the basis of a VGG16 convolutional neural network, wherein the DME characteristic extraction model comprises the following steps:
constructing a deep DME model on the basis of a VGG16 convolutional neural network;
training the deep DME model based on a training set in the preprocessed preset OCT image dataset, and performing internal verification;
and carrying out external verification on the trained deep DME model based on a test set in a preset OCT image data set so as to obtain a DME feature extraction model.
Specifically, the "constructing the deep dme model on the basis of the VGG16 convolutional neural network" includes:
estimating the degree of difference between the VGG16 convolutional neural network output value and the true value by adopting a cross entropy loss function; generally, all values of the loss function are non-negative, and if the robustness is better, the corresponding loss value is smaller, so that the obtained result is more convincing;
the ReLU activation function is adopted, so that the nonlinearity of the VGG16 convolutional neural network is improved, and the sparsity of the VGG16 convolutional neural network is increased; by the method, the problems of overfitting effect and gradient disappearance of the model are effectively avoided, and the calculation speed is improved;
the Dropout layer is adopted to act on the full-connection layer of the VGG16 convolutional neural network, so that interaction between hidden layer nodes can be effectively reduced, complex co-adaptation relations among neurons are reduced, and overfitting is effectively avoided;
after the Batchnorm layer is adopted to act on each convolution layer in the VGG16 convolution neural network, the distribution of the neuron input values of each layer of the neural network layer is pulled back to the standard normal distribution; therefore, the training speed is improved, gradient explosion and gradient disappearance are effectively prevented, and the overfitting effect is avoided;
adjusting the VGG16 convolutional neural network which is pre-trained on the ImageNet data set, and constructing an initial model;
replacing the last full-connection layer, the SoftMax layer and the classification layer of the VGG16 convolutional neural network with a full-connection layer and a regression layer containing 3 neurons, so that the VGG16 convolutional neural network is converted into a regression network; specifically, for example, the last full-connection layer, softMax layer, and classification layer containing 1,000 neurons may be replaced with the full-connection layer and regression layer containing 3 neurons, thereby converting the classification network model into a regression network model;
mapping the OCT image into OCT typing of DME based on the initial model, and respectively adjusting the output of the last fully-connected layer into binary classification function values of DRT, CME and SRD to obtain a deep DME model.
In a specific embodiment, the specific preset OCT image dataset is a sample dataset, i.e. a professional evaluator gives an annotated OCT image of 0 or 1, respectively, according to the presence or absence of 3 DME-typing in each OCT image, each OCT image having a multiple annotation consisting of 3 labels. For example, OCT images where only DRT is present will be labeled as 1/0/0, while OCT images where CME and SRD coexist will be labeled as 0/1/1; thus, the "training the deepdem model based on the training set in the pre-processed preset OCT image dataset" includes:
based on a training set in the preprocessed preset OCT image data set, performing internal inspection on the deep DME model by using a five-fold cross validation method; the training set is randomly divided into five parts, wherein 4 parts are used for training the deep DME model, the remaining 1 part is used for evaluating the performance of the deep DME model, and each part participates in the training of the deep DME model;
determining consistency between a predicted label and a real label obtained each time in internal verification of the deep DME model by using a confusion matrix, and calculating accuracy, sensitivity, specificity and area under a test subject working characteristic curve of each time of running of the model;
the performance of the deepdem model was verified by recording the average results of five runs of the deepdem model.
Specifically, the deep dme model is internally checked by using a five-fold cross-validation method, in which the OCT image of the training set is randomly divided into five independent parts; in each run of the model, four of the parts are used for training of the model, and the remaining one part is used for evaluating the performance of the model until each part participates in training of the model; the confusion matrix is used to describe the consistency between the predictive label and the real label obtained each time in the internal verification of the deep dme model, and the accuracy, sensitivity, specificity and area under the working characteristic curve of the test subject of each operation of the model are calculated.
In a specific embodiment, the "externally validating the deep DME model after training based on a test set in a preset OCT image dataset to obtain a DME feature extraction model" includes:
externally verifying the trained deep DME model by using a test set in a preset OCT image data set;
determining consistency between a predicted tag and a real tag obtained by the deep DME model by using a confusion matrix;
based on the confusion matrix, calculating the accuracy, sensitivity, specificity and area under the working characteristic curve of the subject for detecting the DME typing by the deep DME model.
In a specific embodiment, the "obtaining the DME-typed result based on the binary classification function value and a preset threshold" includes:
for OCT images of the hybrid DME, binary classification function values of the DRT, the CME and the SRD and a preset threshold value are converted into the proportion of 3 types of DME in the hybrid DME by a normalization method.
Specifically, in a specific embodiment, by collecting 11,599 OCT images of different types of DME patients, the deep DME model automatically diagnoses the morphology typing of DME based on an image multi-labeling technology and quantitatively analyzes the respective proportion of 3 types of DME in the mixed DME, so that the defect of the conventional subjective diagnosis method depending on time consumption participation of ophthalmologists is overcome, and an ophthalmologist is helped to formulate a personalized treatment strategy according to the morphology typing of the DME patients.
The deep DME model can automatically detect the morphological typing of DME, so that the workload of an ophthalmologist is reduced, the ophthalmologist is helped to formulate a personalized treatment strategy for a DME patient, the clinical curative effect of DME is improved, the understanding of the patient on diseases is enhanced, and the psychological pressure of medical treatment is relieved. For DME patients expected to respond well to anti-VEGF treatment, the promising outcome of the patient's anti-VEGF treatment should be emphasized and patients should be encouraged to develop standard treatment regimens. For DME patients who are predicted to have poor response to VEGF treatment, additional treatment regimens should be recommended in time when they are not optimally receiving standard treatment regimens.
Compared with the traditional subjective diagnosis method requiring the participation of ophthalmologists, the deep DME model can automatically extract OCT image characteristics to objectively and accurately detect the morphological typing of DME and quantitatively analyze the proportion of each type of DME in the mixed DME, so that the deep DME model is used by clinicians in various areas and at various levels. The deep DME model of the present invention achieves superior performance in the task of detecting different types of DME. In internal verification, the accuracy of detecting DRT typing is 93.0%, the sensitivity is 93.5% and the specificity is 92.3%; the accuracy of CME typing detection is 95.1%, the sensitivity is 94.5%, and the specificity is 95.6%; the accuracy of detecting SRD typing is 98.8%, the sensitivity is 96.7%, and the specificity is 99.3%; the areas under the test operating characteristic curves of the subjects for detecting DRT, CME and SRD typing are respectively 0.971, 0.974 and 0.994. In external verification, the accuracy of detecting DRT typing is 90.2%, the sensitivity is 80.1%, and the specificity is 97.6%; the accuracy of CME typing detection is 95.4%, the sensitivity is 93.4%, and the specificity is 97.2%; the accuracy of detecting SRD typing is 95.9%, the sensitivity is 94.9%, and the specificity is 96.5%; the areas under the test subject operating characteristics curves for detecting DRT, CME and SRD typing were 0.970, 0.997 and 0.997, respectively. In the future, the method is expected to be applied to automatic diagnosis of other multi-factor eye or whole body diseases after the sample is continuously enlarged and repeated training is carried out for a plurality of times. In addition, the deep dme model of the present invention can also improve the ophthalmic diagnosis and treatment level in remote areas lacking medical resources by reducing the time and expertise required for retinal image classification.
Example 2
The embodiment 2 of the invention also discloses a terminal, which comprises a memory and a processor, wherein the processor executes the method in the embodiment 1 when running the program stored in the memory. In particular, embodiment 2 of the present invention also discloses other relevant features, and for brevity, reference is made to the description in embodiment 1.
In this way, the embodiment of the invention provides a method and a terminal for processing DME typing based on a deep neural network, wherein the method comprises the following steps: preprocessing an OCT image to be identified; performing image feature extraction on the OCT image subjected to pretreatment through a trained DME feature extraction model; the DME feature extraction model is obtained based on deep learning network training; based on the extracted image features, obtaining a binary classification function value corresponding to whether preset DME appears in the OCT image; the preset DME includes: DRT, CME and SRD; and obtaining a binary classification task result based on the binary classification function value and a preset threshold value. The scheme realizes the rapid and accurate identification of different types of DMEs.
Those skilled in the art will appreciate that the drawing is merely a schematic illustration of a preferred implementation scenario and that the modules or flows in the drawing are not necessarily required to practice the invention.
Those skilled in the art will appreciate that modules in an apparatus in an implementation scenario may be distributed in an apparatus in an implementation scenario according to an implementation scenario description, or that corresponding changes may be located in one or more apparatuses different from the implementation scenario. The modules of the implementation scenario may be combined into one module, or may be further split into a plurality of sub-modules.
The above-mentioned inventive sequence numbers are merely for description and do not represent advantages or disadvantages of the implementation scenario.
The foregoing disclosure is merely illustrative of some embodiments of the invention, and the invention is not limited thereto, as modifications may be made by those skilled in the art without departing from the scope of the invention.

Claims (8)

1. A method for processing DME typing based on a deep neural network, comprising:
preprocessing an OCT image to be identified;
performing image feature extraction on the OCT image subjected to pretreatment through a trained DME feature extraction model; the DME feature extraction model is obtained based on deep learning network training;
based on the extracted image features, obtaining a binary classification function value corresponding to whether preset DME appears in the OCT image; the preset DME includes: DRT, CME and SRD;
obtaining a DME typing result based on the binary classification function value and a preset threshold value;
the deep DME model is constructed on the basis of a VGG16 convolutional neural network and specifically comprises the following steps:
estimating the degree of difference between the VGG16 convolutional neural network output value and the true value by adopting a cross entropy loss function;
the ReLU activation function is adopted, so that the nonlinearity of the VGG16 convolutional neural network is improved, and the sparsity of the VGG16 convolutional neural network is increased;
the Dropout layer is adopted to act on a full connection layer of the VGG16 convolutional neural network;
after the Batchnorm layer is adopted to act on each convolution layer in the VGG16 convolution neural network, the distribution of the neuron input values of each layer of the neural network layer is pulled back to the standard normal distribution;
adjusting the VGG16 convolutional neural network which is pre-trained on the ImageNet data set, and constructing an initial model;
replacing the last full-connection layer, the SoftMax layer and the classification layer of the VGG16 convolutional neural network with a full-connection layer and a regression layer containing 3 neurons, so that the VGG16 convolutional neural network is converted into a regression network;
mapping the OCT image into OCT typing of DME based on the initial model, and respectively adjusting the output of the last fully-connected layer into binary classification function values of DRT, CME and SRD to obtain a deep DME model.
2. The method of claim 1, further comprising, prior to "preprocessing the OCT image to be identified:
acquiring OCT images of a DME-confirmed patient;
setting the OCT image of the patient as the OCT image to be identified.
3. The method of claim 1, wherein "preprocessing the OCT image to be identified" comprises:
removing saturated pixel values, reducing noise and cutting edges of the OCT image to be identified;
and reading the processed OCT images, and uniformly adjusting the read OCT images to a preset size.
4. The method of claim 1, wherein constructing the deepfdme model based on a VGG16 convolutional neural network, further comprises;
training the deep DME model based on a training set in the preprocessed preset OCT image dataset, and performing internal verification;
and carrying out external verification on the trained deep DME model based on a test set in a preset OCT image data set so as to obtain a DME feature extraction model.
5. The method of claim 4, wherein the training the deepdem model based on the training set in the pre-processed pre-set OCT image dataset and performing internal verification comprises:
based on a training set in the preprocessed preset OCT image data set, performing internal inspection on the deep DME model by using a five-fold cross validation method; the training set is randomly divided into five parts, wherein 4 parts are used for training the deep DME model, the remaining 1 part is used for evaluating the performance of the deep DME model, and each part participates in the training of the deep DME model;
determining consistency between a predicted label and a real label obtained each time in internal verification of the deep DME model by using a confusion matrix, and calculating accuracy, sensitivity, specificity and area under a test subject working characteristic curve of each time of running of the model;
the performance of the deepdem model was verified by recording the average results of five runs of the deepdem model.
6. The method of claim 4, wherein the externally validating the deep DME model after training based on a test set in a preset OCT image dataset to obtain a DME feature extraction model comprises:
externally verifying the trained deep DME model by using a test set in a preset OCT image data set;
determining consistency between a predicted tag and a real tag obtained by the deep DME model by using a confusion matrix;
based on the confusion matrix, calculating the accuracy, sensitivity, specificity and area under the working characteristic curve of the subject for detecting the DME typing by the deep DME model.
7. The method of claim 1, wherein the obtaining a DME-typed result based on the binary classification function value and a preset threshold value comprises:
for OCT images of the hybrid DME, binary classification function values of the DRT, the CME and the SRD and a preset threshold value are converted into the proportion of 3 types of DME in the hybrid DME by a normalization method.
8. A terminal comprising a memory and a processor, the processor executing the method of any of claims 1-7 when running a program stored in the memory.
CN202010851829.8A 2020-08-21 2020-08-21 Method and terminal for processing DME (DME) typing based on deep neural network Active CN112132782B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010851829.8A CN112132782B (en) 2020-08-21 2020-08-21 Method and terminal for processing DME (DME) typing based on deep neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010851829.8A CN112132782B (en) 2020-08-21 2020-08-21 Method and terminal for processing DME (DME) typing based on deep neural network

Publications (2)

Publication Number Publication Date
CN112132782A CN112132782A (en) 2020-12-25
CN112132782B true CN112132782B (en) 2023-09-05

Family

ID=73851033

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010851829.8A Active CN112132782B (en) 2020-08-21 2020-08-21 Method and terminal for processing DME (DME) typing based on deep neural network

Country Status (1)

Country Link
CN (1) CN112132782B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115083005B (en) * 2022-06-13 2023-07-04 广东省人民医院 ROP image classification system and method based on deep learning

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3065086A1 (en) * 2015-03-02 2016-09-07 Medizinische Universität Wien Computerized device and method for processing image data
CN109376767A (en) * 2018-09-20 2019-02-22 中国科学技术大学 Retina OCT image classification method based on deep learning
CN109726743A (en) * 2018-12-12 2019-05-07 苏州大学 A kind of retina OCT image classification method based on Three dimensional convolution neural network
CN110110600A (en) * 2019-04-04 2019-08-09 平安科技(深圳)有限公司 The recognition methods of eye OCT image lesion, device and storage medium
CN110309862A (en) * 2019-06-11 2019-10-08 广东省人民医院(广东省医学科学院) DME prognosis information forecasting system and its application method based on ensemble machine learning
CN111198959A (en) * 2019-12-30 2020-05-26 郑州轻工业大学 Two-stage image retrieval method based on convolutional neural network
CN111476283A (en) * 2020-03-31 2020-07-31 上海海事大学 Glaucoma fundus image identification method based on transfer learning

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10660576B2 (en) * 2017-01-30 2020-05-26 Cognizant Technology Solutions India Pvt. Ltd. System and method for detecting retinopathy

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3065086A1 (en) * 2015-03-02 2016-09-07 Medizinische Universität Wien Computerized device and method for processing image data
CN109376767A (en) * 2018-09-20 2019-02-22 中国科学技术大学 Retina OCT image classification method based on deep learning
CN109726743A (en) * 2018-12-12 2019-05-07 苏州大学 A kind of retina OCT image classification method based on Three dimensional convolution neural network
CN110110600A (en) * 2019-04-04 2019-08-09 平安科技(深圳)有限公司 The recognition methods of eye OCT image lesion, device and storage medium
CN110309862A (en) * 2019-06-11 2019-10-08 广东省人民医院(广东省医学科学院) DME prognosis information forecasting system and its application method based on ensemble machine learning
CN111198959A (en) * 2019-12-30 2020-05-26 郑州轻工业大学 Two-stage image retrieval method based on convolutional neural network
CN111476283A (en) * 2020-03-31 2020-07-31 上海海事大学 Glaucoma fundus image identification method based on transfer learning

Also Published As

Publication number Publication date
CN112132782A (en) 2020-12-25

Similar Documents

Publication Publication Date Title
Gargeya et al. Automated identification of diabetic retinopathy using deep learning
Sarki et al. Convolutional neural network for multi-class classification of diabetic eye disease
CN112133441B (en) Method and terminal for establishing MH postoperative crack state prediction model
Ardiyanto et al. Deep learning-based diabetic retinopathy assessment on embedded system
Akter et al. Glaucoma diagnosis using multi-feature analysis and a deep learning technique
Jayakumari et al. Automated diabetic retinopathy detection and classification using imagenet convolution neural network using fundus images
Al-Timemy et al. A hybrid deep learning construct for detecting keratoconus from corneal maps
Rahim et al. Detection of diabetic retinopathy and maculopathy in eye fundus images using deep learning and image augmentation
Nasir et al. Deep DR: detection of diabetic retinopathy using a convolutional neural network
Sharma et al. Machine learning approach for detection of diabetic retinopathy with improved pre-processing
CN112132782B (en) Method and terminal for processing DME (DME) typing based on deep neural network
CN110403611B (en) Method and apparatus for predicting glycated hemoglobin component value in blood, computer device, and storage medium
Asirvatham et al. Hybrid deep learning network to classify eye diseases
Varma et al. A reliable automatic cataract detection using deep learning
Rakhmetulayeva et al. IMPLEMENTATION OF CONVOLUTIONAL NEURAL NETWORK FOR PREDICTING GLAUCOMA FROM FUNDUS IMAGES.
Romero-Aroca et al. Validation of a deep learning algorithm for diabetic retinopathy
Latha et al. Automated macular disease detection using retinal optical coherence tomography images by fusion of deep learning networks
Quellec et al. Instant automatic diagnosis of diabetic retinopathy
Rathi et al. A review paper on prediction of diabetic retinopathy using data mining techniques
Mazar Pasha et al. Diabetic Retinopathy Severity Categorization in Retinal Images Using Convolution Neural Network.
Fayemiwo et al. Identification and Classification of Diabetic Retinopathy Using Machine Learning
Mani et al. An automated hybrid decoupled convolutional network for laceration segmentation and grading of retinal diseases using optical coherence tomography (OCT) images
Bhulakshmi et al. A systematic review on diabetic retinopathy detection and classification based on deep learning techniques using fundus images
Sharma et al. Optimised Machine Learning Algorithm for Detection And Diagnosis of Diabetes
Venneti et al. Amdnet: Age-related macular degeneration diagnosis through retinal fundus images using lightweight convolutional neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant