CN109390053B - Fundus image processing method, fundus image processing apparatus, computer device, and storage medium - Google Patents

Fundus image processing method, fundus image processing apparatus, computer device, and storage medium Download PDF

Info

Publication number
CN109390053B
CN109390053B CN201810340025.4A CN201810340025A CN109390053B CN 109390053 B CN109390053 B CN 109390053B CN 201810340025 A CN201810340025 A CN 201810340025A CN 109390053 B CN109390053 B CN 109390053B
Authority
CN
China
Prior art keywords
fundus image
neural network
feature set
lesion
fundus
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810340025.4A
Other languages
Chinese (zh)
Other versions
CN109390053A (en
Inventor
贾伟平
盛斌
李华婷
戴领
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Jiaotong University
Shanghai Sixth Peoples Hospital
Original Assignee
Shanghai Jiaotong University
Shanghai Sixth Peoples Hospital
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Jiaotong University, Shanghai Sixth Peoples Hospital filed Critical Shanghai Jiaotong University
Priority to PCT/CN2018/086739 priority Critical patent/WO2019024568A1/en
Priority to US16/302,410 priority patent/US11200665B2/en
Publication of CN109390053A publication Critical patent/CN109390053A/en
Application granted granted Critical
Publication of CN109390053B publication Critical patent/CN109390053B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B3/00Apparatus for testing the eyes; Instruments for examining the eyes
    • A61B3/10Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions
    • A61B3/12Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for looking at the eye fundus, e.g. ophthalmoscopes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/2163Partitioning the feature space
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • G16H30/20ICT specially adapted for the handling or processing of medical images for handling medical images, e.g. DICOM, HL7 or PACS
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • G16H30/40ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/20ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/70ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for mining of medical data, e.g. analysing previous cases of other patients
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • G06N20/10Machine learning using kernel methods, e.g. support vector machines [SVM]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20112Image segmentation details
    • G06T2207/20132Image cropping
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30041Eye; Retina; Ophthalmic
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30096Tumor; Lesion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/03Recognition of patterns in medical or anatomical images

Abstract

The application relates to a fundus image processing method, a fundus image processing device, a computer device and a storage medium. The method comprises the following steps: receiving an acquired fundus image; identifying a fundus image through a first neural network to generate a first feature set of the fundus image; identifying the fundus image through a second neural network, and generating a second feature set of the fundus image, wherein the first feature set and the second feature set represent different lesion attributes of the fundus image; combining the first feature set and the second feature set to obtain a combined feature set of the fundus image; and inputting the combined feature set into a classifier to obtain a classification result. By adopting the method, the accuracy of classifying the fundus images can be improved.

Description

Fundus image processing method, fundus image processing apparatus, computer device, and storage medium
Technical Field
The present application relates to the field of artificial intelligence technologies, and in particular, to a method and an apparatus for processing an eye fundus image, a computer device, and a storage medium.
Background
In recent years, artificial intelligence has been remarkably developed in various fields. An important branch of artificial intelligence is that the human brain is simulated by machine learning for analytical learning, so as to achieve the purpose of interpreting data (such as images, sounds and texts).
At present, regarding the identification of the fundus images, the main identification method is to use the experience of doctors to diagnose whether patients have fundus diseases and the severity of the fundus diseases by means of visual observation, and the manual identification method is time-consuming, labor-consuming and inefficient. The identification of the eye diseases through the machine learning mode is limited to the construction of a single machine learning model, and the identification accuracy is low.
Disclosure of Invention
In view of the above, it is necessary to provide a fundus image recognition method, apparatus, computer device, and storage medium capable of improving the accuracy of classifying fundus images in view of the above technical problems.
A method of fundus image processing, the method comprising:
receiving an acquired fundus image;
identifying the fundus image through a first neural network to generate a first feature set of the fundus image;
identifying the fundus image through a second neural network, and generating a second feature set of the fundus image, wherein the first feature set and the second feature set represent different lesion attributes of the fundus image;
combining the first feature set and the second feature set to obtain a combined feature set of the fundus image;
and inputting the combined feature set into a classifier to obtain a classification result.
In one embodiment, the first set of features characterizes a lesion type attribute of the fundus image, and the second set of features characterizes a lesion level attribute of the fundus image;
inputting the combined feature set into a classifier to obtain a classification result as follows:
and inputting the combined feature set with the lesion type attribute and the lesion level attribute into a multi-stage classifier which is composed of a plurality of two-stage classifiers according to set classification logic to obtain a multi-stage classification result of the fundus image.
In one embodiment, the identifying the fundus image by the first neural network, the obtaining the first feature set of the fundus image comprises:
quadrant segmentation is carried out on the fundus image to generate a quadrant image group;
inputting each quadrant image in the quadrant image group into a first neural network to obtain a feature vector of each quadrant image;
combining the feature vectors of each quadrant image generates a first feature set of the fundus image.
In one embodiment, the received fundus images include a left eye fundus image and a right eye fundus image from the same patient;
inputting the combined feature set into a classifier, and obtaining a classification result comprises:
connecting the combined feature set of the left eye fundus image and the combined feature set of the right eye fundus image to generate a combined feature sequence of the fundus images;
and inputting the combined characteristic sequence into a classifier to obtain a classification result.
In one embodiment, the received fundus images include a first view left eye fundus image, a second view left eye fundus image, a first view right eye fundus image, and a second view right eye fundus image from the same patient;
inputting the combined feature set into a classifier, and obtaining a classification result comprises:
connecting the combined feature set of the first-view left eye fundus image, the combined feature set of the second-view left eye fundus image, the combined feature set of the first-view right eye fundus image and the combined feature set of the second-view right eye fundus image to generate a combined feature sequence of the fundus images;
and inputting the combined characteristic sequence into a classifier to obtain a classification result.
In one embodiment, the identifying the fundus image by the second neural network, generating a second set of features for the fundus image, comprises:
identifying the lesion grade attribute of the fundus image through a second neural network, and outputting a lesion grade vector of the fundus image, wherein when the set fundus lesion containsnGrade of lesion, grade of lesionThe length of the vector isn-1, wherein,imedian anterior in feature vector of grade lesioniIs 1, the rest is 0.
A fundus image processing apparatus, the apparatus comprising:
the image acquisition module is used for receiving the acquired fundus images;
the first neural network identification module is used for identifying the fundus image through a first neural network to generate a first feature set of the fundus image;
the second neural network identification module is used for identifying the fundus image through a second neural network and generating a second feature set of the fundus image, wherein the first feature set and the second feature set represent different lesion attributes of the fundus image;
the characteristic combination module is used for combining the first characteristic set and the second characteristic set to obtain a combined characteristic set of the fundus image;
and the classification module is used for inputting the combined feature set into a classifier to obtain a classification result.
In one embodiment, the first set of features characterizes a lesion type attribute of the fundus image, and the second set of features characterizes a lesion level attribute of the fundus image;
the classification module is further used for inputting the combined feature set with the lesion type attribute and the lesion grade attribute into a multi-stage classifier which is composed of a plurality of two-type classifiers according to set classification logics, and obtaining a multi-stage classification result of the fundus image.
A computer device comprising a memory and a processor, the memory storing a computer program, the processor implementing the following steps when executing the computer program:
receiving an acquired fundus image;
identifying the fundus image through a first neural network to generate a first feature set of the fundus image;
identifying the fundus image through a second neural network, and generating a second feature set of the fundus image, wherein the first feature set and the second feature set represent different lesion attributes of the fundus image;
combining the first feature set and the second feature set to obtain a combined feature set of the fundus image;
and inputting the combined feature set into a classifier to obtain a classification result.
A computer-readable storage medium, on which a computer program is stored which, when being executed by a processor, carries out the following steps.
Receiving an acquired fundus image;
identifying the fundus image through a first neural network to generate a first feature set of the fundus image;
identifying the fundus image through a second neural network, and generating a second feature set of the fundus image, wherein the first feature set and the second feature set represent different lesion attributes of the fundus image;
combining the first feature set and the second feature set to obtain a combined feature set of the fundus image;
and inputting the combined feature set into a classifier to obtain a classification result.
By training two different neural networks, namely the first neural network and the second neural network, the two neural networks can abstract lesion features representing different attributes from the fundus image, namely extracting the lesion features of the fundus image from different angles. The fundus image feature at this stage has essentially preliminarily identified fundus lesions. On the basis, the abstracted pathological change characteristics with different attributes are combined to obtain a combined characteristic set of the fundus image, the combined characteristic set containing more characteristics is used as a characteristic value of the fundus image and is input into a classifier to be identified and classified again, and the classification result is more accurate after multiple pathological change characteristics are combined and are identified through a plurality of neural networks.
Drawings
FIG. 1 is a diagram showing an environment in which a fundus image processing method is applied in one embodiment;
FIG. 2 is a diagram showing an environment in which a fundus image processing method according to another embodiment is applied;
FIG. 3 is a flowchart illustrating a fundus image processing method according to an embodiment;
FIG. 4 is a schematic view of an acquired fundus image;
fig. 5 is a flowchart schematically illustrating a fundus image processing method in another embodiment;
fig. 6 is a schematic view of a fundus image after quadrant cutting;
fig. 7 is a block diagram showing the configuration of a fundus image processing apparatus in one embodiment;
FIG. 8 is a diagram illustrating an internal structure of a computer device according to an embodiment.
Detailed Description
In order to make the objects, technical solutions and advantages of the present application more apparent, the present application is described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the present application and are not intended to limit the present application.
The fundus image processing method provided by the present application can be applied to the application environment as shown in fig. 1. The application environment includes an image capture device 110a, a server 120a, and a terminal 130a, and the image capture device 110a and the terminal 130a may communicate with the server 120a through a network. The server 120a may be an independent server or a server cluster composed of a plurality of servers, and the terminal 130a may be, but is not limited to, various personal computers, notebook computers, smart phones, tablet computers, and portable wearable devices. The image acquisition device 110a can acquire fundus images, the server 120a stores a first neural network, a second neural network and a classifier which are trained in advance, and the server identifies the fundus images through the neural networks to obtain lesion classification results contained in the fundus images. The terminal 130a receives and displays the classification result generated by the server 120 a.
In another embodiment, the fundus image processing method provided by the present application can also be applied to an application environment as shown in fig. 2, the application environment including an image capturing device 110b and a terminal 120b, and the image capturing device 110b can communicate with the terminal 120b through a network. The image acquisition device 110b can acquire fundus images, the terminal 120b stores a first neural network, a second neural network and a classifier which are trained in advance, and the server identifies the fundus images through the neural networks to obtain and display lesion classification results contained in the fundus images.
As shown in fig. 3, the present application provides a fundus image processing method, including the steps of:
step S210: an acquired fundus image is received.
The acquisition of the fundus image may be generated by a hand-held/stationary medical imaging device acquisition, the acquired fundus image being as shown in fig. 4. The fundus image acquired by the medical imaging equipment comprises an effective fundus image in a middle circular area and a surrounding white or black area which is a camera shading part and has no diagnostic significance. Before model prediction, the fundus image may be preprocessed, for example, to remove pixels that are not diagnostic.
Step S220: a fundus image is identified by a first neural network, generating a first set of features for the fundus image.
Step S230: identifying the fundus image via a second neural network, generating a second set of features for the fundus image, wherein the first set of features and the second set of features characterize different lesion attributes of the fundus image.
The first neural network and the second neural network are both constructed by training historical fundus images. The neural network training process is a process of learning and training a certain set fundus lesion attribute of a sample.
In this embodiment, the first neural network is trained to be able to recognize the set lesion attribute of the fundus image. The acquired fundus images are input into a first neural network for identification, and the set lesion genus diseases of the fundus images identified by the first neural network are represented by a first feature set. Likewise, the second neural network recognizes that the lesion property of the fundus image is represented by a second feature set.
In this embodiment, it is understood that the first feature set and the second feature set are both used to describe lesion properties of the acquired fundus image, but the lesion properties of the fundus image identified by the first neural network and the second neural network are not the same, and are complementary to each other.
The above feature set may be a "feature vector" or a "feature sequence", the meaning of which should be understood in the broadest way.
Step S240: the first feature set and the second feature set are combined to obtain a combined feature set of the fundus image.
And fusing the first feature set produced by the first neural network and the second feature set produced by the second neural network to generate a combined feature set. The "combined feature set" herein may be a "feature sequence", "feature vector", or the like. In one embodiment, the combination of the first feature set and the second feature set is a vector sum of features.
Step S250: and inputting the combined feature set into a classifier to obtain a classification result.
The classifier is used as a classifier for finally judging the classification result of the fundus image.
In this embodiment, by training two different neural networks, namely the first neural network and the second neural network, the two neural networks can abstract features representing different lesion attributes from the fundus image, that is, the lesion features are extracted from the fundus image from different angles. The fundus image characteristics at this stage have been able to substantially reflect the fundus image lesion classification. On the basis, the abstracted features with different lesion attributes are combined to obtain a combined feature set of the fundus image, the combined feature set containing more features is used as a feature value of the fundus image and is input into a classifier to be classified and identified again, and the classification result is more accurate by combining multiple lesion features and identifying through a plurality of neural networks.
In one embodiment, the classifier in step S150 may be a class two classification model. The fundus images are classified into two stages, namely pathological changes and non-pathological changes, or mild pathological changes and severe pathological changes. Specifically, the class two classification model may linearly divide the samples into two classes. Taking an SVM as an example, the basic model is defined as a linear classifier with larger intervals on the feature space, and the learning strategy is larger in interval and can be finally converted into the solution of a convex quadratic programming problem. The purpose of the SVM is: finding a hyperplane allows the samples to be divided into two classes with the largest separation. And w, which we find, represents the coefficients of the hyperplane we need to find. Namely:
Figure 757026DEST_PATH_IMAGE001
when the original sample space may not have a hyperplane that can correctly divide the two types of samples, the samples can be mapped from the original space to a higher-dimensional feature space, so that the samples can be linearly divided into two types in the new high-dimensional space, i.e. the samples are linearly divided in the space. Still further, the selection of kernel functions becomes the maximum variable of the support vector machine (if kernel functions have to be used, i.e. coring), so what kernel functions to choose will affect the final result. The most commonly used kernel functions are: linear kernels, polynomial kernels, gaussian kernels, laplacian kernels, sigmoid kernels, new kernel functions derived by operations such as linear combinations or direct products between kernel functions, and the like.
In another embodiment, the classifier may be a multi-level classification network composed of a plurality of two-class classification models according to a set classification logic. If the fundus images are classified in multiple stages, if the fundus images are classified into 5 types, the fundus images are respectively marked as 0-4 stages without pathological changes, mild pathological changes, moderate pathological changes, severe pathological changes, PDR and pathological changes above the degree.
The set classification logic can be multi-label classification logic of 1-VS-ALL, and each sub-class two classification model contained in the multi-level classification model can be separated from other classes to be assigned to a certain class sample. If the classifier is a 5-level classification network, it contains a 5-SVM two-classification network, i.e., one SVM is trained for each classification. Respectively, 0| 1234-is graded into 0 type sample, 1| 0234-is graded into 1 type sample, 2| 0134-is graded into 2 type sample, 3| 0124-is graded into 3 type sample, and 4| 0123-is graded into 4 type sample.
When the SVM is trained, the combined feature set obtained after the processing of the first neural network and the second neural network is used as a feature vector of the fundus image to train an SVM classifier. When the SVM is trained, if the positive and negative samples are not uniformly distributed, different weights are given to the positive sample and the negative sample, for the SVM 0|1234, the positive sample is a type 0 sample (a sample without a pathological change), and the negative sample is a sample with a pathological change. If the ratio of the current positive sample number to the total sample number is d, the weight assigned to the current positive sample number is 1/(2 d). The sample weight is set to alleviate the uneven distribution of data, which is equivalent to increasing the number of the samples with less data, so that the loss value of the samples is equivalent to that of most samples.
In one embodiment, the first Neural network and the second Neural network are Convolutional Neural Networks (CNNs). The convolutional neural network is one of artificial neural networks, and has a weight sharing network structure which is more similar to a biological neural network, so that the complexity of a network model is reduced, and the number of weights is reduced. The acquired fundus images can be directly used as the input of the network, and the complex characteristic extraction and data reconstruction process in the traditional recognition algorithm is avoided.
Further, the first neural network is a convolutional neural network capable of identifying a type of lesion contained in the fundus image. The second neural network is a convolutional neural network capable of identifying a lesion level of a fundus lesion included in the fundus image. That is, the first set of features characterizes a lesion type attribute of the fundus image, and the second set of features characterizes a lesion level attribute of the fundus image. Combining the characteristics of lesion types contained in the fundus image obtained through CNN prediction and the characteristics of lesion classification of the fundus image, wherein the combined characteristic vector contains lesion characteristics of multiple dimensions of the fundus image, and inputting the combined characteristic vector into the SVM to obtain more accurate and stable fundus lesion classification.
Further, the type of fundus lesion identified by the first neural network may include: microangiomas, hard extravasation, soft extravasation, and hemorrhage. Based on this, the first feature set output by the first neural network may be a feature vector with a length of 4, and the first neural network is trained such that each element of the output feature vector represents a corresponding lesion type in turn. For example, if the feature vector output by the first neural network is [1,0,0,0], it indicates that the fundus image includes microangiomas, and does not include hard effusion, soft effusion, and hemorrhage.
In one embodiment, a lesion level attribute of the fundus image is identified via a second neural network, and a lesion level vector of the fundus image is output, wherein when the fundus lesion inclusion is setnIn case of grade lesion, the length of the generated lesion grade vector isn-1, wherein,imedian anterior in feature vector of grade lesioniIs 1, the rest is 0 configuration. For example, the fundus lesion level that the second neural network can identify may include: no pathological changes, mild pathological changes, moderate pathological changes, severe pathological changes, PDR and the above pathological changes are respectively marked as 0-4 grades. Based on this, the second set of features output by the second neural network may be a length-6 feature vector. Unlike One-hot encoding methods used in general multi-level classification, the present application uses a progressive encoding method. That is, for class 0, the training target for the corresponding second neural network is the vector [0,0,0, 0]]For class 1 [1,0,0,0,]for class 2 [1,1,0]. Namely foriClass, target vector middle frontiThe bits are 1 and the remainder are 0. That is, when the fundus image lesion includesnIn stage lesions, the second set of features generated by the second neural network should be of lengthn-1, wherein,imedian anterior in feature vector of grade lesioniIs 1, the rest is 0.
The fundus lesion grading label coding mode of the second neural network is in accordance with the phenomena that lesions are continuously deepened and new lesion types appear under the condition that old lesion types exist.
The training process of the first convolutional neural network, the second convolutional neural network and the classifier is explained as follows.
The training for the first neural network is: and preprocessing the fundus image in advance to obtain a training sample. And carrying out artificial marking on the types of the lesions on the training samples, marking the types of the lesions contained in each sample, wherein each type of the lesions represents a label, and obtaining target output corresponding to the training samples based on the codes of the labels. If the sample image contains microangiomas and hard extravasations, the target output for the sample should be [1,1,0,0 ]. And inputting the processed picture into a CNN network in the training process, carrying out forward propagation, then calculating the difference between the CNN network output and the target output, carrying out derivation on each part in the network, and updating the network parameters by using an SGD algorithm.
The above-described preprocessing of the fundus image includes:
1. an information Area of the image, i.e. an Area of Interest (AOI), is acquired. The AOI of the fundus image, i.e., the circular area in the middle of the fundus picture, contains only the effective fundus image, and the surrounding white or black portion is the camera-obstructing portion, and has no diagnostic significance.
2. And (5) zooming the picture. Fundus images have a high resolution, usually higher than 1000 × 2000, and cannot be directly input as CNN, so the size of the image can be reduced to a desired size, which may be 299 × 299.
3. And normalizing the single picture. The method is mainly used for avoiding the influence of image judgment caused by illumination and the like. The step calculates the average value and standard deviation of pixel intensity in AOI for each channel in the picture RGB channels. For each pixel, the intensity value is subtracted from the average value and then divided by the standard deviation to obtain the intensity value after normalization.
4. Random noise is added. In order to reduce the overfitting problem in the training process and carry out multiple sampling in the prediction process, Gaussian noise with the average value of 0 and the standard deviation of 5% of the image labeling difference is added into the image obtained in the last step. The method can not influence the image discrimination, and can reduce the insufficient bloom problem caused by the overfitting problem.
5. And (4) randomly rotating. The picture AOI part is circular, so that the picture can be rotated by any angle by taking the picture center as the center of a circle. The image rotation does not bring any influence to the picture diagnosis, and the influence of the overfitting problem can be reduced.
Similarly, the pre-processing of the fundus image is also required before the training of the second neural network and the classifier, and therefore, the pre-processing of the image is not described in detail when the training of the second neural network and the classifier is stated.
The training of the second neural network is: and preprocessing the fundus image in advance to obtain a training sample. And manually marking the training samples, marking the lesion grade corresponding to each sample, and obtaining target output corresponding to the training samples based on the progressive coding mode. If the fundus image in a sample is level 3, the target output of that sample should be (1, 1,1, 0). And inputting the processed picture into a CNN network in the training process, carrying out forward propagation, then calculating the difference between the CNN network output and the target output, carrying out derivation on each part in the network, and updating the network parameters by using an SGD algorithm.
In one embodiment, as shown in fig. 5, there is provided a fundus image processing method, including the steps of:
step S310: an acquired fundus image is received.
Step S320: quadrant segmentation is carried out on the fundus image to generate a quadrant image group, each quadrant image in the quadrant image group is input into a first neural network to obtain a feature vector of each quadrant image, and the feature vectors of each quadrant image are combined to generate a first feature set of the fundus image.
Quadrant division is to divide the fundus image into four regions using the horizontal axis and the vertical axis in a cartesian coordinate system, as shown in fig. 6. The fundus image in the area is a quadrant image. The quadrant image is scaled to a set size, such as 299. After processing, the four quadrant images form a quadrant image group.
And inputting quadrant images in the quadrant image group into a first neural network for prediction, wherein each quadrant image generates a feature vector. The first neural network may be a convolutional neural network that identifies a lesion type of the image, and the feature vector of the output quadrant image of the first neural network may be a length-4 vector, where each element in the vector corresponds to a lesion type, e.g., [1,0,0,0 ]. The first neural network and the specific definition of the output of the first neural network refer to the above definitions, and are not described herein again.
Before the quadrant image is input into the first neural network for prediction, preprocessing needs to be performed on the quadrant image, where the preprocessing may include a singulation process, adding random noise, and random rotation.
The feature vector of the combined quadrant image may be a long vector of length 16 connecting the feature vectors of the four quadrant images. May be the feature vector of the first quadrant image + the feature vector of the second quadrant image + the feature vector of the third quadrant image + the feature vector of the fourth quadrant image. The first feature vector generated by combining the feature vectors of the quadrant images can not only represent the types of the lesions contained in the images, but also represent the distribution of different types of lesions.
Step S330: identifying the fundus image via a second neural network, generating a second set of features for the fundus image, wherein the first set of features and the second set of features characterize different lesion attributes of the fundus image.
The specific limitations of this step can refer to the above limitations, which are not described herein again.
Step S340: the first feature set and the second feature set are combined to obtain a combined feature set of the fundus image.
The combined feature set herein includes a first attribute feature of the four quadrant images and a second attribute feature of the fundus image.
Step S350: and inputting the combined feature set into a classifier to obtain a classification result.
The specific limitations of this step can refer to the above limitations, which are not described herein again.
In this embodiment, the combined feature set including more lesion features is input to the classifier, so that the obtained classification result is more accurate.
It should be understood that although the steps in the flowcharts of fig. 3 and 5 are shown in order as indicated by the arrows, the steps are not necessarily performed in order as indicated by the arrows. The steps are not performed in the exact order shown and described, and may be performed in other orders, unless explicitly stated otherwise. Moreover, at least some of the steps in fig. 3 and 5 may include multiple sub-steps or multiple stages that are not necessarily performed at the same time, but may be performed at different times, and the order of performing the sub-steps or stages is not necessarily sequential, but may be performed alternately or alternately with other steps or at least some of the sub-steps or stages of other steps.
In one embodiment, the acquired fundus images may be pairs of fundus images, including a left eye image and a right eye image from the same patient.
And respectively executing the steps S120-S14 or S220-S240 on the left-eye image and the left-eye image to obtain a combined feature set of the left-eye image and a combined feature set of the right-eye image, connecting the combined feature set of the left-eye image and the combined feature set of the right-eye image to generate a combined feature sequence, and inputting the combined feature sequence into a classifier to obtain a classification result.
The classifier in this embodiment is obtained by training a combined feature set of both eyes obtained after processing by the first neural network and the second neural network as a feature vector of the fundus image. That is, the training of the classifier in this embodiment requires to input a feature vector with a length of two eyes (2 times the length of a feature vector of a single eye), and during prediction, a feature vector with a corresponding length needs to be input for prediction.
The combined feature sequence in this embodiment includes lesion features of two different attributes of the left eye fundus image and lesion features of two different attributes of the right eye fundus image, that is, the binocular images (lesions of both eyes have strong correlation) are fused, and the multiple CNN networks and quadrant lesion features are fused, so that the accuracy of lesion classification is further improved.
In one embodiment, the acquired fundus images are two sets of pairs of fundus images in different fields of view, including a left eye image and a right eye image from a first field of view and a left eye image and a right eye image from a second field of view.
And respectively executing the steps S120-S14 or S220-S240 on the binocular double-view images to obtain four groups of combined feature sets, connecting the combined feature sets to generate a combined feature sequence, and inputting the combined feature sequence into a classifier to obtain a classification result.
The classifier in this embodiment is obtained by training a combined feature set of both eyes and both visual fields obtained after processing by the first neural network and the second neural network as a feature vector of the fundus image. That is, the training of the classifier in this embodiment requires to input the feature vectors with the lengths of both eyes and both fields (4 times the length of the feature vector of a single eye), and during prediction, the feature vectors with corresponding lengths need to be input for prediction.
If monocular or single-view data exists in the training data or the data to be predicted, the feature value corresponding to the unavailable/nonexistent view is set to be the same as the existing view, and the feature value corresponding to the unavailable/nonexistent view is set to be the same as the existing value of a certain monocular to generate the feature vector with the corresponding length.
The combined feature sequence in this embodiment includes lesion features of two different attributes of left eye fundus images in different visual fields and lesion features of two different attributes of right eye fundus images in different visual fields, that is, a dual-visual-field binocular image is fused, and a plurality of CNN networks and quadrant lesion features are fused, so that the accuracy of lesion classification is further improved.
In one embodiment, as shown in fig. 7, there is provided a fundus image processing apparatus including:
an image acquisition module 410 for receiving acquired fundus images.
A first neural network identification module 420 for identifying the fundus image via a first neural network to generate a first set of features for the fundus image.
A second neural network identification module 430 for identifying the fundus image via the second neural network to generate a second set of features for the fundus image, wherein the first set of features and the second set of features characterize different lesion attributes of the fundus image.
And the characteristic combination module 440 is used for combining the first characteristic set and the second characteristic set to obtain a combined characteristic set of the fundus image.
The classification module 450 is configured to input the combined feature set into a classifier to obtain a classification result.
In one embodiment, the first neural network is a convolutional neural network capable of identifying the type of lesion contained in the fundus image, the second neural network is a convolutional neural network capable of identifying the grade of the fundus lesion, and the classifier is a multi-stage classification network composed of a plurality of two-class classifiers according to a set classification logic.
In one embodiment, the first neural network identification module 420 is further configured to perform quadrant segmentation on the fundus image to generate a quadrant image group; inputting each quadrant image in the quadrant image group into a first neural network to obtain a feature vector of each image; combining the feature vectors for each quadrant image generates a first feature set for the fundus image.
In one embodiment, the received fundus images include a left eye fundus image and a right eye fundus image from the same patient. The classification module 450 is further configured to connect the combined feature set of the left eye fundus image and the combined feature set of the right eye fundus image to generate a combined feature sequence of the fundus images; and inputting the combined characteristic sequence into a classifier to obtain a classification result.
In one embodiment, the received fundus images include a first view left eye fundus image, a second view left eye fundus image, a first view right eye fundus image, and a second view right eye fundus image from the same patient; the classification module 450 is further configured to connect the combined feature set of the first-view left eye fundus image, the combined feature set of the second-view left eye fundus image, the combined feature set of the first-view right eye fundus image, and the combined feature set of the second-view right eye fundus image, so as to generate a combined feature sequence of the eye fundus images; and inputting the combined characteristic sequence into a classifier to obtain a classification result.
In one embodiment, the second neural network is a convolutional neural network capable of identifying the level of fundus lesions when the fundus image comprises lesionsnSecond generation in case of grade lesionThe feature set is of lengthn-a feature vector of 1, and (c),imedian anterior in feature vector of grade lesioniIs 1, the rest is 0.
The respective modules in the fundus image processing apparatus described above may be entirely or partially realized by software, hardware, and a combination thereof. The network interface may be an ethernet card or a wireless network card. The modules can be embedded in a hardware form or independent from a processor in the computer device, and can also be stored in a memory in the computer device in a software form, so that the processor can call and execute operations corresponding to the modules. The processor can be a Central Processing Unit (CPU), a microprocessor, a singlechip and the like.
In one embodiment, a computer device is provided, which may be a server or a terminal, and its internal structure diagram may be as shown in fig. 8. The computer device includes a processor, a memory, a network interface, and a database connected by a system bus. Wherein the processor of the computer device is configured to provide computing and control capabilities. The memory of the computer device comprises a nonvolatile storage medium and an internal memory. The non-volatile storage medium stores an operating system, a computer program, and a database. The internal memory provides an environment for the operation of an operating system and computer programs in the non-volatile storage medium. The database of the computer device is used for storing neural network model data. The network interface of the computer equipment is used for connecting and communicating with an external image acquisition terminal through a network. The computer program is executed by a processor to implement a fundus image processing method.
Those skilled in the art will appreciate that the architecture shown in fig. 8 is merely a block diagram of some of the structures associated with the disclosed aspects and is not intended to limit the computing devices to which the disclosed aspects apply, as particular computing devices may include more or less components than those shown, or may combine certain components, or have a different arrangement of components.
In one embodiment, a computer device is provided, comprising a memory and a processor, the memory having a computer program stored therein, the processor implementing the following steps when executing the computer program:
receiving an acquired fundus image;
identifying a fundus image through a first neural network to generate a first feature set of the fundus image;
identifying the fundus image through a second neural network, and generating a second feature set of the fundus image, wherein the first feature set and the second feature set represent different lesion attributes of the fundus image;
combining the first feature set and the second feature set to obtain a combined feature set of the fundus image;
and inputting the combined feature set into a classifier to obtain a classification result.
In one embodiment, the first neural network is a convolutional neural network capable of identifying the type of lesion contained in the fundus image, the second neural network is a convolutional neural network capable of identifying the grade of the fundus lesion, and the classifier is a multi-stage classification network composed of a plurality of two-class classifiers according to a set classification logic.
In one embodiment, the processor, when executing the identifying the fundus image by the first neural network, resulting in the first feature set of the fundus image, further performs the steps of:
quadrant segmentation is carried out on the fundus image to generate a quadrant image group;
inputting each quadrant image in the quadrant image group into a first neural network to obtain a characteristic vector corresponding to each quadrant image;
combining the feature vectors for each quadrant image generates a first feature set for the fundus image.
In one embodiment, the acquired fundus images include a left eye fundus image and a right eye fundus image from the same patient;
the processor inputs the combined feature set into the classifier, and when a classification result is obtained, the following steps are also realized: connecting the combined characteristic set of the left eye fundus image and the combined characteristic set of the right eye fundus image to generate a combined characteristic sequence of the fundus images; and inputting the combined characteristic sequence into a classifier to obtain a classification result.
In one embodiment, the acquired fundus images include a first view left eye fundus image, a second view left eye fundus image, a first view right eye fundus image, and a second view right eye fundus image from the same patient;
the processor inputs the combined feature set into the classifier, and when a classification result is obtained, the following steps are also realized: connecting the combined feature set of the left eye fundus image in the first visual field, the combined feature set of the left eye fundus image in the second visual field, the combined feature set of the right eye fundus image in the first visual field and the combined feature set of the right eye fundus image in the second visual field to generate a combined feature sequence of the eye fundus images; and inputting the combined characteristic sequence into a classifier to obtain a classification result.
In one embodiment, the second neural network is a convolutional neural network capable of identifying the level of fundus lesions when the fundus image comprises lesionsnIn stage lesions, a second feature set of length is generatedn-1, wherein,imedian anterior in feature vector of grade lesioniIs 1, the rest is 0.
In one embodiment, a computer-readable storage medium is provided, having a computer program stored thereon, which when executed by a processor, performs the steps of:
receiving an acquired fundus image;
identifying a fundus image through a first neural network to generate a first feature set of the fundus image;
identifying the fundus image through a second neural network, and generating a second feature set of the fundus image, wherein the first feature set and the second feature set represent different lesion attributes of the fundus image;
combining the first feature set and the second feature set to obtain a combined feature set of the fundus image;
and inputting the combined feature set into a classifier to obtain a classification result.
In one embodiment, the first neural network is a convolutional neural network capable of identifying the type of lesion contained in the fundus image, the second neural network is a convolutional neural network capable of identifying the grade of the fundus lesion, and the classifier is a multi-stage classification network composed of a plurality of two-class classifiers according to a set classification logic.
In one embodiment, the processor, when executing the identifying the fundus image by the first neural network, resulting in the first feature set of the fundus image, further performs the steps of:
quadrant segmentation is carried out on the fundus image to generate a quadrant image group;
inputting each quadrant image in the quadrant image group into a first neural network to obtain a characteristic vector corresponding to each quadrant image;
the combined feature vector generates a first feature set of the fundus image.
In one embodiment, the acquired fundus images include a left eye fundus image and a right eye fundus image from the same patient;
the processor inputs the combined feature set into the classifier, and when a classification result is obtained, the following steps are also realized: connecting the combined characteristic set of the left eye fundus image and the combined characteristic set of the right eye fundus image to generate a combined characteristic sequence of the fundus images; and inputting the combined characteristic sequence into a classifier to obtain a classification result.
In one embodiment, the acquired fundus images include a first view left eye fundus image, a second view left eye fundus image, a first view right eye fundus image, and a second view right eye fundus image from the same patient;
the processor inputs the combined feature set into the classifier, and when a classification result is obtained, the following steps are also realized: connecting the combined feature set of the left eye fundus image in the first visual field, the combined feature set of the left eye fundus image in the second visual field, the combined feature set of the right eye fundus image in the first visual field and the combined feature set of the right eye fundus image in the second visual field to generate a combined feature sequence of the eye fundus images; and inputting the combined characteristic sequence into a classifier to obtain a classification result.
In one embodiment, the second neural network is a convolutional neural network capable of identifying the level of fundus lesions when the fundus image comprises lesionsnIn stage lesions, a second feature set of length is generatedn-1, wherein,icharacteristics of grade lesionsVector middle frontiIs 1, the rest is 0.
It will be understood by those skilled in the art that all or part of the processes of the methods of the embodiments described above can be implemented by hardware related to instructions of a computer program, which can be stored in a non-volatile computer-readable storage medium, and when executed, can include the processes of the embodiments of the methods described above. Any reference to memory, storage, database, or other medium used in the embodiments provided herein may include non-volatile and/or volatile memory, among others. Non-volatile memory can include read-only memory (ROM), Programmable ROM (PROM), Electrically Programmable ROM (EPROM), Electrically Erasable Programmable ROM (EEPROM), or flash memory. Volatile memory can include Random Access Memory (RAM) or external cache memory. By way of illustration and not limitation, RAM is available in a variety of forms such as Static RAM (SRAM), Dynamic RAM (DRAM), Synchronous DRAM (SDRAM), Double Data Rate SDRAM (DDRSDRAM), Enhanced SDRAM (ESDRAM), Synchronous Link DRAM (SLDRAM), Rambus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM).
The technical features of the above embodiments can be arbitrarily combined, and for the sake of brevity, all possible combinations of the technical features in the above embodiments are not described, but should be considered as the scope of the present specification as long as there is no contradiction between the combinations of the technical features.
The above examples only express several embodiments of the present application, and the description thereof is more specific and detailed, but not construed as limiting the scope of the invention. It should be noted that, for a person skilled in the art, several variations and modifications can be made without departing from the concept of the present application, which falls within the scope of protection of the present application. Therefore, the protection scope of the present patent shall be subject to the appended claims.

Claims (10)

1. A method of fundus image processing, the method comprising:
receiving an acquired fundus image;
identifying the fundus image through a first neural network to generate a first feature set of the fundus image;
identifying the fundus image through a second neural network, and generating a second feature set of the fundus image, wherein the first feature set and the second feature set represent different lesion attributes of the fundus image; the first neural network is a convolutional neural network capable of identifying a type of a lesion contained in the fundus image, and the second neural network is a convolutional neural network capable of identifying a lesion level of a fundus lesion contained in the fundus image; the fundus lesion type identified by the first neural network comprises: microangiomas, hard extravasation, soft extravasation, and hemorrhage; the first feature set output by the first neural network is a feature vector with the length of 4, and the first neural network is trained to enable each element of the output feature vector to sequentially represent the corresponding lesion type;
combining the first feature set and the second feature set to obtain a combined feature set of the fundus image;
inputting the combined feature set into a classifier to obtain a classification result;
the identifying the fundus image by the second neural network generates a second feature set of the fundus image comprising: identifying the lesion level attribute of the fundus image through a second neural network, and outputting a lesion level vector of the fundus image, wherein when the fundus lesion is set to contain n-level lesions, the length of the generated lesion level vector is n-1, wherein the front i in the feature vector of the i-level lesion is 1, and the rest is 0.
2. The method of claim 1, wherein the first set of features characterizes a lesion type attribute of the fundus image and the second set of features characterizes a lesion level attribute of the fundus image;
inputting the combined feature set into a classifier to obtain a classification result, wherein the classification result comprises:
and inputting the combined feature set with the lesion type attribute and the lesion level attribute into a multi-stage classifier which is composed of a plurality of two-stage classifiers according to set classification logic to obtain a multi-stage classification result of the fundus image.
3. The method of claim 2, wherein identifying the fundus image via the first neural network to obtain the first set of features for the fundus image comprises:
quadrant segmentation is carried out on the fundus image to generate a quadrant image group;
inputting each quadrant image in the quadrant image group into a first neural network to obtain a feature vector of each quadrant image;
combining the feature vectors of each quadrant image generates a first feature set of the fundus image.
4. A method according to any one of claims 1 to 3, wherein the fundus images received comprise left and right eye fundus images from the same patient;
inputting the combined feature set into a classifier, and obtaining a classification result comprises:
connecting the combined feature set of the left eye fundus image and the combined feature set of the right eye fundus image to generate a combined feature sequence of the fundus images;
and inputting the combined characteristic sequence into a classifier to obtain a classification result.
5. The method of any one of claims 1-3, wherein the received fundus images include a first view left eye fundus image, a second view left eye fundus image, a first view right eye fundus image, and a second view right eye fundus image from the same patient;
inputting the combined feature set into a classifier, and obtaining a classification result comprises:
connecting the combined feature set of the first-view left eye fundus image, the combined feature set of the second-view left eye fundus image, the combined feature set of the first-view right eye fundus image and the combined feature set of the second-view right eye fundus image to generate a combined feature sequence of the fundus images;
and inputting the combined characteristic sequence into a classifier to obtain a classification result.
6. The method of claim 1,
the combination of the first set of features and the second set of features is a vector sum of the features.
7. An eye fundus image processing apparatus, characterized in that the apparatus comprises:
the image acquisition module is used for receiving the acquired fundus images;
the first neural network identification module is used for identifying the fundus image through a first neural network to generate a first feature set of the fundus image;
the second neural network identification module is used for identifying the fundus image through a second neural network and generating a second feature set of the fundus image, wherein the first feature set and the second feature set represent different lesion attributes of the fundus image; the first neural network is a convolutional neural network capable of identifying a type of a lesion contained in the fundus image, and the second neural network is a convolutional neural network capable of identifying a lesion level of a fundus lesion contained in the fundus image; the fundus lesion type identified by the first neural network comprises: microangiomas, hard extravasation, soft extravasation, and hemorrhage; the first feature set output by the first neural network is a feature vector with the length of 4, and the first neural network is trained to enable each element of the output feature vector to sequentially represent the corresponding lesion type;
the characteristic combination module is used for combining the first characteristic set and the second characteristic set to obtain a combined characteristic set of the fundus image;
the classification module is used for inputting the combined feature set into a classifier to obtain a classification result;
the identifying the fundus image by the second neural network generates a second feature set of the fundus image comprising: identifying the lesion level attribute of the fundus image through a second neural network, and outputting a lesion level vector of the fundus image, wherein when the fundus lesion is set to contain n-level lesions, the length of the generated lesion level vector is n-1, wherein the front i in the feature vector of the i-level lesion is 1, and the rest is 0.
8. The apparatus of claim 7, wherein the first set of features characterizes a lesion type attribute of the fundus image and the second set of features characterizes a lesion level attribute of the fundus image;
the classification module is further used for inputting the combined feature set with the lesion type attribute and the lesion grade attribute into a multi-stage classifier which is composed of a plurality of two-type classifiers according to set classification logics, and obtaining a multi-stage classification result of the fundus image.
9. A computer device comprising a memory and a processor, the memory storing a computer program, wherein the processor implements the steps of the method of any one of claims 1 to 6 when executing the computer program.
10. A computer-readable storage medium, on which a computer program is stored, which, when being executed by a processor, carries out the steps of the method of any one of claims 1 to 6.
CN201810340025.4A 2017-08-02 2018-04-16 Fundus image processing method, fundus image processing apparatus, computer device, and storage medium Active CN109390053B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2018/086739 WO2019024568A1 (en) 2017-08-02 2018-05-14 Ocular fundus image processing method and apparatus, computer device, and storage medium
US16/302,410 US11200665B2 (en) 2017-08-02 2018-05-14 Fundus image processing method, computer apparatus, and storage medium

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201710653516 2017-08-02
CN201710653516X 2017-08-02

Publications (2)

Publication Number Publication Date
CN109390053A CN109390053A (en) 2019-02-26
CN109390053B true CN109390053B (en) 2021-01-08

Family

ID=65416517

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810340025.4A Active CN109390053B (en) 2017-08-02 2018-04-16 Fundus image processing method, fundus image processing apparatus, computer device, and storage medium

Country Status (1)

Country Link
CN (1) CN109390053B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110263755B (en) * 2019-06-28 2021-04-27 上海鹰瞳医疗科技有限公司 Eye ground image recognition model training method, eye ground image recognition method and eye ground image recognition device
CN110276333B (en) * 2019-06-28 2021-10-15 上海鹰瞳医疗科技有限公司 Eye ground identity recognition model training method, eye ground identity recognition method and equipment
CN110570421B (en) * 2019-09-18 2022-03-22 北京鹰瞳科技发展股份有限公司 Multitask fundus image classification method and apparatus
CN110796161A (en) * 2019-09-18 2020-02-14 平安科技(深圳)有限公司 Recognition model training method, recognition device, recognition equipment and recognition medium for eye ground characteristics
CN113449774A (en) * 2021-06-02 2021-09-28 北京鹰瞳科技发展股份有限公司 Fundus image quality control method, device, electronic apparatus, and storage medium

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10115194B2 (en) * 2015-04-06 2018-10-30 IDx, LLC Systems and methods for feature detection in retinal images
CN104881683B (en) * 2015-05-26 2018-08-28 清华大学 Cataract eye fundus image sorting technique based on assembled classifier and sorter
CN106934798B (en) * 2017-02-20 2020-08-21 苏州体素信息科技有限公司 Diabetic retinopathy classification and classification method based on deep learning
CN106874889B (en) * 2017-03-14 2019-07-02 西安电子科技大学 Multiple features fusion SAR target discrimination method based on convolutional neural networks
CN108771530B (en) * 2017-05-04 2021-03-30 深圳硅基智能科技有限公司 Fundus lesion screening system based on deep neural network

Also Published As

Publication number Publication date
CN109390053A (en) 2019-02-26

Similar Documents

Publication Publication Date Title
US11200665B2 (en) Fundus image processing method, computer apparatus, and storage medium
CN109390053B (en) Fundus image processing method, fundus image processing apparatus, computer device, and storage medium
US11842487B2 (en) Detection model training method and apparatus, computer device and storage medium
US10482603B1 (en) Medical image segmentation using an integrated edge guidance module and object segmentation network
US10430946B1 (en) Medical image segmentation and severity grading using neural network architectures with semi-supervised learning techniques
CN110120040B (en) Slice image processing method, slice image processing device, computer equipment and storage medium
Costa et al. End-to-end adversarial retinal image synthesis
WO2021159774A1 (en) Object detection model training method and apparatus, object detection method and apparatus, computer device, and storage medium
CN110008971B (en) Image processing method, image processing device, computer-readable storage medium and computer equipment
Raja et al. An automated early detection of glaucoma using support vector machine based visual geometry group 19 (VGG-19) convolutional neural network
Xiang et al. Towards interpretable skin lesion classification with deep learning models
CN113642537B (en) Medical image recognition method and device, computer equipment and storage medium
CN112241952B (en) Brain midline identification method, device, computer equipment and storage medium
CN111325766B (en) Three-dimensional edge detection method, three-dimensional edge detection device, storage medium and computer equipment
CN113706564A (en) Meibomian gland segmentation network training method and device based on multiple supervision modes
CN113781488A (en) Tongue picture image segmentation method, apparatus and medium
CN110929730A (en) Image processing method, image processing device, computer equipment and storage medium
Sundar et al. An effective deep learning model for grading abnormalities in retinal fundus images using variational auto‐encoders
Pal et al. Advanced Convolutional Neural Network Model to Identify Melanoma Skin Cancer
Zunaed et al. Learning to generalize towards unseen domains via a content-aware style invariant model for disease detection from chest X-rays
Schaefer et al. Scleroderma capillary pattern identification using texture descriptors and ensemble classification
CN113674228A (en) Identification method and device for brain blood supply area, storage medium and electronic equipment
CN114170224A (en) System and method for cellular pathology classification using generative staining normalization
Ouassit et al. Liver Segmentation
Zheng et al. Enhancing Contrastive Learning for Ordinal Regression via Ordinal Content Preserved Data Augmentation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant