WO2018201647A1 - Method for detecting retinopathy degree level, device and storage medium - Google Patents

Method for detecting retinopathy degree level, device and storage medium Download PDF

Info

Publication number
WO2018201647A1
WO2018201647A1 PCT/CN2017/100044 CN2017100044W WO2018201647A1 WO 2018201647 A1 WO2018201647 A1 WO 2018201647A1 CN 2017100044 W CN2017100044 W CN 2017100044W WO 2018201647 A1 WO2018201647 A1 WO 2018201647A1
Authority
WO
WIPO (PCT)
Prior art keywords
retinopathy
picture
degree
model
preset
Prior art date
Application number
PCT/CN2017/100044
Other languages
French (fr)
Chinese (zh)
Inventor
王健宗
吴天博
黄章成
肖京
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2018201647A1 publication Critical patent/WO2018201647A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/03Recognition of patterns in medical or anatomical images

Definitions

  • the present invention relates to the field of computer technologies, and in particular, to a method for detecting a degree of retinopathy degree, an electronic device, and a computer readable storage medium.
  • diabetic retinopathy is a leading cause of blindness in the eye.
  • the identification of diabetic retinopathy usually requires feature extraction of the ocular image (for example, ocular vascular structure).
  • the extraction of features such as optic disc and retinal center groove, the feature extraction algorithm has poor performance and poor performance.
  • the main object of the present invention is to provide a method for detecting the degree of retinopathy degree, an electronic device and a computer readable storage medium, which are intended to accurately and accurately identify the degree of retinopathy of a patient.
  • an electronic device includes: a memory, a processor, and a memory of a retinopathy degree level detecting system operable on the processor.
  • the program when the program of the retinopathy degree level detection system is executed by the processor, implements the following steps:
  • the received retinopathy picture is identified by using a predetermined recognition model, and the recognition result is output; wherein the predetermined recognition model is pre-marked by different retinopathy a convolutional neural network model obtained by training a predetermined number of sample pictures of a degree level;
  • the degree of retinopathy corresponding to the output recognition result is determined.
  • the training process of the predetermined recognition model is as follows:
  • the present invention also provides a method for detecting a degree of retinopathy degree, the method comprising the following steps:
  • the received retinopathy picture is identified by using a predetermined recognition model, and the recognition result is output; wherein the predetermined recognition model is pre-marked by different retinopathy a convolutional neural network model obtained by training a predetermined number of sample pictures of a degree level;
  • the degree of retinopathy corresponding to the output recognition result is determined.
  • the training process of the predetermined recognition model is as follows:
  • the present invention also provides a computer readable storage medium storing a program of a detection system for a degree of retinopathy degree, a program of a detection system for a degree of retinopathy Any step of the method of detecting the level of retinopathy as described above when executed by the processor.
  • the retinopathy degree level detecting method, the electronic device and the computer readable storage medium proposed by the present invention are received by a deep convolutional neural network model based on a preset number of sample pictures marked with different retinopathy degree levels.
  • the retinopathy picture is identified, and the corresponding retinopathy degree level is determined according to the recognition result. Since only the pre-trained deep convolutional neural network model is needed to recognize the received retinal lesion image, it is not necessary to perform complex feature extraction operation on the eye image, which is simpler and can determine corresponding different retinopathy according to the recognition result.
  • Degree level can effectively identify the degree of retinopathy of patients.
  • FIG. 1 is a schematic flow chart of a preferred embodiment of a method for detecting a degree of retinopathy of the present invention
  • FIG. 2 is a schematic view of a preferred embodiment of an electronic device of the present invention.
  • FIG. 3 is a functional block diagram of a preferred embodiment of the retinopathy lesion level detecting system of FIG. 2.
  • the invention provides a method for detecting the degree of retinopathy.
  • FIG. 1 is a schematic flow chart of a preferred embodiment of a method for detecting a degree of retinopathy of the present invention.
  • the method for detecting the degree of retinopathy is as follows:
  • Step S10 After receiving the retinopathy picture to be identified, the received retinopathy picture is identified by using a predetermined recognition model, and the recognition result is output; wherein the predetermined recognition model is pre-marked by A deep convolutional neural network model obtained by training a preset number of sample images of different degrees of retinopathy.
  • the retinopathy degree level detecting system receives a retinopathy degree level detection request sent by the user and includes a retinopathy degree image to be recognized, for example, receiving a degree of retinopathy transmitted by the user through a terminal such as a mobile phone, a tablet computer, or a self-service terminal device.
  • a level detection request such as receiving a retinopathy degree level detection request sent by a user on a pre-installed client in a terminal such as a mobile phone, a tablet computer, a self-service terminal device, or receiving a user in a terminal such as a mobile phone, a tablet computer, or a self-service terminal device.
  • Retinopathy level level detection request sent on the browser system.
  • the retinopathy degree level detection system recognizes the received retinopathy image to be recognized by using the pre-trained recognition model after receiving the request for the retinopathy degree level detection issued by the user, and recognizes the retinopathy to be recognized that is to be recognized.
  • the recognition model can be continuously trained, learned, verified, optimized, etc. by identifying a plurality of preset number of sample pictures marked with different degrees of retinopathy degree, so as to train them to accurately identify different levels of retinopathy.
  • the model of the annotation may employ a Convolutional Neural Network (CNN) model or the like.
  • CNN Convolutional Neural Network
  • Step S20 Determine a retinopathy degree level corresponding to the output recognition result according to a mapping relationship between the predetermined recognition result and the retinopathy degree level.
  • the mapping result of the predetermined recognition result and the retinopathy degree level may be determined according to the output recognition result.
  • the degree of retinopathy is determined by the degree of retinopathy level of the retinopathy image received.
  • the recognition result includes a first recognition result (eg, labeled "0"), a second recognition result (eg, labeled "1"), and a third recognition result (eg, an annotation a "2"), a fourth recognition result (for example, labeled "3"), and a fifth recognition result (for example, labeled "4"),
  • the retinopathy degree level includes a first level, a second level, and a Three levels, fourth level and fifth level.
  • the mapping relationship between the different recognition results and the retinopathy degree level may be determined in advance, if the first recognition result corresponds to the first level, the second recognition result corresponds to the second level, and the third recognition result corresponds to the third level, and the fourth recognition result corresponds to The fourth level, the fifth recognition result corresponds to the fifth level.
  • the first level may correspond to normal and mild non-proliferative diabetic retinopathy, and the first level corresponding retinopathy picture appears as only individual hemangiomas, hard exudation, retinal hemorrhage, and the like.
  • the second grade can correspond to non-proliferative diabetic retinopathy without clinically significant macular edema.
  • the second grade corresponding retinopathy picture shows microangioma, hard exudation, retinal hemorrhage, sickle or beaded vein.
  • the third grade can correspond to non-proliferative diabetic retinopathy with clinically significant macular edema (CSME).
  • the third grade corresponding retinopathy picture shows retinal thickening in the macular area and its vicinity, and microangioma, soft infiltration Out, retinal hemorrhage.
  • the fourth grade can correspond to non-high-risk proliferative retinopathy.
  • the fourth-level retinal lesion picture shows neovascularization in the outer area of the optic papilla and proliferative changes in retinal microvessel formation in other areas.
  • the fifth grade can correspond to high-risk proliferative retinopathy.
  • the fifth grade corresponding retinopathy picture shows neovascularization, vitreous or preretinal hemorrhage in the optic papilla area.
  • the corresponding different retinopathy degree levels can be determined according to the obtained different recognition results, thereby realizing various refinements. Accurate identification of the degree of retinopathy.
  • the received retinal lesion image is identified by a deep convolutional neural network model based on a preset number of sample images labeled with different degrees of retinopathy degree, and the corresponding degree of retinopathy is determined according to the recognition result. . Since only the pre-trained deep convolutional neural network model is needed to recognize the received retinal lesion image, it is not necessary to perform complex feature extraction operation on the eye image, which is simpler and can determine corresponding different retinopathy according to the recognition result. Degree level, can effectively identify the degree of retinopathy of patients.
  • the training process of the predetermined recognition model is as follows:
  • preset level of retinopathy such as first level, second level, third level, fourth level, and fifth level, or slight, mild, moderate, severe, etc.
  • performing image preprocessing on each sample picture may include:
  • Scaling a shorter side length of each sample picture to a first preset size eg, 640 pixels
  • a second preset size eg, a second picture of 256*256 pixels
  • the standard parameter value corresponding to each predetermined preset type parameter for example, color, brightness, and/or contrast, etc.
  • the standard parameter value corresponding to the color is a1
  • the standard parameter value corresponding to the brightness is a2
  • the contrast corresponds
  • the standard parameter value is a3
  • each predetermined preset type parameter value of each second picture is adjusted to a corresponding standard parameter value, and a corresponding third picture is obtained to eliminate the sample picture as a medical picture at the time of shooting.
  • the picture caused by the condition is not clear, improve the model training
  • the effectiveness of the practice for example, adjusting the brightness value of each second picture to the standard parameter value a2, adjusting the contrast value of each second picture to the standard parameter value a3;
  • each fourth picture is a training picture of the corresponding sample picture.
  • the function of the flip and twist operation is to simulate various forms of pictures in the actual business scene. Through these flip and twist operations, the size of the data set can be increased, thereby improving the authenticity and practicability of the model training.
  • the predetermined recognition model that is, the deep convolutional neural network model
  • the predetermined recognition model includes an input layer and a plurality of network layers
  • the network layer includes a convolution layer, a pooling layer, a fully connected layer, and a classification.
  • the layer optionally, the deep convolutional neural network model may also include a network layer (ie, a Dropout layer) having a random drop of some connection weighting mechanism, the role of which is to improve the recognition accuracy of the model.
  • the deep convolutional neural network model consists of one input layer, eleven convolutional layers, five pooling layers, and one network layer having random connection discarding certain connection weighting mechanisms (ie, Dropout layer), 1 fully connected layer, 1 classifier layer.
  • connection weighting mechanisms ie, Dropout layer
  • Layer Name indicates the name of the network layer
  • Input indicates the data input layer of the network
  • Conv indicates the convolution layer of the model
  • Conv1 indicates the first convolution layer of the model
  • MaxPool indicates the maximum pooling layer of the model
  • MaxPool1 indicates the first One based on the maximum pooling layer
  • Dropout means a network layer with random discarding of some connection weighting mechanism
  • Avgpool5 means the fifth pooling layer but pooled by means of averaging
  • Fc represents the fully connected layer in the model
  • Fc1 Represents the first fully connected layer
  • Softmax represents the Softmax classifier layer
  • Batch Size represents the number of input images of the current layer
  • Kernel Size represents the scale of the current layer convolution kernel (for example, Kernel Size can be equal to 3, indicating the scale of the convolution kernel) 3x3)
  • Stride Size represents the moving step size of the convolution kernel, that is, the distance moved to the next convolution position after one convolution
  • Output Size
  • pooling mode of the pooling layer in this embodiment includes, but is not limited to, Mean pooling, Max pooling, Overlapping, L2 pooling, Local. Contrast Normalization, Stochasticpooling, Def-pooling, and more.
  • each of the network layers (for example, a convolution layer, a pooling layer, a network layer with a random drop of some connection weight mechanism, a fully connected layer, and a classifier)
  • the corresponding activation function f(x) of the layer, etc. is:
  • is the leak rate and x is a numerical input of the neurons in the deep convolutional neural network model.
  • is set to 0.5.
  • each of the network layers (for example, a convolution layer, a pooling layer, a network layer with a random drop of some connection weight mechanism, a fully connected layer, and a classifier)
  • the corresponding cross entropy H(P, Q) of the layer, etc. is:
  • H(P) is the expectation of probability distribution P
  • H(P) - ⁇ x ⁇ X P(x)log P(x)
  • x is the sample space of probability distribution P
  • P(x) represents the probability that sample x is selected
  • Q) is x is the probability distribution P and Q any sample in the common sample space X
  • P(x) represents the probability that the sample x is selected on the probability distribution P
  • Q(x) represents the probability that the sample x is selected on the probability distribution Q.
  • the cross entropy loss function corresponding to each of the network layers for:
  • x is the input to the model, Indicates the corresponding label of the input
  • W represents the preset model parameters
  • X represents the model input space
  • f(x:W) represents the transformed output of the model to the input x
  • denotes the statistic factor
  • 2 means summing matrix elements:
  • ⁇ W i+1 represents the update increment of the weight matrix at time i+1
  • is the potential energy term
  • is the weight attenuation coefficient
  • is the learning rate of the model
  • W i is the state value of the weight matrix at time i
  • D i represents the i-th batch input
  • cross entropy can be used as a loss function in neural networks (machine learning).
  • P represents the distribution of real markers
  • Q is the predicted marker distribution of the trained model
  • the cross entropy loss function can measure P and Q. The similarity to ensure the accuracy of the model training.
  • the cross entropy as a loss function can avoid the problem of the learning rate reduction of the mean square error loss function when the gradient is lowered, and therefore, the efficiency of the model training can be ensured.
  • the deep convolutional neural network model includes at least one fully connected layer, and an initial value of each weight in the predetermined recognition model is from a preset weight range (eg, (0, 1) Weight range) Performing random sampling determination, the probability that the connection weight of the fully connected layer is discarded (Dropout) is set to a first preset value (for example, 0.5), and the weight attenuation coefficient in the cross entropy loss function Set to a second preset value (eg, 0.0005), the potential energy term in the cross entropy loss function is set to a third preset value (eg, 0.9).
  • a preset weight range eg, (0, 1) Weight range
  • the probability that the connection weight of the fully connected layer is discarded is set to a first preset value (for example, 0.5)
  • the weight attenuation coefficient in the cross entropy loss function Set to a second preset value (eg, 0.0005)
  • the potential energy term in the cross entropy loss function is set to a third prese
  • the predetermined scoring function of the recognition model for:
  • O i,j represents the first prediction as i and the second prediction is the number of pictures actually appearing in j
  • O represents a matrix of N*N
  • O i,j represents the matrix element in matrix O
  • N represents the prediction of participation.
  • E i, j represents the number of images in which the first prediction is i and the second prediction is j
  • E is the N* of the desired prediction result.
  • the N matrix, E i,j represents the matrix elements in the matrix E.
  • Scoring function in this embodiment The recognition accuracy of the predetermined recognition model is detected to ensure that the recognition accuracy of the trained recognition model is maintained at a high level to ensure accurate recognition of the degree of retinopathy of the patient.
  • the invention further provides an electronic device. Please refer to FIG. 2, which is a schematic diagram of a preferred embodiment of the electronic device of the present invention.
  • the retinopathy degree level detecting method is applied to an electronic device 1.
  • the electronic device 1 may include, but is not limited to, a memory 11, a processor 12, and a display 13.
  • Figure 2 shows only the electronic device 1 with components 11-13, but it should be understood that not all illustrated components may be implemented, and more or fewer components may be implemented instead.
  • the memory 11 may in some embodiments be an internal storage unit of the electronic device 1, such as a hard disk or memory of the electronic device 1.
  • the memory 11 may also be an external storage device of the electronic device 1 in other embodiments, such as a plug-in hard disk equipped on the electronic device 1, a smart memory card (SMC), and a secure digital device. (Secure Digital, SD) card, flash card, etc.
  • SMC smart memory card
  • secure digital device Secure Digital, SD
  • the memory 11 may also include both an internal storage unit of the electronic device 1 and an external storage device.
  • the memory 11 is configured to store application software and various types of data installed in the electronic device 1, such as program codes of the retinopathy degree level detecting system 10, and the like.
  • the memory 11 can also be used to temporarily store data that has been output or is about to be output.
  • the processor 12 in some embodiments, may be a central processing unit (CPU), a microprocessor or other data processing chip for running program code or processing data stored in the memory 11, for example
  • the retinopathy degree level detecting system 10 and the like are executed.
  • the display 13 in some embodiments may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode) touch sensor, or the like.
  • the display 13 is used to display information processed in the electronic device 1 and a user interface for displaying visualization, such as an application menu interface, an application icon interface, and the like.
  • the components 11-13 of the electronic device 1 communicate with one another via a system bus.
  • the program of the retinopathy degree level detecting system 10 is stored in the memory 11; when the processor 12 executes the program of the retinopathy degree level detecting system 10 stored in the memory 11, the following steps are implemented:
  • the received retinopathy picture is identified by using a predetermined recognition model, and the recognition result is output; wherein the predetermined recognition model is pre-marked by different retinopathy a deep convolutional neural network model obtained by training a predetermined number of sample pictures of a degree level; and
  • the degree of retinopathy corresponding to the output recognition result is determined.
  • the retinopathy degree level detecting system receives a retinopathy degree level detection request sent by the user and includes a retinopathy degree image to be recognized, for example, receiving a degree of retinopathy transmitted by the user through a terminal such as a mobile phone, a tablet computer, or a self-service terminal device.
  • a level detection request such as receiving a retinopathy degree level detection request sent by a user on a pre-installed client in a terminal such as a mobile phone, a tablet computer, a self-service terminal device, or receiving a user in a terminal such as a mobile phone, a tablet computer, or a self-service terminal device.
  • Retinopathy level level detection request sent on the browser system.
  • the retinopathy degree level detection system recognizes the received retinopathy image to be recognized by using the pre-trained recognition model after receiving the request for the retinopathy degree level detection issued by the user, and recognizes the retinopathy to be recognized that is to be recognized.
  • the recognition model can be continuously trained, learned, verified, optimized, etc. by identifying a plurality of preset number of sample pictures marked with different degrees of retinopathy degree, so as to train them to accurately identify different levels of retinopathy.
  • the model of the annotation may employ a Convolutional Neural Network (CNN) model or the like.
  • CNN Convolutional Neural Network
  • the mapping result of the predetermined recognition result and the retinopathy degree level may be determined according to the output recognition result.
  • the degree of retinopathy is determined by the degree of retinopathy level of the retinopathy image received.
  • the recognition result includes a first recognition result (eg, labeled "0"), a second recognition result (eg, labeled "1"), and a third recognition result (eg, an annotation a "2"), a fourth recognition result (for example, labeled "3"), and a fifth recognition result (for example, labeled "4"),
  • the retinopathy degree level includes a first level, a second level, The third level, the fourth level, and the fifth level.
  • the mapping relationship between the different recognition results and the retinopathy degree level may be determined in advance, if the first recognition result corresponds to the first level, the second recognition result corresponds to the second level, and the third recognition result corresponds to the third level, and the fourth recognition result corresponds to The fourth level, the fifth recognition result corresponds to the fifth level.
  • the first level may correspond to normal and mild non-proliferative diabetic retinopathy, and the first level corresponding retinopathy picture appears as only individual hemangiomas, hard exudation, retinal hemorrhage, and the like.
  • the second grade can correspond to non-proliferative diabetic retinopathy without clinically significant macular edema.
  • the second grade corresponding retinopathy picture shows microangioma, hard exudation, retinal hemorrhage, sickle or beaded vein.
  • the third grade can correspond to non-proliferative diabetic retinopathy with clinically significant macular edema (CSME).
  • the third grade corresponding retinopathy picture shows retinal thickening in the macular area and its vicinity, and microangioma, soft infiltration Out, retinal hemorrhage.
  • the fourth grade can correspond to non-high-risk proliferative retinopathy.
  • the fourth-level retinal lesion picture shows neovascularization in the outer area of the optic papilla and proliferative changes in retinal microvessel formation in other areas.
  • the fifth grade can correspond to high-risk proliferative retinopathy.
  • the fifth grade corresponding retinopathy picture shows neovascularization, vitreous or preretinal hemorrhage in the optic papilla area.
  • the corresponding different retinas can be determined according to the obtained different recognition results.
  • the degree of lesion severity thus enabling accurate identification of a variety of refinement levels of retinopathy.
  • the received retinal lesion image is identified by a deep convolutional neural network model based on a preset number of sample images labeled with different degrees of retinopathy degree, and the corresponding degree of retinopathy is determined according to the recognition result. . Since only the pre-trained deep convolutional neural network model is needed to recognize the received retinal lesion image, it is not necessary to perform complex feature extraction operation on the eye image, which is simpler and can determine corresponding different retinopathy according to the recognition result. Degree level, can effectively identify the degree of retinopathy of patients.
  • the training process of the predetermined recognition model is as follows:
  • preset level of retinopathy such as first level, second level, third level, fourth level, and fifth level, or slight, mild, moderate, severe, etc.
  • performing image preprocessing on each sample picture may include:
  • Scaling a shorter side length of each sample picture to a first preset size eg, 640 pixels
  • a second preset size eg, a second picture of 256*256 pixels
  • the standard parameter value corresponding to each predetermined preset type parameter for example, color, brightness, and/or contrast, etc.
  • the standard parameter value corresponding to the color is a1
  • the standard parameter value corresponding to the brightness is a2
  • the contrast corresponds
  • the standard parameter value is a3
  • each predetermined preset type parameter value of each second picture is adjusted to a corresponding standard parameter value, and a corresponding third picture is obtained to eliminate the sample picture as a medical picture at the time of shooting.
  • the picture caused by the condition is not clear, and the effectiveness of the model training is improved; for example, the brightness value of each second picture is adjusted to the standard parameter value a2, and the contrast value of each second picture is adjusted to the standard parameter value a3;
  • each fourth picture is a training picture of the corresponding sample picture.
  • the function of the flip and twist operation is to simulate various forms of pictures in the actual business scene. Through these flip and twist operations, the size of the data set can be increased, thereby improving the authenticity and practicability of the model training.
  • the training ends, or if the accuracy rate is less than the preset accuracy rate, increase each The number of sample pictures corresponding to the level of retinopathy is re-executed in steps B, C, D, and E until the accuracy of the trained recognition model is greater than or equal to the preset accuracy.
  • the predetermined recognition model that is, the deep convolutional neural network model
  • the predetermined recognition model includes an input layer and a plurality of network layers
  • the network layer includes a convolution layer, a pooling layer, a fully connected layer, and a classification.
  • the layer optionally, the deep convolutional neural network model may also include a network layer (ie, a Dropout layer) having a random drop of some connection weighting mechanism, the role of which is to improve the recognition accuracy of the model.
  • the deep convolutional neural network model consists of one input layer, eleven convolutional layers, five pooling layers, and one network layer having random connection discarding certain connection weighting mechanisms (ie, Dropout layer), 1 fully connected layer, 1 classifier layer.
  • connection weighting mechanisms ie, Dropout layer
  • Layer Name indicates the name of the network layer
  • Input indicates the data input layer of the network
  • Conv indicates the convolution layer of the model
  • Conv1 indicates the first convolution layer of the model
  • MaxPool indicates the maximum pooling layer of the model
  • MaxPool1 indicates the first One based on the maximum pooling layer
  • Dropout means a network layer with random discarding of some connection weighting mechanism
  • Avgpool5 means the fifth pooling layer but pooled by means of averaging
  • Fc represents the fully connected layer in the model
  • Fc1 Represents the first fully connected layer
  • Softmax represents the Softmax classifier layer
  • Batch Size represents the number of input images of the current layer
  • Kernel Size represents the scale of the current layer convolution kernel (for example, Kernel Size can be equal to 3, indicating the scale of the convolution kernel) 3x3)
  • Stride Size represents the moving step size of the convolution kernel, that is, the distance moved to the next convolution position after one convolution
  • Output Size
  • pooling mode of the pooling layer in this embodiment includes, but is not limited to, Mean pooling, Max pooling, Overlapping, L2pooling, Local Contrast. Normalization, Stochasticpooling, Def-pooling, and more.
  • each of the network layers (for example, a convolution layer, a pooling layer, a network layer with a random drop of some connection weight mechanism, a fully connected layer, and a classifier)
  • the corresponding activation function f(x) of the layer, etc. is:
  • is the leak rate and x is a numerical input of the neurons in the deep convolutional neural network model.
  • is set to 0.5.
  • each of the network layers (for example, a convolution layer, a pooling layer, a network layer with a random drop of some connection weight mechanism, a fully connected layer, and a classifier)
  • the corresponding cross entropy H(P, Q) of the layer, etc. is:
  • H(P) is the expectation of probability distribution P
  • H(P) - ⁇ x ⁇ X P(x)log P(x)
  • x is the sample space of probability distribution P
  • P(x) represents the probability that sample x is selected
  • Q) is x is any one of the probability distribution P and the Q common sample space X
  • P(x) represents the probability that the sample x is selected on the probability distribution P
  • Q(x) represents the probability that the sample x is selected on the probability distribution Q.
  • the cross entropy loss function corresponding to each of the network layers for:
  • x is the input to the model, Indicates the corresponding label of the input
  • W represents the preset model parameters
  • X represents the model input space
  • f(x:W) represents the transformed output of the model to the input x
  • denotes the statistic factor
  • 2 means summing matrix elements:
  • ⁇ W i+1 represents the update increment of the weight matrix at time i+1
  • is the potential energy term
  • is the weight attenuation coefficient
  • is the learning rate of the model
  • W i is the state value of the weight matrix at time i
  • D i represents the i-th batch input
  • cross entropy can be used as a loss function in neural networks (machine learning).
  • P represents the distribution of real markers
  • Q is the predicted marker distribution of the trained model
  • the cross entropy loss function can measure P and Q. The similarity to ensure the accuracy of the model training.
  • the cross entropy as a loss function can avoid the problem of the learning rate reduction of the mean square error loss function when the gradient is lowered, and therefore, the efficiency of the model training can be ensured.
  • the deep convolutional neural network model includes at least one fully connected layer, and an initial value of each weight in the predetermined recognition model is from a preset weight range (eg, (0, 1) Weight range) Performing random sampling determination, the probability that the connection weight of the fully connected layer is discarded (Dropout) is set to a first preset value (for example, 0.5), and the weight attenuation coefficient in the cross entropy loss function Set to a second preset value (eg, 0.0005), the potential energy term in the cross entropy loss function is set to a third preset value (eg, 0.9).
  • a preset weight range eg, (0, 1) Weight range
  • the probability that the connection weight of the fully connected layer is discarded is set to a first preset value (for example, 0.5)
  • the weight attenuation coefficient in the cross entropy loss function Set to a second preset value (eg, 0.0005)
  • the potential energy term in the cross entropy loss function is set to a third prese
  • the predetermined scoring function of the recognition model for:
  • O i,j represents the first prediction as i and the second prediction is the number of pictures actually appearing in j
  • O represents a matrix of N*N
  • O i,j represents the matrix element in matrix O
  • N represents the prediction of participation.
  • E i, j represents the number of images in which the first prediction is i and the second prediction is j
  • E is the N* of the desired prediction result.
  • the N matrix, Ei,j represents the matrix elements in the matrix E.
  • Scoring function in this embodiment The recognition accuracy of the predetermined recognition model is detected to ensure that the recognition accuracy of the trained recognition model is maintained at a high level to ensure accurate recognition of the degree of retinopathy of the patient.
  • the retinopathy degree level detection system 10 can be segmented into one or more modules, the one or more modules being stored in the memory 11 and being processed by one or more processors (This embodiment is performed by the processor 12) to complete the present invention.
  • module refers to a series of computer program instructions that are capable of performing a particular function, and are more suitable than the program for describing the execution of the retinopathy level detection system 10 in the electronic device 1.
  • FIG. 3 is a functional block diagram of a preferred embodiment of the retinopathy lesion level detecting system 10 of FIG.
  • the retinopathy degree level detecting system 10 can be divided into The module 01 is identified and the module 02 is determined.
  • the functions or operational steps implemented by the module 01-02 are similar to the above, and are not described in detail herein, by way of example, for example:
  • the identification module 01 is configured to: after receiving the retinopathy picture to be identified, identify the received retinopathy picture by using a predetermined recognition model, and output a recognition result; wherein the predetermined recognition model is pre-determined a convolutional neural network model trained to pre-set a predetermined number of sample images with different levels of retinopathy; and
  • the determining module 02 is configured to determine a retinopathy degree level corresponding to the output recognition result according to the mapping relationship between the predetermined recognition result and the retinopathy degree level.
  • the present invention also provides a computer readable storage medium having stored thereon a program of a detection system for a level of retinopathy, the program of the detection system of the degree of retinopathy level being implemented by a processor Any step of the method of detecting the degree of retinopathy.
  • the specific implementation manner of the computer readable storage medium of the present invention is substantially the same as the specific implementation method of the above-mentioned method for detecting the degree of retinopathy degree, and details are not described herein again.
  • the foregoing embodiment method can be implemented by means of software plus a necessary general hardware platform, and can also be implemented by hardware, but in many cases, the former is A better implementation.
  • the technical solution of the present invention which is essential or contributes to the prior art, may be embodied in the form of a software product stored in a storage medium (such as ROM/RAM, disk,
  • the optical disc includes a number of instructions for causing a terminal device (which may be a cell phone, a computer, a server, an air conditioner, or a network device, etc.) to perform the methods described in various embodiments of the present invention.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

A method for detecting retinopathy degree level, comprising: after a retinopathy image to be identified is received, using a predetermined identification model to identify the received retinopathy image, and outputting an identification result (S10), wherein the predetermined identification model is a convolutional neural network model obtained by training, in advance, a preset number of sample images marked by different retinopathy degree levels; and determining the retinopathy degree level corresponding to the output identification result according to a predetermined mapping relationship between the identification result and the retinopathy degree level (S20). With the described method, it is not necessary to perform complex feature extraction operations on eye images, the process is simplified, corresponding different retinopathy degree levels may be determined according to the identification result, and refined identification of the retinopathy degree levels of a patient may be achieved. Further provided are an electronic device and a computer readable storage medium.

Description

视网膜病变程度等级检测方法、装置及存储介质Method, device and storage medium for detecting degree of retinopathy
本申请基于巴黎公约申明享有2017年5月5日递交的申请号为CN201710312327.6、名称为“视网膜病变程度等级检测系统及方法”的中国专利申请的优先权,该中国专利申请的整体内容以参考的方式结合在本申请中。This application is based on the priority of the Chinese Patent Application entitled "Retinopathy Degree Level Detection System and Method" filed on May 5, 2017, with the application number of CN201710312327.6, which is filed on May 5, 2017. The manner of reference is incorporated in the present application.
技术领域Technical field
本发明涉及计算机技术领域,尤其涉及一种视网膜病变程度等级检测方法、电子装置及计算机可读存储介质。The present invention relates to the field of computer technologies, and in particular, to a method for detecting a degree of retinopathy degree, an electronic device, and a computer readable storage medium.
背景技术Background technique
根据对发达国家超过9300万劳动力人口的调查,糖尿病性视网膜病变是导致眼睛失明的一个首要因素,目前,针对糖尿病性视网膜病变的识别通常需要对眼部图像进行特征提取(例如,眼部血管结构,视神经盘,视网膜中心凹槽等特征的提取),特征提取的算法复杂运行性能差,同时,难以对患者的视网膜病变程度进行精细化识别,识别精度难以达到要求。According to a survey of more than 93 million laborers in developed countries, diabetic retinopathy is a leading cause of blindness in the eye. Currently, the identification of diabetic retinopathy usually requires feature extraction of the ocular image (for example, ocular vascular structure). The extraction of features such as optic disc and retinal center groove, the feature extraction algorithm has poor performance and poor performance. At the same time, it is difficult to refine the degree of retinopathy of patients, and the recognition accuracy is difficult to meet the requirements.
发明内容Summary of the invention
本发明的主要目的在于提供一种视网膜病变程度等级检测方法、电子装置及计算机可读存储介质,旨在简单有效地对患者的视网膜病变程度进行精细化识别。The main object of the present invention is to provide a method for detecting the degree of retinopathy degree, an electronic device and a computer readable storage medium, which are intended to accurately and accurately identify the degree of retinopathy of a patient.
为实现上述目的,本发明提供的一种电子装置,其特征在于,所述装置包括:存储器、处理器,所述存储器上存储有可在所述处理器上运行的视网膜病变程度等级检测系统的程序,所述视网膜病变程度等级检测系统的程序被所述处理器执行时实现如下步骤:In order to achieve the above object, an electronic device according to the present invention includes: a memory, a processor, and a memory of a retinopathy degree level detecting system operable on the processor. The program, when the program of the retinopathy degree level detection system is executed by the processor, implements the following steps:
在收到待识别的视网膜病变图片后,对收到的视网膜病变图片利用预先确定的识别模型进行识别,并输出识别结果;其中,所述预先确定的识别模型为预先通过对标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的卷积神经网络模型;After receiving the picture of the retinopathy to be identified, the received retinopathy picture is identified by using a predetermined recognition model, and the recognition result is output; wherein the predetermined recognition model is pre-marked by different retinopathy a convolutional neural network model obtained by training a predetermined number of sample pictures of a degree level;
根据预先确定的识别结果与视网膜病变程度等级的映射关系,确定输出的识别结果对应的视网膜病变程度等级。According to the mapping relationship between the predetermined recognition result and the degree of retinopathy degree, the degree of retinopathy corresponding to the output recognition result is determined.
优选地,所述预先确定的识别模型的训练过程如下:Preferably, the training process of the predetermined recognition model is as follows:
A、为各个预设的视网膜病变程度等级设定对应的预设数量的样本图片,为每个样本图片标注对应的视网膜病变程度等级;A. setting a corresponding preset number of sample pictures for each preset retinopathy degree level, and marking a corresponding retinopathy degree level for each sample picture;
B、将各张样本图片进行图片预处理以获得待模型训练的训练图片;B. Perform image preprocessing on each sample image to obtain a training picture to be trained by the model;
C、将所有训练图片分为第一比例的训练集和第二比例的验证集;C. Divide all training pictures into a training set of a first ratio and a verification set of a second ratio;
D、利用所述训练集训练所述预先确定的识别模型;D. training the predetermined recognition model by using the training set;
E、利用所述验证集验证训练的识别模型的准确率,若准确率大于或者等 于预设准确率,则训练结束,或者,若准确率小于预设准确率,则增加各个视网膜病变程度等级对应的样本图片数量并重新执行上述步骤B、C、D、E。E. verifying the accuracy of the trained recognition model by using the verification set, if the accuracy is greater than or equal At the preset accuracy rate, the training ends, or if the accuracy rate is less than the preset accuracy rate, the number of sample pictures corresponding to each retinopathy degree level is increased and the above steps B, C, D, E are re-executed.
此外,为实现上述目的,本发明还提供一种视网膜病变程度等级的检测方法,所述方法包括以下步骤:In addition, in order to achieve the above object, the present invention also provides a method for detecting a degree of retinopathy degree, the method comprising the following steps:
在收到待识别的视网膜病变图片后,对收到的视网膜病变图片利用预先确定的识别模型进行识别,并输出识别结果;其中,所述预先确定的识别模型为预先通过对标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的卷积神经网络模型;After receiving the picture of the retinopathy to be identified, the received retinopathy picture is identified by using a predetermined recognition model, and the recognition result is output; wherein the predetermined recognition model is pre-marked by different retinopathy a convolutional neural network model obtained by training a predetermined number of sample pictures of a degree level;
根据预先确定的识别结果与视网膜病变程度等级的映射关系,确定输出的识别结果对应的视网膜病变程度等级。According to the mapping relationship between the predetermined recognition result and the degree of retinopathy degree, the degree of retinopathy corresponding to the output recognition result is determined.
优选地,所述预先确定的识别模型的训练过程如下:Preferably, the training process of the predetermined recognition model is as follows:
A、为各个预设的视网膜病变程度等级设定对应的预设数量的样本图片,为每个样本图片标注对应的视网膜病变程度等级;A. setting a corresponding preset number of sample pictures for each preset retinopathy degree level, and marking a corresponding retinopathy degree level for each sample picture;
B、将各张样本图片进行图片预处理以获得待模型训练的训练图片;B. Perform image preprocessing on each sample image to obtain a training picture to be trained by the model;
C、将所有训练图片分为第一比例的训练集和第二比例的验证集;C. Divide all training pictures into a training set of a first ratio and a verification set of a second ratio;
D、利用所述训练集训练所述预先确定的识别模型;D. training the predetermined recognition model by using the training set;
E、利用所述验证集验证训练的识别模型的准确率,若准确率大于或者等于预设准确率,则训练结束,或者,若准确率小于预设准确率,则增加各个视网膜病变程度等级对应的样本图片数量并重新执行上述步骤B、C、D、E。E. verifying the accuracy of the training recognition model by using the verification set. If the accuracy rate is greater than or equal to the preset accuracy rate, the training ends, or if the accuracy rate is less than the preset accuracy rate, increasing the level of each retinopathy level The number of sample pictures and re-execute steps B, C, D, E above.
另外,为实现上述目的,本发明还提供一种计算机可读存储介质,所述计算机可读存储介质上存储有视网膜病变程度等级的检测系统的程序,所述视网膜病变程度等级的检测系统的程序被处理器执行时实现上述视网膜病变程度等级的检测方法的任一步骤。In addition, in order to achieve the above object, the present invention also provides a computer readable storage medium storing a program of a detection system for a degree of retinopathy degree, a program of a detection system for a degree of retinopathy Any step of the method of detecting the level of retinopathy as described above when executed by the processor.
本发明提出的视网膜病变程度等级检测方法、电子装置及计算机可读存储介质,通过基于标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的深度卷积神经网络模型来对收到的视网膜病变图片进行识别,并根据识别结果确定对应的视网膜病变程度等级。由于只需根据预先训练得到的深度卷积神经网络模型对收到的视网膜病变图片进行识别,无需对眼部图像进行复杂的特征提取运算,更加简单,且能根据识别结果确定对应的不同视网膜病变程度等级,能有效地对患者的视网膜病变程度进行精细化识别。The retinopathy degree level detecting method, the electronic device and the computer readable storage medium proposed by the present invention are received by a deep convolutional neural network model based on a preset number of sample pictures marked with different retinopathy degree levels. The retinopathy picture is identified, and the corresponding retinopathy degree level is determined according to the recognition result. Since only the pre-trained deep convolutional neural network model is needed to recognize the received retinal lesion image, it is not necessary to perform complex feature extraction operation on the eye image, which is simpler and can determine corresponding different retinopathy according to the recognition result. Degree level, can effectively identify the degree of retinopathy of patients.
附图说明DRAWINGS
图1为本发明视网膜病变程度等级的检测方法较佳实施例的流程示意图;1 is a schematic flow chart of a preferred embodiment of a method for detecting a degree of retinopathy of the present invention;
图2为本发明电子装置较佳实施例的示意图;2 is a schematic view of a preferred embodiment of an electronic device of the present invention;
图3为图2中视网膜病变程度等级检测系统较佳实施例的功能模块示意图。3 is a functional block diagram of a preferred embodiment of the retinopathy lesion level detecting system of FIG. 2.
本发明目的的实现、功能特点及优点将结合实施例,参照附图做进一步说明。 The implementation, functional features, and advantages of the present invention will be further described in conjunction with the embodiments.
具体实施方式detailed description
为了使本发明所要解决的技术问题、技术方案及有益效果更加清楚、明白,以下结合附图和实施例,对本发明进行进一步详细说明。应当理解,此处所描述的具体实施例仅仅用以解释本发明,并不用于限定本发明。The present invention will be further described in detail below with reference to the accompanying drawings and embodiments, in order to make the present invention. It is understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
本发明提供一种视网膜病变程度等级的检测方法。The invention provides a method for detecting the degree of retinopathy.
参照图1,图1为本发明视网膜病变程度等级的检测方法较佳实施例的流程示意图。Referring to FIG. 1, FIG. 1 is a schematic flow chart of a preferred embodiment of a method for detecting a degree of retinopathy of the present invention.
在本实施例中,该视网膜病变程度等级的检测方法包括:In this embodiment, the method for detecting the degree of retinopathy is as follows:
步骤S10、在收到待识别的视网膜病变图片后,对收到的视网膜病变图片利用预先确定的识别模型进行识别,并输出识别结果;其中,所述预先确定的识别模型为预先通过对标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的深度卷积神经网络模型。Step S10: After receiving the retinopathy picture to be identified, the received retinopathy picture is identified by using a predetermined recognition model, and the recognition result is output; wherein the predetermined recognition model is pre-marked by A deep convolutional neural network model obtained by training a preset number of sample images of different degrees of retinopathy.
本实施例中,视网膜病变程度等级检测系统接收用户发出的包含待识别的视网膜病变图片的视网膜病变程度等级检测请求,例如,接收用户通过手机、平板电脑、自助终端设备等终端发送的视网膜病变程度等级检测请求,如接收用户在手机、平板电脑、自助终端设备等终端中预先安装的客户端上发送来的视网膜病变程度等级检测请求,或接收用户在手机、平板电脑、自助终端设备等终端中的浏览器系统上发送来的视网膜病变程度等级检测请求。In this embodiment, the retinopathy degree level detecting system receives a retinopathy degree level detection request sent by the user and includes a retinopathy degree image to be recognized, for example, receiving a degree of retinopathy transmitted by the user through a terminal such as a mobile phone, a tablet computer, or a self-service terminal device. A level detection request, such as receiving a retinopathy degree level detection request sent by a user on a pre-installed client in a terminal such as a mobile phone, a tablet computer, a self-service terminal device, or receiving a user in a terminal such as a mobile phone, a tablet computer, or a self-service terminal device. Retinopathy level level detection request sent on the browser system.
视网膜病变程度等级检测系统在收到用户发出的视网膜病变程度等级检测请求后,利用预先训练好的识别模型对收到的待识别的视网膜病变图片进行识别,识别出收到的待识别的视网膜病变图片在识别模型中的识别结果。该识别模型可预先通过对大量标注有不同视网膜病变程度等级的预设数量样本图片进行识别来不断进行训练、学习、验证、优化等,以将其训练成能准确识别出不同视网膜病变程度等级对应的标注的模型。例如,该识别模型可采用深度卷积神经网络模型(Convolutional Neural Network,CNN)模型等。The retinopathy degree level detection system recognizes the received retinopathy image to be recognized by using the pre-trained recognition model after receiving the request for the retinopathy degree level detection issued by the user, and recognizes the retinopathy to be recognized that is to be recognized. The recognition result of the picture in the recognition model. The recognition model can be continuously trained, learned, verified, optimized, etc. by identifying a plurality of preset number of sample pictures marked with different degrees of retinopathy degree, so as to train them to accurately identify different levels of retinopathy. The model of the annotation. For example, the recognition model may employ a Convolutional Neural Network (CNN) model or the like.
步骤S20、根据预先确定的识别结果与视网膜病变程度等级的映射关系,确定输出的识别结果对应的视网膜病变程度等级。Step S20: Determine a retinopathy degree level corresponding to the output recognition result according to a mapping relationship between the predetermined recognition result and the retinopathy degree level.
在利用预先训练好的深度卷积神经网络模型对收到的视网膜病变图片进行识别获取到识别结果后,可根据预先确定的识别结果与视网膜病变程度等级的映射关系,确定输出的识别结果对应的视网膜病变程度等级,确定的视网膜病变程度等级即为收到的视网膜病变图片所对应的视网膜病变程度等级。例如,在一种实施方式中,所述识别结果包括第一识别结果(例如,标注为“0”)、第二识别结果(例如,标注为“1”)、第三识别结果(例如,标注为“2”)、第四识别结果(例如,标注为“3”)及第五识别结果(例如,标注为“4”),所述视网膜病变程度等级包括第一等级、第二等级、第三等级、第四等级及第五等级。可预先确定不同识别结果与视网膜病变程度等级的映射关系,如所述第一识别结果对应第一等级,第二识别结果对应第二等级,第三识别结果对应第三等级,第四识别结果对应第四等级,第五识别结果对应第五等级。 例如,具体地,第一等级可以对应正常和轻度的非增殖性糖尿病视网膜病变,该第一等级对应的视网膜病变图片表现为仅有个别血管瘤,硬性渗出,视网膜出血等。第二等级可以对应无临床意义黄斑水肿的非增殖性糖尿病视网膜病变,该第二等级对应的视网膜病变图片表现有微血管瘤,硬性渗出,视网膜出血,袢状或串珠状静脉。第三等级可以对应有临床意义的黄斑水肿(CSME)的非增殖性糖尿病视网膜病,该第三等级对应的视网膜病变图片表现黄斑区及其附近有视网膜增厚,并有微血管瘤,软性渗出,视网膜出血。第四等级可以对应非高危险期的增生性视网膜病,该第四等级对应的视网膜病变图片表现为视乳头外区有新生血管形成,其他区域内视网膜微血管形成的增殖型改变。第五等级可以对应高危险期的增生性视网膜病,该第五等级对应的视网膜病变图片表现为视乳头区有新生血管形成,玻璃体或视网膜前出血。After obtaining the recognition result by using the pre-trained deep convolutional neural network model to obtain the recognition result, the mapping result of the predetermined recognition result and the retinopathy degree level may be determined according to the output recognition result. The degree of retinopathy is determined by the degree of retinopathy level of the retinopathy image received. For example, in one embodiment, the recognition result includes a first recognition result (eg, labeled "0"), a second recognition result (eg, labeled "1"), and a third recognition result (eg, an annotation a "2"), a fourth recognition result (for example, labeled "3"), and a fifth recognition result (for example, labeled "4"), the retinopathy degree level includes a first level, a second level, and a Three levels, fourth level and fifth level. The mapping relationship between the different recognition results and the retinopathy degree level may be determined in advance, if the first recognition result corresponds to the first level, the second recognition result corresponds to the second level, and the third recognition result corresponds to the third level, and the fourth recognition result corresponds to The fourth level, the fifth recognition result corresponds to the fifth level. For example, specifically, the first level may correspond to normal and mild non-proliferative diabetic retinopathy, and the first level corresponding retinopathy picture appears as only individual hemangiomas, hard exudation, retinal hemorrhage, and the like. The second grade can correspond to non-proliferative diabetic retinopathy without clinically significant macular edema. The second grade corresponding retinopathy picture shows microangioma, hard exudation, retinal hemorrhage, sickle or beaded vein. The third grade can correspond to non-proliferative diabetic retinopathy with clinically significant macular edema (CSME). The third grade corresponding retinopathy picture shows retinal thickening in the macular area and its vicinity, and microangioma, soft infiltration Out, retinal hemorrhage. The fourth grade can correspond to non-high-risk proliferative retinopathy. The fourth-level retinal lesion picture shows neovascularization in the outer area of the optic papilla and proliferative changes in retinal microvessel formation in other areas. The fifth grade can correspond to high-risk proliferative retinopathy. The fifth grade corresponding retinopathy picture shows neovascularization, vitreous or preretinal hemorrhage in the optic papilla area.
这样,在利用预先训练好的识别模型对收到的视网膜病变图片进行识别获取到识别结果后,即可根据获取到的不同识别结果确定对应的不同视网膜病变程度等级,从而实现对多种细化的视网膜病变程度等级的准确识别。In this way, after the received retinopathy image is recognized by the pre-trained recognition model to obtain the recognition result, the corresponding different retinopathy degree levels can be determined according to the obtained different recognition results, thereby realizing various refinements. Accurate identification of the degree of retinopathy.
本实施例通过基于标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的深度卷积神经网络模型来对收到的视网膜病变图片进行识别,并根据识别结果确定对应的视网膜病变程度等级。由于只需根据预先训练得到的深度卷积神经网络模型对收到的视网膜病变图片进行识别,无需对眼部图像进行复杂的特征提取运算,更加简单,且能根据识别结果确定对应的不同视网膜病变程度等级,能有效地对患者的视网膜病变程度进行精细化识别。In this embodiment, the received retinal lesion image is identified by a deep convolutional neural network model based on a preset number of sample images labeled with different degrees of retinopathy degree, and the corresponding degree of retinopathy is determined according to the recognition result. . Since only the pre-trained deep convolutional neural network model is needed to recognize the received retinal lesion image, it is not necessary to perform complex feature extraction operation on the eye image, which is simpler and can determine corresponding different retinopathy according to the recognition result. Degree level, can effectively identify the degree of retinopathy of patients.
进一步地,在其他实施例中,所述预先确定的识别模型的训练过程如下:Further, in other embodiments, the training process of the predetermined recognition model is as follows:
A、为各个预设的视网膜病变程度等级(如第一等级、第二等级、第三等级、第四等级及第五等级,或轻微、轻度、中度、重度等)准备对应的预设数量的样本图片,为每个样本图片标注对应的视网膜病变程度等级;A. Prepare corresponding presets for each preset level of retinopathy (such as first level, second level, third level, fourth level, and fifth level, or slight, mild, moderate, severe, etc.) a number of sample images, each of which is labeled with a corresponding degree of retinopathy;
B、将各张样本图片进行图片预处理以获得待模型训练的训练图片。通过对各张样本图片进行图片预处理如缩放、裁剪、翻转及/或扭曲等操作后才进行模型训练,以有效提高模型训练的真实性及准确率。例如在一种实施方式中,对各张样本图片进行图片预处理可以包括:B. Perform image preprocessing on each sample image to obtain a training picture to be trained by the model. The model training is carried out by performing image preprocessing such as scaling, cropping, flipping and/or twisting on each sample image to effectively improve the authenticity and accuracy of the model training. For example, in an embodiment, performing image preprocessing on each sample picture may include:
将各张样本图片的较短边长缩放到第一预设大小(例如,640像素)以获得对应的第一图片,在各张第一图片上随机裁剪出一个第二预设大小(例如,256*256像素)的第二图片;Scaling a shorter side length of each sample picture to a first preset size (eg, 640 pixels) to obtain a corresponding first picture, and randomly cropping a second preset size on each first picture (eg, a second picture of 256*256 pixels);
根据各个预先确定的预设类型参数(例如,颜色、亮度及/或对比度等)对应的标准参数值(例如,颜色对应的标准参数值为a1,亮度对应的标准参数值为a2,对比度对应的标准参数值为a3),将各张第二图片的各个预先确定的预设类型参数值调整为对应的标准参数值,获得对应的第三图片,以消除作为医学图片的样本图片在拍摄时外界条件导致的图片不清晰,提高模型训 练的有效性;例如,将各张第二图片的亮度值调整为标准参数值a2,将各张第二图片的对比度值调整为标准参数值a3;The standard parameter value corresponding to each predetermined preset type parameter (for example, color, brightness, and/or contrast, etc.) (for example, the standard parameter value corresponding to the color is a1, and the standard parameter value corresponding to the brightness is a2, and the contrast corresponds The standard parameter value is a3), and each predetermined preset type parameter value of each second picture is adjusted to a corresponding standard parameter value, and a corresponding third picture is obtained to eliminate the sample picture as a medical picture at the time of shooting. The picture caused by the condition is not clear, improve the model training The effectiveness of the practice; for example, adjusting the brightness value of each second picture to the standard parameter value a2, adjusting the contrast value of each second picture to the standard parameter value a3;
对各张第三图片进行预设方向(例如,水平和垂直方向)的翻转,及按照预设的扭曲角度(例如,30度)对各张第三图片进行扭曲操作,获得各张第三图片对应的第四图片,各张第四图片即为对应的样本图片的训练图片。其中,翻转和扭曲操作的作用是模拟实际业务场景下各种形式的图片,通过这些翻转和扭曲操作可以增大数据集的规模,从而提高模型训练的真实性和实用性。Performing a preset direction (for example, horizontal and vertical directions) on each of the third pictures, and performing a twist operation on each of the third pictures according to a preset twist angle (for example, 30 degrees) to obtain each third picture. The corresponding fourth picture, each fourth picture is a training picture of the corresponding sample picture. Among them, the function of the flip and twist operation is to simulate various forms of pictures in the actual business scene. Through these flip and twist operations, the size of the data set can be increased, thereby improving the authenticity and practicability of the model training.
C、将所有训练图片分为第一比例(例如,50%)的训练集、第二比例(例如,25%)的验证集;C. Divide all training pictures into a training set of a first ratio (for example, 50%) and a verification set of a second ratio (for example, 25%);
D、利用所述训练集训练所述预先确定的识别模型;D. training the predetermined recognition model by using the training set;
E、利用所述验证集验证训练的识别模型的准确率,若准确率大于或者等于预设准确率,则训练结束,或者,若准确率小于预设准确率,则增加各个视网膜病变程度等级对应的样本图片数量并重新执行上述步骤B、C、D、E,直至训练的识别模型的准确率大于或者等于预设准确率。E. verifying the accuracy of the training recognition model by using the verification set. If the accuracy rate is greater than or equal to the preset accuracy rate, the training ends, or if the accuracy rate is less than the preset accuracy rate, increasing the level of each retinopathy level The number of sample pictures and the above steps B, C, D, E are re-executed until the accuracy of the trained recognition model is greater than or equal to the preset accuracy.
进一步地,在其他实施例中,所述预先确定的识别模型即深度卷积神经网络模型包括输入层和多个网络层,所述网络层包括卷积层、池化层、全连接层及分类器层,可选的,深度卷积神经网络模型还可以包括具有随机丢弃某些连接权重机制的网络层(即Dropout层),该网络层的作用是提升模型的识别精度。Further, in other embodiments, the predetermined recognition model, that is, the deep convolutional neural network model, includes an input layer and a plurality of network layers, and the network layer includes a convolution layer, a pooling layer, a fully connected layer, and a classification. The layer, optionally, the deep convolutional neural network model may also include a network layer (ie, a Dropout layer) having a random drop of some connection weighting mechanism, the role of which is to improve the recognition accuracy of the model.
在一种具体的实施方式中,所述深度卷积神经网络模型由1个输入层,11个卷积层,5个池化层,1个具有随机丢弃某些连接权重机制的网络层(即Dropout层),1个全连接层,1个分类器层构成。该深度卷积神经网络模型的详细结构如下表1所示:In a specific embodiment, the deep convolutional neural network model consists of one input layer, eleven convolutional layers, five pooling layers, and one network layer having random connection discarding certain connection weighting mechanisms (ie, Dropout layer), 1 fully connected layer, 1 classifier layer. The detailed structure of the deep convolutional neural network model is shown in Table 1 below:
Layer NameLayer Name Batch SizeBatch Size Kernel SizeKernel Size Stride SizeStride Size Output SizeOutput Size
InputInput 6464      
Conv1Conv1 6464 77 22 112112
MaxPool1MaxPool1 6464 33 22 5656
Conv2Conv2 192192 33 11 5656
Maxpool2Maxpool2 192192 22 22 2828
Convolution3Convolution3 256256 33 22 2828
Convolution4Convolution4 480480 33 22 2828
Maxpool3Maxpool3 480480 22 22 2828
Convolution5Convolution5 512512 33 22 1414
Convolution6Convolution6 512512 33 22 1414
Convolution7Convolution7 512512 33 22 1414
Convolution8Convolution8 512512 33 22 1414
Convolution9Convolution9 512512 33 22 1414
Maxpool4Maxpool4 832832 22 22 77
Convolution10Convolution10 832832 33 22 77
Convolution11Convolution11 10241024 33 22 77
Avgpool5Avgpool5 10241024 77 11 11
DropoutDropout 10241024     11
Fc1Fc1 55     11
SoftmaxSoftmax 55     11
表1Table 1
其中:Layer Name表示网络层的名称,Input表示网络的数据输入层,Conv表示模型的卷积层,Conv1表示模型的第1个卷积层,MaxPool表示模型的最大值池化层,MaxPool1表示第一个基于最大值池化层,Dropout表示具有随机丢弃某些连接权重机制的网络层,Avgpool5表示第5个池化层但采用取均值方式进行池化,Fc表示模型中的全连接层,Fc1表示第1个全连接层,Softmax表示Softmax分类器层;Batch Size表示当前层的输入图像数目;Kernel Size表示当前层卷积核的尺度(例如,Kernel Size可以等于3,表示卷积核的尺度为3x3);Stride Size表示卷积核的移动步长,即做完一次卷积之后移动到下一个卷积位置的距离;Output Size表示网络层输出特征映射的尺寸。需要说明的是,本实施例中池化层的池化方式包括但不限于Mean pooling(均值采样)、Max pooling(最大值采样)、Overlapping(重叠采样)、L2 pooling(均方采样)、Local Contrast Normalization(归一化采样)、Stochasticpooling(随即采样)、Def-pooling(形变约束采样)等等。Among them: Layer Name indicates the name of the network layer, Input indicates the data input layer of the network, Conv indicates the convolution layer of the model, Conv1 indicates the first convolution layer of the model, MaxPool indicates the maximum pooling layer of the model, and MaxPool1 indicates the first One based on the maximum pooling layer, Dropout means a network layer with random discarding of some connection weighting mechanism, Avgpool5 means the fifth pooling layer but pooled by means of averaging, Fc represents the fully connected layer in the model, Fc1 Represents the first fully connected layer, Softmax represents the Softmax classifier layer; Batch Size represents the number of input images of the current layer; Kernel Size represents the scale of the current layer convolution kernel (for example, Kernel Size can be equal to 3, indicating the scale of the convolution kernel) 3x3); Stride Size represents the moving step size of the convolution kernel, that is, the distance moved to the next convolution position after one convolution; Output Size represents the size of the network layer output feature map. It should be noted that the pooling mode of the pooling layer in this embodiment includes, but is not limited to, Mean pooling, Max pooling, Overlapping, L2 pooling, Local. Contrast Normalization, Stochasticpooling, Def-pooling, and more.
进一步地,在其他实施例中,为了提高模型的识别精度,各个所述网络层(例如,卷积层、池化层、具有随机丢弃某些连接权重机制的网络层、全连接层及分类器层等)对应的激活函数f(x)为:Further, in other embodiments, in order to improve the recognition accuracy of the model, each of the network layers (for example, a convolution layer, a pooling layer, a network layer with a random drop of some connection weight mechanism, a fully connected layer, and a classifier) The corresponding activation function f(x) of the layer, etc. is:
f(x)=max(α*x,0)f(x)=max(α*x,0)
其中,α为泄漏率,x表示该深度卷积神经网络模型中神经元的一个数值输入。在本实施例的一个优选实施方式中,将α设定为0.5。经过相同测试数据集的对比测试,相较于其他现有的激活函数,通过本实施例的激活函数f(x),该深度卷积神经网络模型的识别准确率大约有3%的提升。Where α is the leak rate and x is a numerical input of the neurons in the deep convolutional neural network model. In a preferred embodiment of the embodiment, α is set to 0.5. Through the comparison test of the same test data set, the recognition accuracy of the deep convolutional neural network model is improved by about 3% by the activation function f(x) of the present embodiment compared to other existing activation functions.
进一步地,在其他实施例中,为了提高模型的识别精度,各个所述网络层(例如,卷积层、池化层、具有随机丢弃某些连接权重机制的网络层、全连接层及分类器层等)对应的交叉熵H(P,Q)为:Further, in other embodiments, in order to improve the recognition accuracy of the model, each of the network layers (for example, a convolution layer, a pooling layer, a network layer with a random drop of some connection weight mechanism, a fully connected layer, and a classifier) The corresponding cross entropy H(P, Q) of the layer, etc. is:
H(P,Q)=H(P)+DKL(P||Q)H(P,Q)=H(P)+D KL (P||Q)
其中,P,Q为两个概率分布,H(P)为概率分布P的期望,H(P)=-∑x∈XP(x)log P(x),x为概率分布P的样本空间X中任意一个样本,P(x)表示样本x被选取的概率;DKL(P||Q)的表达式为
Figure PCTCN2017100044-appb-000001
x为概率 分布P和Q公共样本空间X中任意一个样本,P(x)表示样本x在概率分布P上被选取的概率,Q(x)表示样本x在概率分布Q上被选取的概率。
Where P and Q are two probability distributions, H(P) is the expectation of probability distribution P, H(P)=-∑ x∈X P(x)log P(x), and x is the sample space of probability distribution P For any sample in X, P(x) represents the probability that sample x is selected; the expression of D KL (P||Q) is
Figure PCTCN2017100044-appb-000001
x is the probability distribution P and Q any sample in the common sample space X, P(x) represents the probability that the sample x is selected on the probability distribution P, and Q(x) represents the probability that the sample x is selected on the probability distribution Q.
进一步地,为了保证模型训练的效率和准确性,各个所述网络层对应的交叉熵损失函数
Figure PCTCN2017100044-appb-000002
为:
Further, in order to ensure the efficiency and accuracy of the model training, the cross entropy loss function corresponding to each of the network layers
Figure PCTCN2017100044-appb-000002
for:
Figure PCTCN2017100044-appb-000003
Figure PCTCN2017100044-appb-000003
其中,x表示模型的输入,
Figure PCTCN2017100044-appb-000004
表示输入对应的标签,W表示预设的模型参数,X表示模型输入空间,f(x:W)表示模型对输入x的做了变换后的输出,ζ表示规约化因子,||W||2表示对矩阵元素求和:
Where x is the input to the model,
Figure PCTCN2017100044-appb-000004
Indicates the corresponding label of the input, W represents the preset model parameters, X represents the model input space, f(x:W) represents the transformed output of the model to the input x, ζ denotes the statistic factor, ||W|| 2 means summing matrix elements:
Figure PCTCN2017100044-appb-000005
Figure PCTCN2017100044-appb-000005
Wi+1=Wi+ΔWi+1 W i+1 =W i +ΔW i+1
其中,ΔWi+1表示在i+1时刻权值矩阵的更新增量,α为势能项,β为权值衰减系数,γ为模型的学习率,Wi表示在i时刻权值矩阵状态值,Di表示第i批输入,
Figure PCTCN2017100044-appb-000006
表示第i批输入对应的平均梯度。
Where ΔW i+1 represents the update increment of the weight matrix at time i+1, α is the potential energy term, β is the weight attenuation coefficient, γ is the learning rate of the model, and W i is the state value of the weight matrix at time i , D i represents the i-th batch input,
Figure PCTCN2017100044-appb-000006
Indicates the average gradient corresponding to the i-th batch input.
本实施例中,交叉熵可在神经网络(机器学习)中作为损失函数,例如,P表示真实标记的分布,Q则为训练后的模型的预测标记分布,交叉熵损失函数可以衡量P与Q的相似性,以保证模型训练的准确性。而且,交叉熵作为损失函数在梯度下降时能避免均方误差损失函数学习速率降低的问题,因此,能保证模型训练的效率。In this embodiment, cross entropy can be used as a loss function in neural networks (machine learning). For example, P represents the distribution of real markers, Q is the predicted marker distribution of the trained model, and the cross entropy loss function can measure P and Q. The similarity to ensure the accuracy of the model training. Moreover, the cross entropy as a loss function can avoid the problem of the learning rate reduction of the mean square error loss function when the gradient is lowered, and therefore, the efficiency of the model training can be ensured.
进一步地,在其他实施例中,所述深度卷积神经网络模型包括至少一个全连接层,所述预先确定的识别模型中的各权重的初始值从预设的权重范围(例如,(0,1)权重范围)进行随机采样确定,所述全连接层的连接权重被丢弃(Dropout)的概率设置为第一预设值(例如,0.5),所述交叉熵损失函数中的权值衰减系数设置为第二预设值(例如,0.0005),所述交叉熵损失函数中的势能项设置为第三预设值(例如,0.9)。Further, in other embodiments, the deep convolutional neural network model includes at least one fully connected layer, and an initial value of each weight in the predetermined recognition model is from a preset weight range (eg, (0, 1) Weight range) Performing random sampling determination, the probability that the connection weight of the fully connected layer is discarded (Dropout) is set to a first preset value (for example, 0.5), and the weight attenuation coefficient in the cross entropy loss function Set to a second preset value (eg, 0.0005), the potential energy term in the cross entropy loss function is set to a third preset value (eg, 0.9).
进一步地,在其他实施例中,所述预先确定的识别模型的打分函数
Figure PCTCN2017100044-appb-000007
为:
Further, in other embodiments, the predetermined scoring function of the recognition model
Figure PCTCN2017100044-appb-000007
for:
Figure PCTCN2017100044-appb-000008
Figure PCTCN2017100044-appb-000008
其中,
Figure PCTCN2017100044-appb-000009
Oi,j表示第一次预测为i并且第二次预测为j实际出现的图片数目,O表示一个N*N的矩阵,Oi,j代表矩阵O中的矩阵元素,N表示参与预测的图片数目,预测结果i,j∈{0 1 2 3 4},Ei,j表示第一次预测为i并 且第二次预测为j应该出现的图像数目,E是期望的预测结果的N*N矩阵,Ei,j代表矩阵E中的矩阵元素。
among them,
Figure PCTCN2017100044-appb-000009
O i,j represents the first prediction as i and the second prediction is the number of pictures actually appearing in j, O represents a matrix of N*N, O i,j represents the matrix element in matrix O, and N represents the prediction of participation. The number of pictures, the prediction result i, j ∈ {0 1 2 3 4}, E i, j represents the number of images in which the first prediction is i and the second prediction is j, and E is the N* of the desired prediction result. The N matrix, E i,j represents the matrix elements in the matrix E.
本实施例中通过打分函数
Figure PCTCN2017100044-appb-000010
来检测所述预先确定的识别模型的识别准确率,以保证训练出的所述预先确定的识别模型的识别准确率保持在较高水平,以保证对患者的视网膜病变程度进行准确地识别。
Scoring function in this embodiment
Figure PCTCN2017100044-appb-000010
The recognition accuracy of the predetermined recognition model is detected to ensure that the recognition accuracy of the trained recognition model is maintained at a high level to ensure accurate recognition of the degree of retinopathy of the patient.
本发明进一步提供一种电子装置。请参阅图2,是本发明电子装置较佳实施例的示意图。The invention further provides an electronic device. Please refer to FIG. 2, which is a schematic diagram of a preferred embodiment of the electronic device of the present invention.
在本实施例中,所述视网膜病变程度等级检测方法应用于一种电子装置1中。该电子装置1可包括,但不仅限于,存储器11、处理器12及显示器13。图2仅示出了具有组件11-13的电子装置1,但是应理解的是,并不要求实施所有示出的组件,可以替代的实施更多或者更少的组件。In the present embodiment, the retinopathy degree level detecting method is applied to an electronic device 1. The electronic device 1 may include, but is not limited to, a memory 11, a processor 12, and a display 13. Figure 2 shows only the electronic device 1 with components 11-13, but it should be understood that not all illustrated components may be implemented, and more or fewer components may be implemented instead.
所述存储器11在一些实施例中可以是所述电子装置1的内部存储单元,例如该电子装置1的硬盘或内存。所述存储器11在另一些实施例中也可以是所述电子装置1的外部存储设备,例如所述电子装置1上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,所述存储器11还可以既包括所述电子装置1的内部存储单元也包括外部存储设备。所述存储器11用于存储安装于所述电子装置1的应用软件及各类数据,例如所述视网膜病变程度等级检测系统10的程序代码等。所述存储器11还可以用于暂时地存储已经输出或者将要输出的数据。The memory 11 may in some embodiments be an internal storage unit of the electronic device 1, such as a hard disk or memory of the electronic device 1. The memory 11 may also be an external storage device of the electronic device 1 in other embodiments, such as a plug-in hard disk equipped on the electronic device 1, a smart memory card (SMC), and a secure digital device. (Secure Digital, SD) card, flash card, etc. Further, the memory 11 may also include both an internal storage unit of the electronic device 1 and an external storage device. The memory 11 is configured to store application software and various types of data installed in the electronic device 1, such as program codes of the retinopathy degree level detecting system 10, and the like. The memory 11 can also be used to temporarily store data that has been output or is about to be output.
所述处理器12在一些实施例中可以是一中央处理器(Central Processing Unit,CPU),微处理器或其他数据处理芯片,用于运行所述存储器11中存储的程序代码或处理数据,例如执行所述视网膜病变程度等级检测系统10等。The processor 12, in some embodiments, may be a central processing unit (CPU), a microprocessor or other data processing chip for running program code or processing data stored in the memory 11, for example The retinopathy degree level detecting system 10 and the like are executed.
所述显示器13在一些实施例中可以是LED显示器、液晶显示器、触控式液晶显示器以及OLED(Organic Light-Emitting Diode,有机发光二极管)触摸器等。所述显示器13用于显示在所述电子装置1中处理的信息以及用于显示可视化的用户界面,例如应用菜单界面、应用图标界面等。所述电子装置1的部件11-13通过系统总线相互通信。The display 13 in some embodiments may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode) touch sensor, or the like. The display 13 is used to display information processed in the electronic device 1 and a user interface for displaying visualization, such as an application menu interface, an application icon interface, and the like. The components 11-13 of the electronic device 1 communicate with one another via a system bus.
在图2所示的装置实施例中,存储器11中存储有视网膜病变程度等级检测系统10的程序;处理器12执行存储器11中存储的视网膜病变程度等级检测系统10的程序时实现如下步骤:In the apparatus embodiment shown in FIG. 2, the program of the retinopathy degree level detecting system 10 is stored in the memory 11; when the processor 12 executes the program of the retinopathy degree level detecting system 10 stored in the memory 11, the following steps are implemented:
在收到待识别的视网膜病变图片后,对收到的视网膜病变图片利用预先确定的识别模型进行识别,并输出识别结果;其中,所述预先确定的识别模型为预先通过对标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的深度卷积神经网络模型;及After receiving the picture of the retinopathy to be identified, the received retinopathy picture is identified by using a predetermined recognition model, and the recognition result is output; wherein the predetermined recognition model is pre-marked by different retinopathy a deep convolutional neural network model obtained by training a predetermined number of sample pictures of a degree level; and
根据预先确定的识别结果与视网膜病变程度等级的映射关系,确定输出的识别结果对应的视网膜病变程度等级。 According to the mapping relationship between the predetermined recognition result and the degree of retinopathy degree, the degree of retinopathy corresponding to the output recognition result is determined.
本实施例中,视网膜病变程度等级检测系统接收用户发出的包含待识别的视网膜病变图片的视网膜病变程度等级检测请求,例如,接收用户通过手机、平板电脑、自助终端设备等终端发送的视网膜病变程度等级检测请求,如接收用户在手机、平板电脑、自助终端设备等终端中预先安装的客户端上发送来的视网膜病变程度等级检测请求,或接收用户在手机、平板电脑、自助终端设备等终端中的浏览器系统上发送来的视网膜病变程度等级检测请求。In this embodiment, the retinopathy degree level detecting system receives a retinopathy degree level detection request sent by the user and includes a retinopathy degree image to be recognized, for example, receiving a degree of retinopathy transmitted by the user through a terminal such as a mobile phone, a tablet computer, or a self-service terminal device. A level detection request, such as receiving a retinopathy degree level detection request sent by a user on a pre-installed client in a terminal such as a mobile phone, a tablet computer, a self-service terminal device, or receiving a user in a terminal such as a mobile phone, a tablet computer, or a self-service terminal device. Retinopathy level level detection request sent on the browser system.
视网膜病变程度等级检测系统在收到用户发出的视网膜病变程度等级检测请求后,利用预先训练好的识别模型对收到的待识别的视网膜病变图片进行识别,识别出收到的待识别的视网膜病变图片在识别模型中的识别结果。该识别模型可预先通过对大量标注有不同视网膜病变程度等级的预设数量样本图片进行识别来不断进行训练、学习、验证、优化等,以将其训练成能准确识别出不同视网膜病变程度等级对应的标注的模型。例如,该识别模型可采用深度卷积神经网络模型(Convolutional Neural Network,CNN)模型等。The retinopathy degree level detection system recognizes the received retinopathy image to be recognized by using the pre-trained recognition model after receiving the request for the retinopathy degree level detection issued by the user, and recognizes the retinopathy to be recognized that is to be recognized. The recognition result of the picture in the recognition model. The recognition model can be continuously trained, learned, verified, optimized, etc. by identifying a plurality of preset number of sample pictures marked with different degrees of retinopathy degree, so as to train them to accurately identify different levels of retinopathy. The model of the annotation. For example, the recognition model may employ a Convolutional Neural Network (CNN) model or the like.
在利用预先训练好的深度卷积神经网络模型对收到的视网膜病变图片进行识别获取到识别结果后,可根据预先确定的识别结果与视网膜病变程度等级的映射关系,确定输出的识别结果对应的视网膜病变程度等级,确定的视网膜病变程度等级即为收到的视网膜病变图片所对应的视网膜病变程度等级。例如,在一种实施方式中,所述识别结果包括第一识别结果(例如,标注为“0”)、第二识别结果(例如,标注为“1”)、第三识别结果(例如,标注为“2”)、第四识别结果(例如,标注为“3”)、及第五识别结果(例如,标注为“4”),所述视网膜病变程度等级包括第一等级、第二等级、第三等级、第四等级及第五等级。可预先确定不同识别结果与视网膜病变程度等级的映射关系,如所述第一识别结果对应第一等级,第二识别结果对应第二等级,第三识别结果对应第三等级,第四识别结果对应第四等级,第五识别结果对应第五等级。例如,具体地,第一等级可以对应正常和轻度的非增殖性糖尿病视网膜病变,该第一等级对应的视网膜病变图片表现为仅有个别血管瘤,硬性渗出,视网膜出血等。第二等级可以对应无临床意义黄斑水肿的非增殖性糖尿病视网膜病变,该第二等级对应的视网膜病变图片表现有微血管瘤,硬性渗出,视网膜出血,袢状或串珠状静脉。第三等级可以对应有临床意义的黄斑水肿(CSME)的非增殖性糖尿病视网膜病,该第三等级对应的视网膜病变图片表现黄斑区及其附近有视网膜增厚,并有微血管瘤,软性渗出,视网膜出血。第四等级可以对应非高危险期的增生性视网膜病,该第四等级对应的视网膜病变图片表现为视乳头外区有新生血管形成,其他区域内视网膜微血管形成的增殖型改变。第五等级可以对应高危险期的增生性视网膜病,该第五等级对应的视网膜病变图片表现为视乳头区有新生血管形成,玻璃体或视网膜前出血。After obtaining the recognition result by using the pre-trained deep convolutional neural network model to obtain the recognition result, the mapping result of the predetermined recognition result and the retinopathy degree level may be determined according to the output recognition result. The degree of retinopathy is determined by the degree of retinopathy level of the retinopathy image received. For example, in one embodiment, the recognition result includes a first recognition result (eg, labeled "0"), a second recognition result (eg, labeled "1"), and a third recognition result (eg, an annotation a "2"), a fourth recognition result (for example, labeled "3"), and a fifth recognition result (for example, labeled "4"), the retinopathy degree level includes a first level, a second level, The third level, the fourth level, and the fifth level. The mapping relationship between the different recognition results and the retinopathy degree level may be determined in advance, if the first recognition result corresponds to the first level, the second recognition result corresponds to the second level, and the third recognition result corresponds to the third level, and the fourth recognition result corresponds to The fourth level, the fifth recognition result corresponds to the fifth level. For example, specifically, the first level may correspond to normal and mild non-proliferative diabetic retinopathy, and the first level corresponding retinopathy picture appears as only individual hemangiomas, hard exudation, retinal hemorrhage, and the like. The second grade can correspond to non-proliferative diabetic retinopathy without clinically significant macular edema. The second grade corresponding retinopathy picture shows microangioma, hard exudation, retinal hemorrhage, sickle or beaded vein. The third grade can correspond to non-proliferative diabetic retinopathy with clinically significant macular edema (CSME). The third grade corresponding retinopathy picture shows retinal thickening in the macular area and its vicinity, and microangioma, soft infiltration Out, retinal hemorrhage. The fourth grade can correspond to non-high-risk proliferative retinopathy. The fourth-level retinal lesion picture shows neovascularization in the outer area of the optic papilla and proliferative changes in retinal microvessel formation in other areas. The fifth grade can correspond to high-risk proliferative retinopathy. The fifth grade corresponding retinopathy picture shows neovascularization, vitreous or preretinal hemorrhage in the optic papilla area.
这样,在利用预先训练好的识别模型对收到的视网膜病变图片进行识别获取到识别结果后,即可根据获取到的不同识别结果确定对应的不同视网膜 病变程度等级,从而实现对多种细化的视网膜病变程度等级的准确识别。In this way, after the received retinopathy image is recognized by using the pre-trained recognition model to obtain the recognition result, the corresponding different retinas can be determined according to the obtained different recognition results. The degree of lesion severity, thus enabling accurate identification of a variety of refinement levels of retinopathy.
本实施例通过基于标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的深度卷积神经网络模型来对收到的视网膜病变图片进行识别,并根据识别结果确定对应的视网膜病变程度等级。由于只需根据预先训练得到的深度卷积神经网络模型对收到的视网膜病变图片进行识别,无需对眼部图像进行复杂的特征提取运算,更加简单,且能根据识别结果确定对应的不同视网膜病变程度等级,能有效地对患者的视网膜病变程度进行精细化识别。In this embodiment, the received retinal lesion image is identified by a deep convolutional neural network model based on a preset number of sample images labeled with different degrees of retinopathy degree, and the corresponding degree of retinopathy is determined according to the recognition result. . Since only the pre-trained deep convolutional neural network model is needed to recognize the received retinal lesion image, it is not necessary to perform complex feature extraction operation on the eye image, which is simpler and can determine corresponding different retinopathy according to the recognition result. Degree level, can effectively identify the degree of retinopathy of patients.
进一步地,在其他实施例中,所述预先确定的识别模型的训练过程如下:Further, in other embodiments, the training process of the predetermined recognition model is as follows:
A、为各个预设的视网膜病变程度等级(如第一等级、第二等级、第三等级、第四等级及第五等级,或轻微、轻度、中度、重度等)准备对应的预设数量的样本图片,为每个样本图片标注对应的视网膜病变程度等级;A. Prepare corresponding presets for each preset level of retinopathy (such as first level, second level, third level, fourth level, and fifth level, or slight, mild, moderate, severe, etc.) a number of sample images, each of which is labeled with a corresponding degree of retinopathy;
B、将各张样本图片进行图片预处理以获得待模型训练的训练图片。通过对各张样本图片进行图片预处理如缩放、裁剪、翻转及/或扭曲等操作后才进行模型训练,以有效提高模型训练的真实性及准确率。例如在一种实施方式中,对各张样本图片进行图片预处理可以包括:B. Perform image preprocessing on each sample image to obtain a training picture to be trained by the model. The model training is carried out by performing image preprocessing such as scaling, cropping, flipping and/or twisting on each sample image to effectively improve the authenticity and accuracy of the model training. For example, in an embodiment, performing image preprocessing on each sample picture may include:
将各张样本图片的较短边长缩放到第一预设大小(例如,640像素)以获得对应的第一图片,在各张第一图片上随机裁剪出一个第二预设大小(例如,256*256像素)的第二图片;Scaling a shorter side length of each sample picture to a first preset size (eg, 640 pixels) to obtain a corresponding first picture, and randomly cropping a second preset size on each first picture (eg, a second picture of 256*256 pixels);
根据各个预先确定的预设类型参数(例如,颜色、亮度及/或对比度等)对应的标准参数值(例如,颜色对应的标准参数值为a1,亮度对应的标准参数值为a2,对比度对应的标准参数值为a3),将各张第二图片的各个预先确定的预设类型参数值调整为对应的标准参数值,获得对应的第三图片,以消除作为医学图片的样本图片在拍摄时外界条件导致的图片不清晰,提高模型训练的有效性;例如,将各张第二图片的亮度值调整为标准参数值a2,将各张第二图片的对比度值调整为标准参数值a3;The standard parameter value corresponding to each predetermined preset type parameter (for example, color, brightness, and/or contrast, etc.) (for example, the standard parameter value corresponding to the color is a1, and the standard parameter value corresponding to the brightness is a2, and the contrast corresponds The standard parameter value is a3), and each predetermined preset type parameter value of each second picture is adjusted to a corresponding standard parameter value, and a corresponding third picture is obtained to eliminate the sample picture as a medical picture at the time of shooting. The picture caused by the condition is not clear, and the effectiveness of the model training is improved; for example, the brightness value of each second picture is adjusted to the standard parameter value a2, and the contrast value of each second picture is adjusted to the standard parameter value a3;
对各张第三图片进行预设方向(例如,水平和垂直方向)的翻转,及按照预设的扭曲角度(例如,30度)对各张第三图片进行扭曲操作,获得各张第三图片对应的第四图片,各张第四图片即为对应的样本图片的训练图片。其中,翻转和扭曲操作的作用是模拟实际业务场景下各种形式的图片,通过这些翻转和扭曲操作可以增大数据集的规模,从而提高模型训练的真实性和实用性。Performing a preset direction (for example, horizontal and vertical directions) on each of the third pictures, and performing a twist operation on each of the third pictures according to a preset twist angle (for example, 30 degrees) to obtain each third picture. The corresponding fourth picture, each fourth picture is a training picture of the corresponding sample picture. Among them, the function of the flip and twist operation is to simulate various forms of pictures in the actual business scene. Through these flip and twist operations, the size of the data set can be increased, thereby improving the authenticity and practicability of the model training.
C、将所有训练图片分为第一比例(例如,50%)的训练集、第二比例(例如,25%)的验证集;C. Divide all training pictures into a training set of a first ratio (for example, 50%) and a verification set of a second ratio (for example, 25%);
D、利用所述训练集训练所述预先确定的识别模型;D. training the predetermined recognition model by using the training set;
E、利用所述验证集验证训练的识别模型的准确率,若准确率大于或者等于预设准确率,则训练结束,或者,若准确率小于预设准确率,则增加各个 视网膜病变程度等级对应的样本图片数量并重新执行上述步骤B、C、D、E,直至训练的识别模型的准确率大于或者等于预设准确率。E. Using the verification set to verify the accuracy of the training recognition model, if the accuracy rate is greater than or equal to the preset accuracy rate, the training ends, or if the accuracy rate is less than the preset accuracy rate, increase each The number of sample pictures corresponding to the level of retinopathy is re-executed in steps B, C, D, and E until the accuracy of the trained recognition model is greater than or equal to the preset accuracy.
进一步地,在其他实施例中,所述预先确定的识别模型即深度卷积神经网络模型包括输入层和多个网络层,所述网络层包括卷积层、池化层、全连接层及分类器层,可选的,深度卷积神经网络模型还可以包括具有随机丢弃某些连接权重机制的网络层(即Dropout层),该网络层的作用是提升模型的识别精度。Further, in other embodiments, the predetermined recognition model, that is, the deep convolutional neural network model, includes an input layer and a plurality of network layers, and the network layer includes a convolution layer, a pooling layer, a fully connected layer, and a classification. The layer, optionally, the deep convolutional neural network model may also include a network layer (ie, a Dropout layer) having a random drop of some connection weighting mechanism, the role of which is to improve the recognition accuracy of the model.
在一种具体的实施方式中,所述深度卷积神经网络模型由1个输入层,11个卷积层,5个池化层,1个具有随机丢弃某些连接权重机制的网络层(即Dropout层),1个全连接层,1个分类器层构成。该深度卷积神经网络模型的详细结构如下表1所示:In a specific embodiment, the deep convolutional neural network model consists of one input layer, eleven convolutional layers, five pooling layers, and one network layer having random connection discarding certain connection weighting mechanisms (ie, Dropout layer), 1 fully connected layer, 1 classifier layer. The detailed structure of the deep convolutional neural network model is shown in Table 1 below:
Layer NameLayer Name Batch SizeBatch Size Kernel SizeKernel Size Stride SizeStride Size Output SizeOutput Size
InputInput 6464      
Conv1Conv1 6464 77 22 112112
MaxPool1MaxPool1 6464 33 22 5656
Conv2Conv2 192192 33 11 5656
Maxpool2Maxpool2 192192 22 22 2828
Convolution3Convolution3 256256 33 22 2828
Convolution4Convolution4 480480 33 22 2828
Maxpool3Maxpool3 480480 22 22 2828
Convolution5Convolution5 512512 33 22 1414
Convolution6Convolution6 512512 33 22 1414
Convolution7Convolution7 512512 33 22 1414
Convolution8Convolution8 512512 33 22 1414
Convolution9Convolution9 512512 33 22 1414
Maxpool4Maxpool4 832832 22 22 77
Convolution10Convolution10 832832 33 22 77
Convolution11Convolution11 10241024 33 22 77
Avgpool5Avgpool5 10241024 77 11 11
DropoutDropout 10241024     11
Fc1Fc1 55     11
SoftmaxSoftmax 55     11
表1Table 1
其中:Layer Name表示网络层的名称,Input表示网络的数据输入层,Conv表示模型的卷积层,Conv1表示模型的第1个卷积层,MaxPool表示模型的最大值池化层,MaxPool1表示第一个基于最大值池化层,Dropout表示具有随机丢弃某些连接权重机制的网络层,Avgpool5表示第5个池化层但采用取均值方式进行池化,Fc表示模型中的全连接层,Fc1表示第1个全连接层,Softmax表示Softmax分类器层;Batch Size表示当前层的输入图像数目;Kernel Size表示当前层卷积核的尺度(例如,Kernel Size可以等于3,表示卷积核的尺度 为3x3);Stride Size表示卷积核的移动步长,即做完一次卷积之后移动到下一个卷积位置的距离;Output Size表示网络层输出特征映射的尺寸。需要说明的是,本实施例中池化层的池化方式包括但不限于Mean pooling(均值采样)、Max pooling(最大值采样)、Overlapping(重叠采样)、L2pooling(均方采样)、Local Contrast Normalization(归一化采样)、Stochasticpooling(随即采样)、Def-pooling(形变约束采样)等等。Among them: Layer Name indicates the name of the network layer, Input indicates the data input layer of the network, Conv indicates the convolution layer of the model, Conv1 indicates the first convolution layer of the model, MaxPool indicates the maximum pooling layer of the model, and MaxPool1 indicates the first One based on the maximum pooling layer, Dropout means a network layer with random discarding of some connection weighting mechanism, Avgpool5 means the fifth pooling layer but pooled by means of averaging, Fc represents the fully connected layer in the model, Fc1 Represents the first fully connected layer, Softmax represents the Softmax classifier layer; Batch Size represents the number of input images of the current layer; Kernel Size represents the scale of the current layer convolution kernel (for example, Kernel Size can be equal to 3, indicating the scale of the convolution kernel) 3x3); Stride Size represents the moving step size of the convolution kernel, that is, the distance moved to the next convolution position after one convolution; Output Size represents the size of the network layer output feature map. It should be noted that the pooling mode of the pooling layer in this embodiment includes, but is not limited to, Mean pooling, Max pooling, Overlapping, L2pooling, Local Contrast. Normalization, Stochasticpooling, Def-pooling, and more.
进一步地,在其他实施例中,为了提高模型的识别精度,各个所述网络层(例如,卷积层、池化层、具有随机丢弃某些连接权重机制的网络层、全连接层及分类器层等)对应的激活函数f(x)为:Further, in other embodiments, in order to improve the recognition accuracy of the model, each of the network layers (for example, a convolution layer, a pooling layer, a network layer with a random drop of some connection weight mechanism, a fully connected layer, and a classifier) The corresponding activation function f(x) of the layer, etc. is:
f(x)=max(α*x,0)f(x)=max(α*x,0)
其中,α为泄漏率,x表示该深度卷积神经网络模型中神经元的一个数值输入。在本实施例的一个优选实施方式中,将α设定为0.5。经过相同测试数据集的对比测试,相较于其他现有的激活函数,通过本实施例的激活函数f(x),该深度卷积神经网络模型的识别准确率大约有3%的提升。Where α is the leak rate and x is a numerical input of the neurons in the deep convolutional neural network model. In a preferred embodiment of the embodiment, α is set to 0.5. Through the comparison test of the same test data set, the recognition accuracy of the deep convolutional neural network model is improved by about 3% by the activation function f(x) of the present embodiment compared to other existing activation functions.
进一步地,在其他实施例中,为了提高模型的识别精度,各个所述网络层(例如,卷积层、池化层、具有随机丢弃某些连接权重机制的网络层、全连接层及分类器层等)对应的交叉熵H(P,Q)为:Further, in other embodiments, in order to improve the recognition accuracy of the model, each of the network layers (for example, a convolution layer, a pooling layer, a network layer with a random drop of some connection weight mechanism, a fully connected layer, and a classifier) The corresponding cross entropy H(P, Q) of the layer, etc. is:
H(P,Q)=H(P)+DKL(P||Q)H(P,Q)=H(P)+D KL (P||Q)
其中,P,Q为两个概率分布,H(P)为概率分布P的期望,H(P)=-∑x∈XP(x)log P(x),x为概率分布P的样本空间X中任意一个样本,P(x)表示样本x被选取的概率;DKL(P||Q)的表达式为
Figure PCTCN2017100044-appb-000011
x为概率分布P和Q公共样本空间X中任意一个样本,P(x)表示样本x在概率分布P上被选取的概率,Q(x)表示样本x在概率分布Q上被选取的概率。
Where P and Q are two probability distributions, H(P) is the expectation of probability distribution P, H(P)=-∑ x∈X P(x)log P(x), and x is the sample space of probability distribution P For any sample in X, P(x) represents the probability that sample x is selected; the expression of D KL (P||Q) is
Figure PCTCN2017100044-appb-000011
x is any one of the probability distribution P and the Q common sample space X, P(x) represents the probability that the sample x is selected on the probability distribution P, and Q(x) represents the probability that the sample x is selected on the probability distribution Q.
进一步地,为了保证模型训练的效率和准确性,各个所述网络层对应的交叉熵损失函数
Figure PCTCN2017100044-appb-000012
为:
Further, in order to ensure the efficiency and accuracy of the model training, the cross entropy loss function corresponding to each of the network layers
Figure PCTCN2017100044-appb-000012
for:
Figure PCTCN2017100044-appb-000013
Figure PCTCN2017100044-appb-000013
其中,x表示模型的输入,
Figure PCTCN2017100044-appb-000014
表示输入对应的标签,W表示预设的模型参数,X表示模型输入空间,f(x:W)表示模型对输入x的做了变换后的输出,ζ表示规约化因子,||W||2表示对矩阵元素求和:
Where x is the input to the model,
Figure PCTCN2017100044-appb-000014
Indicates the corresponding label of the input, W represents the preset model parameters, X represents the model input space, f(x:W) represents the transformed output of the model to the input x, ζ denotes the statistic factor, ||W|| 2 means summing matrix elements:
Figure PCTCN2017100044-appb-000015
Figure PCTCN2017100044-appb-000015
Wi+1=Wi+ΔWi+1 W i+1 =W i +ΔW i+1
其中,ΔWi+1表示在i+1时刻权值矩阵的更新增量,α为势能项,β为权值 衰减系数,γ为模型的学习率,Wi表示在i时刻权值矩阵状态值,Di表示第i批输入,
Figure PCTCN2017100044-appb-000016
表示第i批输入对应的平均梯度。
Where ΔW i+1 represents the update increment of the weight matrix at time i+1, α is the potential energy term, β is the weight attenuation coefficient, γ is the learning rate of the model, and W i is the state value of the weight matrix at time i , D i represents the i-th batch input,
Figure PCTCN2017100044-appb-000016
Indicates the average gradient corresponding to the i-th batch input.
本实施例中,交叉熵可在神经网络(机器学习)中作为损失函数,例如,P表示真实标记的分布,Q则为训练后的模型的预测标记分布,交叉熵损失函数可以衡量P与Q的相似性,以保证模型训练的准确性。而且,交叉熵作为损失函数在梯度下降时能避免均方误差损失函数学习速率降低的问题,因此,能保证模型训练的效率。In this embodiment, cross entropy can be used as a loss function in neural networks (machine learning). For example, P represents the distribution of real markers, Q is the predicted marker distribution of the trained model, and the cross entropy loss function can measure P and Q. The similarity to ensure the accuracy of the model training. Moreover, the cross entropy as a loss function can avoid the problem of the learning rate reduction of the mean square error loss function when the gradient is lowered, and therefore, the efficiency of the model training can be ensured.
进一步地,在其他实施例中,所述深度卷积神经网络模型包括至少一个全连接层,所述预先确定的识别模型中的各权重的初始值从预设的权重范围(例如,(0,1)权重范围)进行随机采样确定,所述全连接层的连接权重被丢弃(Dropout)的概率设置为第一预设值(例如,0.5),所述交叉熵损失函数中的权值衰减系数设置为第二预设值(例如,0.0005),所述交叉熵损失函数中的势能项设置为第三预设值(例如,0.9)。Further, in other embodiments, the deep convolutional neural network model includes at least one fully connected layer, and an initial value of each weight in the predetermined recognition model is from a preset weight range (eg, (0, 1) Weight range) Performing random sampling determination, the probability that the connection weight of the fully connected layer is discarded (Dropout) is set to a first preset value (for example, 0.5), and the weight attenuation coefficient in the cross entropy loss function Set to a second preset value (eg, 0.0005), the potential energy term in the cross entropy loss function is set to a third preset value (eg, 0.9).
进一步地,在其他实施例中,所述预先确定的识别模型的打分函数
Figure PCTCN2017100044-appb-000017
为:
Further, in other embodiments, the predetermined scoring function of the recognition model
Figure PCTCN2017100044-appb-000017
for:
Figure PCTCN2017100044-appb-000018
Figure PCTCN2017100044-appb-000018
其中,
Figure PCTCN2017100044-appb-000019
Oi,j表示第一次预测为i并且第二次预测为j实际出现的图片数目,O表示一个N*N的矩阵,Oi,j代表矩阵O中的矩阵元素,N表示参与预测的图片数目,预测结果i,j∈{0 1 2 3 4},Ei,j表示第一次预测为i并且第二次预测为j应该出现的图像数目,E是期望的预测结果的N*N矩阵,Ei,j代表矩阵E中的矩阵元素。
among them,
Figure PCTCN2017100044-appb-000019
O i,j represents the first prediction as i and the second prediction is the number of pictures actually appearing in j, O represents a matrix of N*N, O i,j represents the matrix element in matrix O, and N represents the prediction of participation. The number of pictures, the prediction result i, j ∈ {0 1 2 3 4}, E i, j represents the number of images in which the first prediction is i and the second prediction is j, and E is the N* of the desired prediction result. The N matrix, Ei,j represents the matrix elements in the matrix E.
本实施例中通过打分函数
Figure PCTCN2017100044-appb-000020
来检测所述预先确定的识别模型的识别准确率,以保证训练出的所述预先确定的识别模型的识别准确率保持在较高水平,以保证对患者的视网膜病变程度进行准确地识别。
Scoring function in this embodiment
Figure PCTCN2017100044-appb-000020
The recognition accuracy of the predetermined recognition model is detected to ensure that the recognition accuracy of the trained recognition model is maintained at a high level to ensure accurate recognition of the degree of retinopathy of the patient.
在其他实施例中,所述的视网膜病变程度等级检测系统10可以被分割成一个或多个模块,所述一个或者多个模块被存储于所述存储器11中,并由一个或多个处理器(本实施例为所述处理器12)所执行,以完成本发明。本发明所称的模块是指能够完成特定功能的一系列计算机程序指令段,比程序更适合于描述所述视网膜病变程度等级检测系统10在所述电子装置1中的执行过程。In other embodiments, the retinopathy degree level detection system 10 can be segmented into one or more modules, the one or more modules being stored in the memory 11 and being processed by one or more processors (This embodiment is performed by the processor 12) to complete the present invention. The term "module" as used in the present invention refers to a series of computer program instructions that are capable of performing a particular function, and are more suitable than the program for describing the execution of the retinopathy level detection system 10 in the electronic device 1.
请参阅图3,是图2中视网膜病变程度等级检测系统10较佳实施例的功能模块图。在图3中,所述的视网膜病变程度等级检测系统10可以被分割成 识别模块01、确定模块02。所述模块01-02所实现的功能或操作步骤均与上文类似,此处不再详述,示例性地,例如其中:Please refer to FIG. 3, which is a functional block diagram of a preferred embodiment of the retinopathy lesion level detecting system 10 of FIG. In FIG. 3, the retinopathy degree level detecting system 10 can be divided into The module 01 is identified and the module 02 is determined. The functions or operational steps implemented by the module 01-02 are similar to the above, and are not described in detail herein, by way of example, for example:
识别模块01,用于在收到待识别的视网膜病变图片后,对收到的视网膜病变图片利用预先确定的识别模型进行识别,并输出识别结果;其中,所述预先确定的识别模型为预先通过对标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的卷积神经网络模型;及The identification module 01 is configured to: after receiving the retinopathy picture to be identified, identify the received retinopathy picture by using a predetermined recognition model, and output a recognition result; wherein the predetermined recognition model is pre-determined a convolutional neural network model trained to pre-set a predetermined number of sample images with different levels of retinopathy; and
确定模块02,用于根据预先确定的识别结果与视网膜病变程度等级的映射关系,确定输出的识别结果对应的视网膜病变程度等级。The determining module 02 is configured to determine a retinopathy degree level corresponding to the output recognition result according to the mapping relationship between the predetermined recognition result and the retinopathy degree level.
本发明还提供一种计算机可读存储介质,所述计算机可读存储介质上存储有视网膜病变程度等级的检测系统的程序,所述视网膜病变程度等级的检测系统的程序被处理器执行时实现上述视网膜病变程度等级的检测方法的任一步骤。The present invention also provides a computer readable storage medium having stored thereon a program of a detection system for a level of retinopathy, the program of the detection system of the degree of retinopathy level being implemented by a processor Any step of the method of detecting the degree of retinopathy.
本发明之计算机可读存储介质的具体实施方式与上述视网膜病变程度等级的检测方法的具体实施方式大致相同,在此不再赘述。The specific implementation manner of the computer readable storage medium of the present invention is substantially the same as the specific implementation method of the above-mentioned method for detecting the degree of retinopathy degree, and details are not described herein again.
需要说明的是,在本文中,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者装置不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者装置所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括该要素的过程、方法、物品或者装置中还存在另外的相同要素。It is to be understood that the term "comprises", "comprising", or any other variants thereof, is intended to encompass a non-exclusive inclusion, such that a process, method, article, or device comprising a series of elements includes those elements. It also includes other elements that are not explicitly listed, or elements that are inherent to such a process, method, article, or device. An element that is defined by the phrase "comprising a ..." does not exclude the presence of additional equivalent elements in the process, method, item, or device that comprises the element.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件来实现,但很多情况下前者是更佳的实施方式。基于这样的理解,本发明的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端设备(可以是手机,计算机,服务器,空调器,或者网络设备等)执行本发明各个实施例所述的方法。Through the description of the above embodiments, those skilled in the art can clearly understand that the foregoing embodiment method can be implemented by means of software plus a necessary general hardware platform, and can also be implemented by hardware, but in many cases, the former is A better implementation. Based on such understanding, the technical solution of the present invention, which is essential or contributes to the prior art, may be embodied in the form of a software product stored in a storage medium (such as ROM/RAM, disk, The optical disc includes a number of instructions for causing a terminal device (which may be a cell phone, a computer, a server, an air conditioner, or a network device, etc.) to perform the methods described in various embodiments of the present invention.
以上参照附图说明了本发明的优选实施例,并非因此局限本发明的权利范围。上述本发明实施例序号仅仅为了描述,不代表实施例的优劣。另外,虽然在流程图中示出了逻辑顺序,但是在某些情况下,可以以不同于此处的顺序执行所示出或描述的步骤。The preferred embodiments of the present invention have been described above with reference to the drawings, and are not intended to limit the scope of the invention. The serial numbers of the embodiments of the present invention are merely for the description, and do not represent the advantages and disadvantages of the embodiments. Additionally, although logical sequences are shown in the flowcharts, in some cases the steps shown or described may be performed in a different order than the ones described herein.
本领域技术人员不脱离本发明的范围和实质,可以有多种变型方案实现本发明,比如作为一个实施例的特征可用于另一实施例而得到又一实施例。凡在运用本发明的技术构思之内所作的任何修改、等同替换和改进,均应在本发明的权利范围之内。 A person skilled in the art can implement the invention in various variants without departing from the scope and spirit of the invention. For example, the features of one embodiment can be used in another embodiment to obtain a further embodiment. Any modifications, equivalent substitutions and improvements made within the technical concept of the invention are intended to be included within the scope of the invention.

Claims (20)

  1. 一种电子装置,其特征在于,所述装置包括:存储器、处理器,所述存储器上存储有可在所述处理器上运行的视网膜病变程度等级检测系统的程序,所述视网膜病变程度等级检测系统的程序被所述处理器执行时实现如下步骤:An electronic device, comprising: a memory, a processor, wherein the memory stores a program of a retinopathy degree level detecting system operable on the processor, the retinopathy degree level detecting When the program of the system is executed by the processor, the following steps are implemented:
    在收到待识别的视网膜病变图片后,对收到的视网膜病变图片利用预先确定的识别模型进行识别,并输出识别结果;其中,所述预先确定的识别模型为预先通过对标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的卷积神经网络模型;After receiving the picture of the retinopathy to be identified, the received retinopathy picture is identified by using a predetermined recognition model, and the recognition result is output; wherein the predetermined recognition model is pre-marked by different retinopathy a convolutional neural network model obtained by training a predetermined number of sample pictures of a degree level;
    根据预先确定的识别结果与视网膜病变程度等级的映射关系,确定输出的识别结果对应的视网膜病变程度等级。According to the mapping relationship between the predetermined recognition result and the degree of retinopathy degree, the degree of retinopathy corresponding to the output recognition result is determined.
  2. 如权利要求1所述的电子装置,其特征在于,所述预先确定的识别模型的训练过程如下:The electronic device according to claim 1, wherein the training process of the predetermined recognition model is as follows:
    A、为各个预设的视网膜病变程度等级设定对应的预设数量的样本图片,为每个样本图片标注对应的视网膜病变程度等级;A. setting a corresponding preset number of sample pictures for each preset retinopathy degree level, and marking a corresponding retinopathy degree level for each sample picture;
    B、将各张样本图片进行图片预处理以获得待模型训练的训练图片;B. Perform image preprocessing on each sample image to obtain a training picture to be trained by the model;
    C、将所有训练图片分为第一比例的训练集和第二比例的验证集;C. Divide all training pictures into a training set of a first ratio and a verification set of a second ratio;
    D、利用所述训练集训练所述预先确定的识别模型;D. training the predetermined recognition model by using the training set;
    E、利用所述验证集验证训练的识别模型的准确率,若准确率大于或者等于预设准确率,则训练结束,或者,若准确率小于预设准确率,则增加各个视网膜病变程度等级对应的样本图片数量并重新执行上述步骤B、C、D、E。E. verifying the accuracy of the training recognition model by using the verification set. If the accuracy rate is greater than or equal to the preset accuracy rate, the training ends, or if the accuracy rate is less than the preset accuracy rate, increasing the level of each retinopathy level The number of sample pictures and re-execute steps B, C, D, E above.
  3. 如权利要求2所述的电子装置,其特征在于,所述步骤B包括:The electronic device of claim 2, wherein the step B comprises:
    将各张样本图片的较短边长缩放到第一预设大小以获得对应的第一图片,在各张第一图片上随机裁剪出一个第二预设大小的第二图片;Scaling a shorter side length of each sample picture to a first preset size to obtain a corresponding first picture, and randomly cutting a second picture of a second preset size on each first picture;
    根据各个预先确定的预设类型参数对应的标准参数值,将各张第二图片的各个预先确定的预设类型参数值调整为对应的标准参数值,以获得对应的第三图片;Adjusting each predetermined preset type parameter value of each second picture to a corresponding standard parameter value according to a standard parameter value corresponding to each predetermined preset type parameter, to obtain a corresponding third picture;
    对各张第三图片进行预设方向的翻转操作,及按照预设的扭曲角度对各张第三图片进行扭曲操作,以获得各张第三图片对应的第四图片,将各张第四图片作为待模型训练的训练图片。Performing a preset direction flip operation on each third picture, and performing a twist operation on each third picture according to a preset twist angle to obtain a fourth picture corresponding to each third picture, and each fourth picture As a training picture to be trained by the model.
  4. 如权利要求1或2所述的电子装置,其特征在于,所述深度卷积神经网络模型包括输入层和多个网络层,所述网络层包括卷积层、池化层、全连接层及分类器层。The electronic device according to claim 1 or 2, wherein the deep convolutional neural network model comprises an input layer and a plurality of network layers, the network layer comprising a convolution layer, a pooling layer, a fully connected layer, and Classifier layer.
  5. 如权利要求4所述的电子装置,其特征在于,各个所述网络层对应的 激活函数f(x)为:The electronic device according to claim 4, wherein each of said network layers corresponds to The activation function f(x) is:
    f(x)=max(α*x,0)f(x)=max(α*x,0)
    其中,α为预设的泄漏率,x表示所述深度卷积神经网络模型中神经元的一个数值输入。Where α is the preset leak rate and x is a numerical input of the neurons in the deep convolutional neural network model.
  6. 如权利要求4所述的电子装置,其特征在于,各个所述网络层对应的交叉熵H(P,Q)为:The electronic device according to claim 4, wherein the cross entropy H(P, Q) corresponding to each of the network layers is:
    H(P,Q)=H(P)+DKL(P||Q)H(P,Q)=H(P)+D KL (P||Q)
    其中,P,Q为两个概率分布,H(P)为概率分布P的期望,H(P)=-∑x∈XP(x)logP(x),x为概率分布P的样本空间X中任意一个样本,P(x)表示样本x被选取的概率;DKL(P||Q)的表达式为
    Figure PCTCN2017100044-appb-100001
    x为概率分布P和Q公共样本空间X中任意一个样本,P(x)表示样本x在概率分布P上被选取的概率,Q(x)表示样本x在概率分布Q上被选取的概率。
    Where P and Q are two probability distributions, H(P) is the expectation of the probability distribution P, H(P)=-∑ x∈X P(x)logP(x), and x is the sample space X of the probability distribution P In any of the samples, P(x) represents the probability that the sample x is selected; the expression of D KL (P||Q) is
    Figure PCTCN2017100044-appb-100001
    x is any one of the probability distribution P and the Q common sample space X, P(x) represents the probability that the sample x is selected on the probability distribution P, and Q(x) represents the probability that the sample x is selected on the probability distribution Q.
  7. 如权利要求1或2所述的电子装置,其特征在于,所述预先确定的识别模型的打分函数
    Figure PCTCN2017100044-appb-100002
    为:
    The electronic device according to claim 1 or 2, wherein the predetermined scoring function of the recognition model
    Figure PCTCN2017100044-appb-100002
    for:
    Figure PCTCN2017100044-appb-100003
    Figure PCTCN2017100044-appb-100003
    其中,
    Figure PCTCN2017100044-appb-100004
    Oi,j表示第一次预测为i并且第二次预测为j实际出现的图片数目,O表示一个N*N的矩阵,Oi,j代表矩阵O中的矩阵元素,N表示参与预测的图片数目,预测结果i,j∈{01234},Ei,j表示第一次预测为i并且第二次预测为j应该出现的图像数目,E是期望的预测结果的N*N矩阵,Ei,j代表矩阵E中的矩阵元素。
    among them,
    Figure PCTCN2017100044-appb-100004
    O i,j represents the first prediction as i and the second prediction is the number of pictures actually appearing in j, O represents a matrix of N*N, O i,j represents the matrix element in matrix O, and N represents the prediction of participation. The number of pictures, the prediction result i, j ∈ {01234}, E i, j represents the number of images in which the first prediction is i and the second prediction is j, and E is the N*N matrix of the desired prediction result, E i, j represents the matrix element in the matrix E.
  8. 如权利要求4所述的电子装置,其特征在于,各个所述网络层对应的交叉熵损失函数
    Figure PCTCN2017100044-appb-100005
    为:
    The electronic device according to claim 4, wherein a cross entropy loss function corresponding to each of said network layers
    Figure PCTCN2017100044-appb-100005
    for:
    Figure PCTCN2017100044-appb-100006
    Figure PCTCN2017100044-appb-100006
    其中,x表示模型的输入,
    Figure PCTCN2017100044-appb-100007
    表示输入对应的标签,W表示预设的模型参数,X表示模型输入空间,f(x:W)表示模型对输入x的做了变换后的输出,ζ表示规约化因子,||W||2表示对矩阵元素求和:
    Where x is the input to the model,
    Figure PCTCN2017100044-appb-100007
    Indicates the corresponding label of the input, W represents the preset model parameters, X represents the model input space, f(x:W) represents the transformed output of the model to the input x, ζ denotes the statistic factor, ||W|| 2 means summing matrix elements:
    Figure PCTCN2017100044-appb-100008
    Figure PCTCN2017100044-appb-100008
    Wi+1=Wi+ΔWi+1 W i+1 =W i +ΔW i+1
    其中,ΔWi+1表示在i+1时刻权值矩阵的更新增量,α为势能项,β为权值衰减系数,γ为模型的学习率,Wi表示在i时刻权值矩阵状态值,Di表示第i批输入,
    Figure PCTCN2017100044-appb-100009
    表示第i批输入对应的平均梯度。
    Where ΔW i+1 represents the update increment of the weight matrix at time i+1, α is the potential energy term, β is the weight attenuation coefficient, γ is the learning rate of the model, and W i is the state value of the weight matrix at time i , D i represents the i-th batch input,
    Figure PCTCN2017100044-appb-100009
    Indicates the average gradient corresponding to the i-th batch input.
  9. 如权利要求8所述的电子装置,其特征在于,所述预先确定的识别模型包括至少一个全连接层,所述预先确定的识别模型中的各权重的初始值从预设的权重范围进行随机采样确定,所述全连接层的连接权重被丢弃的概率设置为第一预设值,所述交叉熵损失函数中的权值衰减系数设置为第二预设值,所述交叉熵损失函数中的势能项设置为第三预设值。The electronic device according to claim 8, wherein said predetermined recognition model comprises at least one fully connected layer, and initial values of respective weights in said predetermined recognition model are randomly selected from a preset weight range Sampling determines that a probability that the connection weight of the fully connected layer is discarded is set to a first preset value, and a weight attenuation coefficient in the cross entropy loss function is set to a second preset value, where the cross entropy loss function is The potential energy item is set to a third preset value.
  10. 一种视网膜病变程度等级的检测方法,其特征在于,所述方法包括以下步骤:A method for detecting a degree of retinopathy severity, characterized in that the method comprises the following steps:
    在收到待识别的视网膜病变图片后,对收到的视网膜病变图片利用预先确定的识别模型进行识别,并输出识别结果;其中,所述预先确定的识别模型为预先通过对标注有不同视网膜病变程度等级的预设数量样本图片进行训练得到的卷积神经网络模型;After receiving the picture of the retinopathy to be identified, the received retinopathy picture is identified by using a predetermined recognition model, and the recognition result is output; wherein the predetermined recognition model is pre-marked by different retinopathy a convolutional neural network model obtained by training a predetermined number of sample pictures of a degree level;
    根据预先确定的识别结果与视网膜病变程度等级的映射关系,确定输出的识别结果对应的视网膜病变程度等级。According to the mapping relationship between the predetermined recognition result and the degree of retinopathy degree, the degree of retinopathy corresponding to the output recognition result is determined.
  11. 如权利要求10所述的视网膜病变程度等级的检测方法,其特征在于,所述预先确定的识别模型的训练过程如下:The method for detecting a degree of retinopathy according to claim 10, wherein the training process of the predetermined recognition model is as follows:
    A、为各个预设的视网膜病变程度等级设定对应的预设数量的样本图片,为每个样本图片标注对应的视网膜病变程度等级;A. setting a corresponding preset number of sample pictures for each preset retinopathy degree level, and marking a corresponding retinopathy degree level for each sample picture;
    B、将各张样本图片进行图片预处理以获得待模型训练的训练图片;B. Perform image preprocessing on each sample image to obtain a training picture to be trained by the model;
    C、将所有训练图片分为第一比例的训练集和第二比例的验证集;C. Divide all training pictures into a training set of a first ratio and a verification set of a second ratio;
    D、利用所述训练集训练所述预先确定的识别模型;D. training the predetermined recognition model by using the training set;
    E、利用所述验证集验证训练的识别模型的准确率,若准确率大于或者等于预设准确率,则训练结束,或者,若准确率小于预设准确率,则增加各个视网膜病变程度等级对应的样本图片数量并重新执行上述步骤B、C、D、E。E. verifying the accuracy of the training recognition model by using the verification set. If the accuracy rate is greater than or equal to the preset accuracy rate, the training ends, or if the accuracy rate is less than the preset accuracy rate, increasing the level of each retinopathy level The number of sample pictures and re-execute steps B, C, D, E above.
  12. 如权利要求11所述的视网膜病变程度等级的检测方法,其特征在于,所述步骤B包括:The method for detecting a degree of retinopathy according to claim 11, wherein the step B comprises:
    将各张样本图片的较短边长缩放到第一预设大小以获得对应的第一图片,在各张第一图片上随机裁剪出一个第二预设大小的第二图片;Scaling a shorter side length of each sample picture to a first preset size to obtain a corresponding first picture, and randomly cutting a second picture of a second preset size on each first picture;
    根据各个预先确定的预设类型参数对应的标准参数值,将各张第二图片的各个预先确定的预设类型参数值调整为对应的标准参数值,以获得对应的第三图片; Adjusting each predetermined preset type parameter value of each second picture to a corresponding standard parameter value according to a standard parameter value corresponding to each predetermined preset type parameter, to obtain a corresponding third picture;
    对各张第三图片进行预设方向的翻转操作,及按照预设的扭曲角度对各张第三图片进行扭曲操作,以获得各张第三图片对应的第四图片,将各张第四图片作为待模型训练的训练图片。Performing a preset direction flip operation on each third picture, and performing a twist operation on each third picture according to a preset twist angle to obtain a fourth picture corresponding to each third picture, and each fourth picture As a training picture to be trained by the model.
  13. 如权利要求10或11所述的视网膜病变程度等级的检测方法,其特征在于,所述深度卷积神经网络模型包括输入层和多个网络层。The method of detecting a degree of retinopathy according to claim 10 or 11, wherein the deep convolutional neural network model comprises an input layer and a plurality of network layers.
  14. 如权利要求13所述的视网膜病变程度等级的检测方法,其特征在于,所述网络层包括卷积层、池化层、全连接层及分类器层。The method for detecting a degree of retinopathy according to claim 13, wherein the network layer comprises a convolution layer, a pooling layer, a fully connected layer, and a classifier layer.
  15. 如权利要求14所述的视网膜病变程度等级的检测方法,其特征在于,各个所述网络层对应的激活函数f(x)为:The method for detecting a degree of retinopathy according to claim 14, wherein the activation function f(x) corresponding to each of said network layers is:
    f(x)=max(α*x,0)f(x)=max(α*x,0)
    其中,α为预设的泄漏率,x表示所述深度卷积神经网络模型中神经元的一个数值输入。Where α is the preset leak rate and x is a numerical input of the neurons in the deep convolutional neural network model.
  16. 如权利要求14所述的视网膜病变程度等级的检测方法,其特征在于,各个所述网络层对应的交叉熵H(P,Q)为:The method for detecting a degree of retinopathy according to claim 14, wherein the cross entropy H(P, Q) corresponding to each of the network layers is:
    H(P,Q)=H(P)+DKL(P||Q)H(P,Q)=H(P)+D KL (P||Q)
    其中,P,Q为两个概率分布,H(P)为概率分布P的期望,H(P)=-∑x∈XP(x)logP(x),x为概率分布P的样本空间X中任意一个样本,P(x)表示样本x被选取的概率;DKL(P||Q)的表达式为
    Figure PCTCN2017100044-appb-100010
    x为概率分布P和Q公共样本空间X中任意一个样本,P(x)表示样本x在概率分布P上被选取的概率,Q(x)表示样本x在概率分布Q上被选取的概率。
    Where P and Q are two probability distributions, H(P) is the expectation of the probability distribution P, H(P)=-∑ x∈X P(x)logP(x), and x is the sample space X of the probability distribution P In any of the samples, P(x) represents the probability that the sample x is selected; the expression of D KL (P||Q) is
    Figure PCTCN2017100044-appb-100010
    x is any one of the probability distribution P and the Q common sample space X, P(x) represents the probability that the sample x is selected on the probability distribution P, and Q(x) represents the probability that the sample x is selected on the probability distribution Q.
  17. 如权利要求10或11所述的视网膜病变程度等级的检测方法,其特征在于,所述预先确定的识别模型的打分函数
    Figure PCTCN2017100044-appb-100011
    为:
    The method for detecting a degree of retinopathy of claim 10 or 11, wherein the predetermined scoring function of the recognition model
    Figure PCTCN2017100044-appb-100011
    for:
    Figure PCTCN2017100044-appb-100012
    Figure PCTCN2017100044-appb-100012
    其中,
    Figure PCTCN2017100044-appb-100013
    Oi,j表示第一次预测为i并且第二次预测为j实际出现的图片数目,O表示一个N*N的矩阵,Oi,j代表矩阵O中的矩阵元素,N表示参与预测的图片数目,预测结果i,j∈{01234},Ei,j表示第一次预测为i并且第二次预测为j应该出现的图像数目,E是期望的预测结果的N*N矩阵,Ei,j代表矩阵E中的矩阵元素。
    among them,
    Figure PCTCN2017100044-appb-100013
    O i,j represents the first prediction as i and the second prediction is the number of pictures actually appearing in j, O represents a matrix of N*N, O i,j represents the matrix element in matrix O, and N represents the prediction of participation. The number of pictures, the prediction result i, j ∈ {01234}, E i, j represents the number of images in which the first prediction is i and the second prediction is j, and E is the N*N matrix of the desired prediction result, E i, j represents the matrix element in the matrix E.
  18. 如权利要求14所述的视网膜病变程度等级的检测方法,其特征在于,各个所述网络层对应的交叉熵损失函数
    Figure PCTCN2017100044-appb-100014
    为:
    A method for detecting a degree of retinopathy according to claim 14, wherein a cross entropy loss function corresponding to each of said network layers is provided.
    Figure PCTCN2017100044-appb-100014
    for:
    Figure PCTCN2017100044-appb-100015
    Figure PCTCN2017100044-appb-100015
    其中,x表示模型的输入,
    Figure PCTCN2017100044-appb-100016
    表示输入对应的标签,W表示预设的模型参数,X表示模型输入空间,f(x:W)表示模型对输入x的做了变换后的输出,ζ表示规约化因子,||W||2表示对矩阵元素求和:
    Where x is the input to the model,
    Figure PCTCN2017100044-appb-100016
    Indicates the corresponding label of the input, W represents the preset model parameters, X represents the model input space, f(x:W) represents the transformed output of the model to the input x, ζ denotes the statistic factor, ||W|| 2 means summing matrix elements:
    Figure PCTCN2017100044-appb-100017
    Figure PCTCN2017100044-appb-100017
    Wi+1=Wi+ΔWi+1 W i+1 =W i +ΔW i+1
    其中,ΔWi+1表示在i+1时刻权值矩阵的更新增量,α为势能项,β为权值衰减系数,γ为模型的学习率,Wi表示在i时刻权值矩阵状态值,Di表示第i批输入,
    Figure PCTCN2017100044-appb-100018
    表示第i批输入对应的平均梯度。
    Where ΔW i+1 represents the update increment of the weight matrix at time i+1, α is the potential energy term, β is the weight attenuation coefficient, γ is the learning rate of the model, and W i is the state value of the weight matrix at time i , D i represents the i-th batch input,
    Figure PCTCN2017100044-appb-100018
    Indicates the average gradient corresponding to the i-th batch input.
  19. 如权利要求14所述的视网膜病变程度等级的检测方法,其特征在于,所述预先确定的识别模型包括至少一个全连接层,所述预先确定的识别模型中的各权重的初始值从预设的权重范围进行随机采样确定,所述全连接层的连接权重被丢弃的概率设置为第一预设值,所述交叉熵损失函数中的权值衰减系数设置为第二预设值,所述交叉熵损失函数中的势能项设置为第三预设值。The method for detecting a degree of retinopathy degree according to claim 14, wherein said predetermined recognition model comprises at least one fully connected layer, and initial values of respective weights in said predetermined recognition model are preset The weight range is determined by random sampling, the probability that the connection weight of the fully connected layer is discarded is set to a first preset value, and the weight attenuation coefficient in the cross entropy loss function is set to a second preset value, The potential energy term in the cross entropy loss function is set to a third preset value.
  20. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质上存储有视网膜病变程度等级的检测系统的程序,所述视网膜病变程度等级的检测系统的程序被处理器执行时实现如权利要求10至19中任一项所述的视网膜病变程度等级的检测方法的步骤。 A computer readable storage medium, wherein the computer readable storage medium stores a program of a detection system of a degree of retinopathy level, and the program of the detection system of the level of retinopathy is implemented by a processor, for example The method of the method for detecting a degree of retinopathy according to any one of claims 10 to 19.
PCT/CN2017/100044 2017-05-05 2017-08-31 Method for detecting retinopathy degree level, device and storage medium WO2018201647A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201710312327.6A CN107203778A (en) 2017-05-05 2017-05-05 PVR intensity grade detecting system and method
CN201710312327.6 2017-05-05

Publications (1)

Publication Number Publication Date
WO2018201647A1 true WO2018201647A1 (en) 2018-11-08

Family

ID=59905737

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/100044 WO2018201647A1 (en) 2017-05-05 2017-08-31 Method for detecting retinopathy degree level, device and storage medium

Country Status (2)

Country Link
CN (1) CN107203778A (en)
WO (1) WO2018201647A1 (en)

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110211109A (en) * 2019-05-30 2019-09-06 西安电子科技大学 Image change detection method based on deep neural network structure optimizing
CN110414607A (en) * 2019-07-31 2019-11-05 中山大学 Classification method, device, equipment and the medium of capsule endoscope image
CN110517225A (en) * 2019-07-19 2019-11-29 平安科技(深圳)有限公司 AI image recognition method, device, equipment and storage medium
CN110555478A (en) * 2019-09-05 2019-12-10 东北大学 Fan multi-fault diagnosis method based on depth measurement network of difficult sample mining
CN110599451A (en) * 2019-08-05 2019-12-20 平安科技(深圳)有限公司 Medical image focus detection positioning method, device, equipment and storage medium
CN111242933A (en) * 2020-01-15 2020-06-05 华南理工大学 Retina image artery and vein classification device, equipment and storage medium
CN111275121A (en) * 2020-01-23 2020-06-12 北京百度网讯科技有限公司 Medical image processing method and device and electronic equipment
CN111612021A (en) * 2019-02-22 2020-09-01 中国移动通信有限公司研究院 Error sample identification method and device and terminal
CN111696101A (en) * 2020-06-18 2020-09-22 中国农业大学 Light-weight solanaceae disease identification method based on SE-Inception
CN111754486A (en) * 2020-06-24 2020-10-09 北京百度网讯科技有限公司 Image processing method, image processing device, electronic equipment and storage medium
CN111914814A (en) * 2020-09-01 2020-11-10 平安国际智慧城市科技股份有限公司 Wheat rust detection method and device and computer equipment
CN112052904A (en) * 2020-09-09 2020-12-08 陕西理工大学 Method for identifying plant diseases and insect pests based on transfer learning and convolutional neural network
CN112215239A (en) * 2020-09-15 2021-01-12 浙江工业大学 Retinal lesion fine-grained grading method and device
CN112581448A (en) * 2020-12-17 2021-03-30 中南大学 Retina image hard exudate identification method and imaging method
CN112819797A (en) * 2021-02-06 2021-05-18 国药集团基因科技有限公司 Diabetic retinopathy analysis method, device, system and storage medium
CN113191478A (en) * 2020-01-14 2021-07-30 阿里巴巴集团控股有限公司 Training method, device and system of neural network model
CN115115569A (en) * 2021-03-19 2022-09-27 宏碁智医股份有限公司 Image correlation detection method and detection device
WO2023061080A1 (en) * 2021-10-14 2023-04-20 北京字节跳动网络技术有限公司 Method and apparatus for recognizing tissue image, readable medium, and electronic device
CN118345334A (en) * 2024-06-17 2024-07-16 华兴源创(成都)科技有限公司 Film thickness correction method and device and computer equipment

Families Citing this family (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107729929B (en) 2017-09-30 2021-03-19 百度在线网络技术(北京)有限公司 Method and device for acquiring information
CN107680684B (en) * 2017-10-12 2021-05-07 百度在线网络技术(北京)有限公司 Method and device for acquiring information
EP3698702B1 (en) * 2017-10-17 2023-06-07 FUJIFILM Corporation Medical image processing device and endoscope device
CN109726726B (en) * 2017-10-27 2023-06-20 北京邮电大学 Event detection method and device in video
CN110349156B (en) * 2017-11-30 2023-05-30 腾讯科技(深圳)有限公司 Method and device for identifying lesion characteristics in fundus picture and storage medium
CN108154509B (en) * 2018-01-12 2022-11-11 平安科技(深圳)有限公司 Cancer identification method, device and storage medium
CN110097966B (en) * 2018-01-30 2021-09-14 中国移动通信有限公司研究院 Information reminding method and device and terminal equipment
CN108564570A (en) * 2018-03-29 2018-09-21 哈尔滨工业大学(威海) A kind of method and apparatus of intelligentized pathological tissues positioning
CN108615051B (en) * 2018-04-13 2020-09-15 博众精工科技股份有限公司 Diabetic retina image classification method and system based on deep learning
CN108595583B (en) * 2018-04-18 2022-12-02 平安科技(深圳)有限公司 Dynamic graph page data crawling method, device, terminal and storage medium
CN108392174B (en) * 2018-04-19 2021-01-19 梁建宏 Automatic examination method and system for retinopathy of prematurity
CN109034264B (en) * 2018-08-15 2021-11-19 云南大学 CSP-CNN model for predicting severity of traffic accident and modeling method thereof
WO2020061972A1 (en) * 2018-09-27 2020-04-02 电子科技大学 Convolutional neural network-based diabetic retinopathy diagnosing technique
CN109427060A (en) * 2018-10-30 2019-03-05 腾讯科技(深圳)有限公司 A kind of method, apparatus, terminal device and the medical system of image identification
US10963757B2 (en) 2018-12-14 2021-03-30 Industrial Technology Research Institute Neural network model fusion method and electronic device using the same
CN109567797B (en) * 2019-01-30 2021-10-01 浙江强脑科技有限公司 Epilepsy early warning method and device and computer readable storage medium
EP3948773A1 (en) * 2019-03-29 2022-02-09 Ai Technologies Inc. Image-based detection of ophthalmic and systemic diseases
TWI702615B (en) * 2019-07-26 2020-08-21 長佳智能股份有限公司 Retinopathy assessment model establishment method and system
CN110739071B (en) * 2019-10-10 2022-05-31 北京致远慧图科技有限公司 Method and device for determining optic disc and yellow spot combined positioning model and storage medium
CN110720888A (en) * 2019-10-12 2020-01-24 杭州电子科技大学 Method for predicting macular edema lesion of fundus image based on deep learning
CN110956628B (en) * 2019-12-13 2023-05-09 广州达安临床检验中心有限公司 Picture grade classification method, device, computer equipment and storage medium
CN110992364B (en) * 2019-12-31 2023-11-28 重庆艾可立安医疗器械有限公司 Retina image recognition method, retina image recognition device, computer equipment and storage medium
CN111739616B (en) * 2020-07-20 2020-12-01 平安国际智慧城市科技股份有限公司 Eye image processing method, device, equipment and storage medium
US20220207732A1 (en) * 2020-12-28 2022-06-30 Seyed Ehsan Vaghefi Rezaei Systems and methods for processing of fundus images
WO2022145541A1 (en) * 2020-12-30 2022-07-07 주식회사 메디웨일 Method and device for assisting renal disease diagnosis
CN113569612B (en) * 2021-02-09 2022-09-13 腾讯医疗健康(深圳)有限公司 Training method and device for image recognition neural network and image recognition method

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160292856A1 (en) * 2015-04-06 2016-10-06 IDx, LLC Systems and methods for feature detection in retinal images
CN106530295A (en) * 2016-11-07 2017-03-22 首都医科大学 Fundus image classification method and device of retinopathy
US20170112372A1 (en) * 2015-10-23 2017-04-27 International Business Machines Corporation Automatically detecting eye type in retinal fundus images
CN106934798A (en) * 2017-02-20 2017-07-07 苏州体素信息科技有限公司 Diabetic retinopathy classification stage division based on deep learning

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3116375B1 (en) * 2014-03-14 2023-07-19 Lkc Technologies Inc. System and method for retinopathy detection
CN106295547A (en) * 2016-08-05 2017-01-04 深圳市商汤科技有限公司 A kind of image comparison method and image comparison device
CN106570530A (en) * 2016-11-10 2017-04-19 西南交通大学 Extraction method for extracting hard exudates in ophthalmoscopic image

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160292856A1 (en) * 2015-04-06 2016-10-06 IDx, LLC Systems and methods for feature detection in retinal images
US20170112372A1 (en) * 2015-10-23 2017-04-27 International Business Machines Corporation Automatically detecting eye type in retinal fundus images
CN106530295A (en) * 2016-11-07 2017-03-22 首都医科大学 Fundus image classification method and device of retinopathy
CN106934798A (en) * 2017-02-20 2017-07-07 苏州体素信息科技有限公司 Diabetic retinopathy classification stage division based on deep learning

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
DING, PENGLI ET AL.: "Diabetic Retinal Image Classification Method based on Deep Neural Network", JOURNAL OF COMPUTER APPLICATIONS, vol. 37, no. 3, 10 March 2017 (2017-03-10) *

Cited By (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111612021B (en) * 2019-02-22 2023-10-31 中国移动通信有限公司研究院 Error sample identification method, device and terminal
CN111612021A (en) * 2019-02-22 2020-09-01 中国移动通信有限公司研究院 Error sample identification method and device and terminal
CN110211109A (en) * 2019-05-30 2019-09-06 西安电子科技大学 Image change detection method based on deep neural network structure optimizing
CN110211109B (en) * 2019-05-30 2022-12-06 西安电子科技大学 Image change detection method based on deep neural network structure optimization
CN110517225B (en) * 2019-07-19 2023-07-11 平安科技(深圳)有限公司 AI image recognition method, apparatus, device and storage medium
CN110517225A (en) * 2019-07-19 2019-11-29 平安科技(深圳)有限公司 AI image recognition method, device, equipment and storage medium
CN110414607A (en) * 2019-07-31 2019-11-05 中山大学 Classification method, device, equipment and the medium of capsule endoscope image
CN110599451A (en) * 2019-08-05 2019-12-20 平安科技(深圳)有限公司 Medical image focus detection positioning method, device, equipment and storage medium
CN110555478A (en) * 2019-09-05 2019-12-10 东北大学 Fan multi-fault diagnosis method based on depth measurement network of difficult sample mining
CN110555478B (en) * 2019-09-05 2023-02-03 东北大学 Fan multi-fault diagnosis method based on depth measurement network of difficult sample mining
CN113191478A (en) * 2020-01-14 2021-07-30 阿里巴巴集团控股有限公司 Training method, device and system of neural network model
CN111242933A (en) * 2020-01-15 2020-06-05 华南理工大学 Retina image artery and vein classification device, equipment and storage medium
CN111242933B (en) * 2020-01-15 2023-06-20 华南理工大学 Retinal image artery and vein classification device, apparatus, and storage medium
CN111275121A (en) * 2020-01-23 2020-06-12 北京百度网讯科技有限公司 Medical image processing method and device and electronic equipment
CN111275121B (en) * 2020-01-23 2023-07-18 北京康夫子健康技术有限公司 Medical image processing method and device and electronic equipment
CN111696101A (en) * 2020-06-18 2020-09-22 中国农业大学 Light-weight solanaceae disease identification method based on SE-Inception
CN111754486A (en) * 2020-06-24 2020-10-09 北京百度网讯科技有限公司 Image processing method, image processing device, electronic equipment and storage medium
CN111754486B (en) * 2020-06-24 2023-08-15 北京百度网讯科技有限公司 Image processing method, device, electronic equipment and storage medium
CN111914814A (en) * 2020-09-01 2020-11-10 平安国际智慧城市科技股份有限公司 Wheat rust detection method and device and computer equipment
CN112052904A (en) * 2020-09-09 2020-12-08 陕西理工大学 Method for identifying plant diseases and insect pests based on transfer learning and convolutional neural network
CN112215239A (en) * 2020-09-15 2021-01-12 浙江工业大学 Retinal lesion fine-grained grading method and device
CN112581448A (en) * 2020-12-17 2021-03-30 中南大学 Retina image hard exudate identification method and imaging method
CN112819797A (en) * 2021-02-06 2021-05-18 国药集团基因科技有限公司 Diabetic retinopathy analysis method, device, system and storage medium
CN112819797B (en) * 2021-02-06 2023-09-19 国药集团基因科技有限公司 Method, device, system and storage medium for analyzing diabetic retinopathy
CN115115569A (en) * 2021-03-19 2022-09-27 宏碁智医股份有限公司 Image correlation detection method and detection device
WO2023061080A1 (en) * 2021-10-14 2023-04-20 北京字节跳动网络技术有限公司 Method and apparatus for recognizing tissue image, readable medium, and electronic device
CN118345334A (en) * 2024-06-17 2024-07-16 华兴源创(成都)科技有限公司 Film thickness correction method and device and computer equipment

Also Published As

Publication number Publication date
CN107203778A (en) 2017-09-26

Similar Documents

Publication Publication Date Title
WO2018201647A1 (en) Method for detecting retinopathy degree level, device and storage medium
US11631175B2 (en) AI-based heat map generating system and methods for use therewith
US11961227B2 (en) Method and device for detecting and locating lesion in medical image, equipment and storage medium
WO2019120115A1 (en) Facial recognition method, apparatus, and computer apparatus
US11551377B2 (en) Eye gaze tracking using neural networks
WO2018166114A1 (en) Picture identification method and system, electronic device, and medium
WO2019174130A1 (en) Bill recognition method, server, and computer readable storage medium
WO2019109526A1 (en) Method and device for age recognition of face image, storage medium
TWI754806B (en) System and method for locating iris using deep learning
CN106845414B (en) Method and system for quality metrics for biometric verification
WO2019062080A1 (en) Identity recognition method, electronic device, and computer readable storage medium
WO2020140370A1 (en) Method and device for automatically detecting petechia in fundus, and computer-readable storage medium
WO2021082691A1 (en) Segmentation method and apparatus for lesion area of eye oct image, and terminal device
WO2020238044A1 (en) Method and device for constructing 3d unet network model for tumor detection, and storage medium
WO2022105118A1 (en) Image-based health status identification method and apparatus, device and storage medium
CN111598038B (en) Facial feature point detection method, device, equipment and storage medium
CN112052850B (en) License plate recognition method and device, electronic equipment and storage medium
US10607077B1 (en) Identity authentication using an inlier neural network
US20200302149A1 (en) Fusing multi-spectral images for identity authentication
WO2022126903A1 (en) Method and device for image anomaly area detection, electronic device, and storage medium
US10521580B1 (en) Open data biometric identity validation
CN112132812B (en) Certificate verification method and device, electronic equipment and medium
CN112613471B (en) Face living body detection method, device and computer readable storage medium
WO2024125217A1 (en) Light spot tracking method and apparatus, and electronic device and storage medium
US20240161035A1 (en) Multi-model medical scan analysis system and methods for use therewith

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17908360

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 19.02.2020)

122 Ep: pct application non-entry in european phase

Ref document number: 17908360

Country of ref document: EP

Kind code of ref document: A1