CN111444960A - Skin disease image classification system based on multi-mode data input - Google Patents
Skin disease image classification system based on multi-mode data input Download PDFInfo
- Publication number
- CN111444960A CN111444960A CN202010222477.XA CN202010222477A CN111444960A CN 111444960 A CN111444960 A CN 111444960A CN 202010222477 A CN202010222477 A CN 202010222477A CN 111444960 A CN111444960 A CN 111444960A
- Authority
- CN
- China
- Prior art keywords
- image
- data
- text
- layer
- classification
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 208000017520 skin disease Diseases 0.000 title claims abstract description 56
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 27
- 238000013508 migration Methods 0.000 claims abstract description 20
- 230000005012 migration Effects 0.000 claims abstract description 20
- 238000007500 overflow downdraw method Methods 0.000 claims abstract description 4
- 238000000605 extraction Methods 0.000 claims description 38
- 238000000034 method Methods 0.000 claims description 25
- 210000002569 neuron Anatomy 0.000 claims description 23
- 238000012549 training Methods 0.000 claims description 21
- 230000004913 activation Effects 0.000 claims description 16
- 230000006870 function Effects 0.000 claims description 16
- 239000011159 matrix material Substances 0.000 claims description 15
- 230000004927 fusion Effects 0.000 claims description 14
- 239000013598 vector Substances 0.000 claims description 14
- 230000035945 sensitivity Effects 0.000 claims description 12
- 238000012360 testing method Methods 0.000 claims description 9
- 238000007781 pre-processing Methods 0.000 claims description 8
- 238000013528 artificial neural network Methods 0.000 claims description 6
- 230000005540 biological transmission Effects 0.000 claims description 2
- 230000015572 biosynthetic process Effects 0.000 claims description 2
- 230000010365 information processing Effects 0.000 claims description 2
- 238000003786 synthesis reaction Methods 0.000 claims description 2
- 206010048768 Dermatosis Diseases 0.000 abstract description 2
- 230000008569 process Effects 0.000 description 7
- 231100000444 skin lesion Toxicity 0.000 description 7
- 206010040882 skin lesion Diseases 0.000 description 6
- 201000010099 disease Diseases 0.000 description 5
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 5
- 238000012545 processing Methods 0.000 description 5
- 210000003491 skin Anatomy 0.000 description 5
- 238000013145 classification model Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 238000013507 mapping Methods 0.000 description 4
- 238000003745 diagnosis Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 239000000284 extract Substances 0.000 description 3
- 238000010606 normalization Methods 0.000 description 3
- 238000012790 confirmation Methods 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 239000013589 supplement Substances 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- 208000005440 Basal Cell Neoplasms Diseases 0.000 description 1
- 206010004146 Basal cell carcinoma Diseases 0.000 description 1
- 208000007256 Nevus Diseases 0.000 description 1
- 208000012641 Pigmentation disease Diseases 0.000 description 1
- 206010039796 Seborrhoeic keratosis Diseases 0.000 description 1
- 208000009621 actinic keratosis Diseases 0.000 description 1
- 238000003759 clinical diagnosis Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 206010016629 fibroma Diseases 0.000 description 1
- 238000013100 final test Methods 0.000 description 1
- 230000008014 freezing Effects 0.000 description 1
- 238000007710 freezing Methods 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000003902 lesion Effects 0.000 description 1
- 238000000968 medical method and process Methods 0.000 description 1
- 210000002752 melanocyte Anatomy 0.000 description 1
- 201000001441 melanoma Diseases 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 201000003385 seborrheic keratosis Diseases 0.000 description 1
- 238000013526 transfer learning Methods 0.000 description 1
- 230000002792 vascular Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
Abstract
A dermatologic image classification system based on multi-modal data input, comprising: the invention fully utilizes the text supplementary data matched with the image data and improves the accuracy of the classification of the dermatosis image. The input data is divided into image data and text data, and after image data standardization and text data coding are carried out, an image feature extractor based on a depth migration convolutional neural network and a text feature extractor based on a multilayer perceptron unit are respectively constructed. And inputting the image characteristics and the text characteristics into the comprehensive full-connection layer by using a characteristic fusion method, outputting the class labels of the data, and realizing the classification task of the skin disease images.
Description
Technical Field
The invention relates to the technology in the field of image processing, in particular to a skin disease image classification system based on multi-mode data input.
Background
In the field of image processing technology, image recognition and classification are common image processing tasks, and are processes for judging specific class labels to which given images belong by using technical means. The skin mirror image of the pigmentation skin lesion has a significant difference from a natural image, and the skin lesion image has differences of an imaging means, an image visual angle, a size and illumination, background noise interference and the like. Meanwhile, the classification task of skin disease images is a very difficult challenge because the skin disease images have a large variety and some images have small difference, and different images of the same disease type may have large difference due to the reasons.
In the field of research of skin disease image classification, a Convolutional Neural Network (CNN) is generally used for feature extraction and classification of images. However, the image classification model based on the conventional convolutional neural network only focuses on single-mode image input of an image data layer, and all features are extracted and mapped based on the image. In the clinical diagnosis process of skin diseases, along with the acquisition of image data of skin diseases, there are often many important text numerical data, and this part of data may come from medical record information, examination reports, records of diagnosis and treatment processes, and the like. Meanwhile, the image classification based on the traditional convolutional neural network can reduce the classification effect when the problems of multiple skin diseases, insignificant difference among individual diseases, unbalanced data of various diseases and the like are faced, and the text numerical data serving as supplementary information is added into the classification process, so that the classification accuracy is improved. Aiming at the problems, how to construct a multi-modal input data classification model capable of efficiently processing image and text values is one of key technologies for fully utilizing various medical data and improving the classification accuracy of skin disease images under the background of big data
Disclosure of Invention
The invention provides a skin disease image classification system based on multi-mode data input, which fully utilizes text supplementary data matched with image data to improve the accuracy of skin disease image classification, divides input data into image class data and text class data, respectively constructs an image feature extractor based on a Deep Transfer Convolutional Neural Network (DTCNN) and a text feature extractor based on a Multilayer Perceptron unit (M L P) after image data standardization and text data coding, and jointly inputs image features and text features to a comprehensive full-link layer by using a feature fusion method to output class labels of data, thereby realizing the classification task of skin disease images.
The invention is realized by the following technical scheme:
the invention relates to a skin disease image classification system based on multi-modal data input, which comprises: the system comprises a data preprocessing module, an image feature extraction module, a text parameter feature extraction module, a feature fusion module and a classification output module, wherein: the data preprocessing module receives multi-modal data comprising image data and text data corresponding to the image data, a pixel value matrix is obtained through preprocessing and is standardized, text parameters in the text data are encoded to obtain a numerical matrix, and a training sample set and a test sample set are generated through the processed data; the deep migration convolutional neural network in the image feature extraction module is trained by a training sample set and then the feature extraction of the image is carried out; the multi-layer perceptron unit in the text parameter characteristic extraction module is trained by a training sample set and then carries out the characteristic extraction of text parameters; the feature fusion module fuses the image features and the text parameter features through a feature fusion method and outputs the fused image features and the text parameter features to the classification output module together, and the comprehensive full-connection layer of the classification output module utilizes a softmax activation function to output the classification result of the image.
And applying the multi-mode input comprehensive classification model established by the training set data to a test set to obtain class labels of each skin disease image of the test set, and comparing the class labels with real labels to obtain the overall accuracy, average sensitivity and average accuracy of classification prediction.
The standardization refers to that: scaling the value X of the pixel value matrix to the interval of 0-1Wherein: xnormIs a normalized matrix of pixel values, X is an initial matrix of pixel values, muXIs the average, σ, of all pixel value matricesXIs the standard deviation of the matrix of all pixel values.
The coding means: the text data is converted into numerical data by mapping the text information to serial number information.
The ratio of the training sample set to the test sample set is preferably: 8:2.
The deep migration convolutional neural network is as follows: and migrating the models established in other task fields to the target task field by using a migration learning method, so that the target task is quickly and efficiently completed by using the adjusted migration model, and the difficulty of restarting establishing the models is avoided. The method specifically comprises the following steps: selecting a convolutional neural network model pre-trained on a natural image data set, changing a final output layer of the pre-trained model to enable the output layer to output the classification characteristics of the skin disease image, and finely adjusting the migration model by using the image data of the training set in the example to finally obtain the deep migration convolutional neural network.
The multi-layer perceptron unit is an Artificial Neural Network (ANN), and comprises:
the text parameter feature extraction device comprises an input layer for receiving text numerical data, an output layer for receiving text parameter features and a hidden layer arranged between the input layer and the output layer and used for extracting input data features, wherein the layers are fully connected through neurons, and the text parameter feature extraction device comprises: each layer of the network comprises a certain number of neurons, information processing and transmission are carried out on each layer through specific weights, threshold values and activation functions, the input layer receives numerical value input, and the output layer outputs specific results according to feature extraction and combination conditions.
The feature fusion is that: and combining the extracted image features and the text parameter features to obtain a comprehensive feature vector of the image and text data.
The comprehensive full-connection layer is as follows: the last few layers of the convolutional neural network, which are intended to map the distributed feature representation learned by convolutional layers to class space, act as a classifier. The synthesis means that: the fully connected layer of the present invention receives the integrated feature input containing image and text data for final classification.
The softmax activation function refers to: in the multi-classification task, the activation function included in the final fully-connected layer maps a plurality of neuron output values to (0,1), and the sum of the output values is 1. From the perspective of probability, for the output of each sample, the largest value (the largest probability) is selected as the class prediction value of the sample, specifically:wherein: e.g. of the typeiIs the output value of the ith neuron, whose value is the power i of e, ∑jejIs the sum of the outputs of all neurons in the full link layer, SiAnd predicting probability values for the output classes of the ith neurons of the full connection layer.
The total accuracy rate is as follows: according to the real category label of the sample, the ratio of the number of all samples with correct prediction to the number of the whole samples is as follows:wherein: TP is actually the positive type, and the model correctly predicts the number of the positive types; TN is actually the negative category, and the model correctly predicts the number of the negative categories; FP is actually negative category, and the number of the positive categories is predicted by the model error; FN is the actual positive class and the model mispredicts to the number of negative classes. In the multi-classification problem, for a specific class, the class is a positive class, and the non-class is a negative class.
The average sensitivity refers to: for each class, the proportion correctly predicted by the model as a positive class is sensitivity. The average sensitivity is then the average of all individual class sensitivitiesWherein: TP is actually positiveOtherwise, the model predicts correctly the number of positive categories; FN is the actual positive class and the model mispredicts to the number of negative classes.
The average accuracy is as follows: in the positive category of model prediction, the proportion that really belongs to the positive category is the accuracy. The average accuracy is the average of the accuracies of all the individual classesWherein: TP is actually the positive type, and the model correctly predicts the number of the positive types; FP is the actual negative class and the model mispredicts as the number of positive classes.
Technical effects
The invention integrally solves the problems that the accuracy is low due to the fact that when the traditional convolutional neural network is used for classifying the skin disease images, text numerical data matched with the skin disease images are ignored and only single image data is processed, and the skin disease types are large, the difference among individual disease types is not obvious, and the data of the disease types are unbalanced.
Compared with the prior art, the method comprehensively improves the classification accuracy of the skin disease images by utilizing the multi-mode input text numerical data: when skin disease image classification is carried out, skin disease image data and text data can be processed simultaneously, extraction of image features and text parameter features is carried out through establishing a depth migration convolutional neural network and a multilayer perceptron unit, information supplement and interaction existing among modal data are utilized, and the accuracy of skin disease image classification is effectively improved in a data feature fusion mode. The traditional convolution neural network based on single-mode image input ignores text data except images in the skin disease image classification process, and the data can often provide valuable reference information in the medical process.
Drawings
FIG. 1 is a schematic diagram of the system of the present invention;
FIG. 2 is a flow chart of skin condition image classification based on multi-modal data input;
FIG. 3 is a schematic diagram of an example of a class 7 dermatological sample in example;
FIG. 4 is a schematic structural diagram of a Transfer-ResNet50 feature extractor in the embodiment;
FIG. 5 is a schematic diagram of a multi-layered sensor unit according to an embodiment.
Detailed Description
As shown in fig. 1, a skin disease image classification system based on multi-modal data input includes: the system comprises a data preprocessing module, an image feature extraction module, a text parameter feature extraction module, a feature fusion module and a classification output module, wherein: the image characteristic extraction module and the text parameter characteristic extraction module comprehensively utilize skin disease image data and related text data, can receive image and text numerical value multi-modal data input and perform characteristic extraction and fusion, a depth migration convolutional neural network in the image characteristic extraction module effectively represents and extracts skin disease image characteristics, the multilayer perceptron unit in the text parameter characteristic extraction module is used for efficiently extracting text parameter characteristics, and the image characteristics and the text parameter characteristics are integrated through the characteristic fusion module, so that the classification accuracy of skin disease images is improved together.
The image feature extraction module adopts a Transfer learning method, selects a Residual error Network (ResNet) trained in advance based on a natural image data set, removes the last full connection layer of the original Residual error Network by transferring the Network structure and the weights of all layers of ResNet50, adds three batch normalization layers, one discarding layer and two full connection layers with the activation functions of Re L U, and establishes a deep Transfer convolutional neural Network Transfer-ResNet 50.
In the residual error network, the network weight of each layer of ResNet50 is used as an initial value during training, the first 10 layers of ResNet50 networks are frozen, namely, neuron parameters of the first 10 layers are not updated, the characteristic extraction is carried out on the skin disease image by using neuron parameters of 40 layers after the skin disease image data set is trained, and finally a 64-dimensional skin disease image characteristic vector is output.
The text parameter feature extraction module uses a multilayer perceptron unit to extract features of text parameters, the multilayer perceptron unit receives input of coded numerical vectors by building an input layer, two layers of hidden layers are provided, the number of neurons is 64 and 32 respectively, Re L U is used as an activation function, the number of neurons in a feature output layer is 4, Re L U is used as the activation function, and finally a 4-dimensional text parameter feature vector is output.
This example is illustrated by taking as an example a classification of a skin mirror image containing 7 types of pigmented skin lesions, which specifically includes the following steps:
the image data were 10015 skin mirror images containing 7 types of pigmented skin lesions, respectively: 327 solar keratosis images, 514 basal cell tumors, 1099 seborrheic keratosis images, 115 skin fibroma lesions, 6705 melanocyte nevi, 1113 melanoma images and 142 vascular skin lesions, wherein samples of all the categories are shown in FIG. 3;
the text data includes the number of the skin disease image, the skin lesion number, and the diagnosis confirmation type.
2.1) acquiring pixel value matrixes of three color channels of RGB of each image aiming at the skin disease image data:
X=[XR,XG,XB]and calculating the mean value mu according to the pixel value matrix of all the imagesXAnd standard deviation sigmaX. Using formulasNormalizing the matrix of pixel values for each input image to scale the matrix values to [0, 1%]。
2.2) encoding each item of data for text data, mapping character-type image numbers, skin lesion numbers, and diagnosis confirmation types to corresponding numerical serial numbers, thereby converting the character-type data into numerical data, whereby each input sample becomes a numerical vector.
2.3) in the processed image data and text data thereof, dividing the processed image data and text data into training sample sets according to the proportion of 80% for training classification prediction models; the remaining 20% of the proportion is the test set used for model final testing and results comparison.
Step 3, constructing a deep migration convolutional neural network feature extractor shown in fig. 4 by using a migration learning method, and extracting features of the skin disease image, wherein the method specifically comprises the following steps:
3.1) selecting a Residual Network (ResNet) trained in advance based on a natural image dataset, and constructing a feature extractor by taking ResNet50 with a 50-layer Network (49 layers of which are convolutional layers) as an example.
ResNet ensures that a deep network at least keeps the performance of a shallow network by introducing a residual learning mode and utilizing identity mapping, thereby overcoming the problems that the learning efficiency is low and the accuracy cannot be effectively improved due to deepening of the network depth. Due to the structural innovation of the residual error network, the pre-training model ResNet50 has good image feature learning capability and generalization capability.
3.2) removing the last full-connection layer by migrating the network structure and the weights of the ResNet50, only using for image feature extraction and not outputting classification categories, meanwhile, adding three batch normalization layers after the network layer of the ResNet50, one discarding layer and two full-connection layers with the activation function of Re L U, and establishing a deep migration convolutional neural network Transfer-ResNet50 as shown in FIG. 4.
The batch normalization layer is as follows: batch standardization and deep learning training skills. For the neurons of each hidden layer of the deep neural network, the input distribution is normalized to the standard normal distribution with the mean value of 0 and the variance of 1, so that the input data of each layer has the same distribution, and the learning process of the network is accelerated.
The discarding layer is as follows: in the method for preventing model overfitting, hidden layer neuron output is set to be 0 according to a certain probability p (in the example, p is 0.25) in the deep neural network training process, and the updating of parameters of the hidden layer neuron is stopped.
The Re L U activation function is a modified linear unit, and when x is less than 0, the output is 0, and when x is greater than 0, the output is x.
3.3) during training, using the network weight of each layer of ResNet50 as an initial value, freezing the first 10 layers of ResNet50 network, namely not updating the neuron parameters of the first 10 layers, using the neuron parameters of the 40 layers after the training of the skin disease image data set, reasonably using the good feature learning capability of ResNet50 to extract the features of the skin disease image, and finally outputting a 64-dimensional feature vector.
the method comprises the steps of constructing an input layer for receiving input of coded numerical vectors, constructing two hidden layers, wherein the neuron number is 64 and 32 respectively, Re L U is used as an activation function, the neuron number of a characteristic output layer is 4, Re L U is used as the activation function, and performing characteristic extraction on text parameters by a multilayer perceptron unit to output a 4-dimensional characteristic vector.
And 5, performing feature fusion based on the image features and the text parameter features obtained in the steps 3 and 4: using the fusion network layer, the 64-dimensional image feature vector and the 4-dimensional text parameter feature vector are combined to form a 68-dimensional comprehensive feature vector.
And 6, inputting the comprehensive characteristic vector obtained in the step 5 into a comprehensive full-connection layer, mapping the comprehensive characteristic vector to 7 dermatosis category spaces of the case by the comprehensive full-connection layer by using a softmax activation function, and outputting a final classification result and corresponding probability.
By adopting the steps, the image feature extractor based on the depth migration convolutional neural network and the text parameter feature extractor based on the multilayer perceptron unit are constructed, feature extraction can be simultaneously performed on the skin disease image data and the text data, the extracted image features and the text parameter features are fused, and finally the classification result of the skin disease image is output.
Based on the 7 types of skin disease images and the corresponding text parameters of the embodiment, the comprehensive classification model is trained on a training set, and the model classification performance is verified by using a test set. Meanwhile, in order to show the performance improvement of the method on the skin disease image classification task, the traditional convolutional neural network method is used for training and testing on the same skin disease image data set, the overall classification Accuracy (Accuracy), the average Sensitivity (Sensitivity) and the average Precision (Precision) of the two methods on the test set are compared, and the comparison result is shown in table 1.
TABLE 1 comparison of skin disease image classification results
As shown in table 1, for the skin disease image multi-classification problem, compared with the traditional convolutional neural network method, the method improves the classification accuracy by 6.94%; the sensitivity of the classification is improved by 16.80%; the accuracy of classification is improved by 21.22%. 3, classification evaluation indexes are obviously improved, and the classification performance of the skin disease image of the model is well improved on the basis of fully utilizing various modal data.
Through specific practical experiments, under the specific environment setting of a Python programming language and a tensrflow deep learning framework, the device/method is started/operated by parameters of an initial learning Rate L earning Rate being 0.001, a data set iteration number Epoch being 50 and a sample size Batchsize of each training being 32, and the obtained experimental data includes that the classification accuracy of 7 types of skin disease images is 85.62%, the sensitivity is 66.90% and the accuracy is 81.02%.
Compared with the prior art, the system constructs a skin disease image classification system capable of processing multi-modal data input, and specifically comprises the following steps: and performing feature extraction and fusion on the skin disease image data and the corresponding text parameter data through multi-modal data feature extraction. The limitation that the skin disease image classification only focuses on the image data is broken through, and an effective technical means is provided for comprehensive utilization of medical data. The image feature extraction module based on the deep migration convolutional neural network can well extract features of skin disease images, the text parameter feature extraction module based on the multilayer perceptron unit can efficiently extract text parameter features, the feature fusion module combines the two types of features, and the accuracy of skin disease image classification is improved by utilizing information supplement and interaction among modal data features. Compared with the traditional convolutional neural network method, the classification accuracy of the method on the 7-class skin disease images is improved by 6.94 percent; the sensitivity is improved by 16.80%; the accuracy is improved by 21.22%.
The foregoing embodiments may be modified in many different ways by those skilled in the art without departing from the spirit and scope of the invention, which is defined by the appended claims and all changes that come within the meaning and range of equivalency of the claims are therefore intended to be embraced therein.
Claims (9)
1. A system for classifying skin conditions based on multi-modal data input, comprising: the system comprises a data preprocessing module, an image feature extraction module, a text parameter feature extraction module, a feature fusion module and a classification output module, wherein: the data preprocessing module receives multi-modal data comprising image data and text data corresponding to the image data, a pixel value matrix is obtained through preprocessing and is standardized, text parameters in the text data are encoded to obtain a numerical matrix, and a training sample set and a test sample set are generated through the processed data; the deep migration convolutional neural network in the image feature extraction module is trained by a training sample set and then the feature extraction of the image is carried out; the multi-layer perceptron unit in the text parameter characteristic extraction module is trained by a training sample set and then carries out the characteristic extraction of text parameters; the feature fusion module fuses the image features and the text parameter features through a feature fusion method and outputs the fused image features and the text parameter features to the classification output module together, and a comprehensive full-connection layer of the classification output module utilizes a softmax activation function to realize classification result output;
the standardization refers to that: scaling the value X of the pixel value matrix to the interval of 0-1Wherein: xnormIs a normalized matrix of pixel values, X is an initial matrix of pixel values, muXIs the average, σ, of all pixel value matricesXIs the standard deviation of the matrix of all pixel values.
2. The skin condition image classification system according to claim 1, wherein the depth migration convolutional neural network is: and migrating the models established in other task fields to the target task field by using a migration learning method, so that the target task is quickly and efficiently completed by using the adjusted migration model, and the difficulty of restarting establishing the models is avoided. The method specifically comprises the following steps: selecting a convolutional neural network model pre-trained on a natural image data set, changing a final output layer of the pre-trained model to enable the output layer to output the classification characteristics of the skin disease image, and finely adjusting the migration model by using the image data of the training set in the example to finally obtain the deep migration convolutional neural network.
3. The skin condition image classification system according to claim 1, wherein the multi-layered perceptron unit is a feed-forward artificial neural network comprising: the text parameter feature extraction device comprises an input layer for receiving text numerical data, an output layer for receiving text parameter features and a hidden layer arranged between the input layer and the output layer and used for extracting input data features, wherein the layers are fully connected through neurons, and the text parameter feature extraction device comprises: each layer of the network comprises a certain number of neurons, information processing and transmission are carried out on each layer through specific weights, threshold values and activation functions, the input layer receives numerical value input, and the output layer outputs specific results according to feature extraction and combination conditions.
4. The skin condition image classification system according to claim 1, wherein the feature fusion is: and combining the extracted image features and the text parameter features to obtain a comprehensive feature vector of the image and text data.
5. The skin condition image classification system according to claim 1, wherein the comprehensive fully-connected layer is: the last few layers of the convolutional neural network, which are intended to map the distributed feature representation learned by convolutional layers to class space, act as a classifier. The synthesis means that: the fully connected layer of the present invention receives the integrated feature input containing image and text data for final classification.
6. The dermatologic image classification system according to claim 1, wherein the softmax activation function is: in the multi-classification task, the activation function included in the final fully-connected layer maps a plurality of neuron output values to (0,1), and the sum of the output values is 1. From the perspective of probability, for the output of each sample, the largest value (the largest probability) is selected as the class prediction value of the sample, specifically:wherein: e.g. of the typeiIs the output value of the ith neuron, whose value is the power i of e, ∑jejIs the sum of the outputs of all neurons in the full link layer, SiAnd predicting probability values for the output classes of the ith neurons of the full connection layer.
7. The skin condition image classification system according to claim 1, wherein the overall accuracy is: according to the real category label of the sample, the ratio of the number of all samples with correct prediction to the number of the whole samples is as follows:wherein: TP is actually the positive type, and the model correctly predicts the number of the positive types; TN is actually the negative category, and the model correctly predicts the number of the negative categories; FP is actually negative category, and the number of the positive categories is predicted by the model error; FN is the actual positive class and the model mispredicts to the number of negative classes. In the multi-classification problem, for a specific class, the class is a positive class, and the non-class is a negative class.
8. The dermatologic image classification system of claim 1, wherein the average sensitivity is: for each class, the proportion correctly predicted by the model as a positive class is sensitivity. The average sensitivity is then the average of all individual class sensitivitiesWherein: TP is actually the positive type, and the model correctly predicts the number of the positive types; FN is the actual positive class and the model mispredicts to the number of negative classes.
9. The skin condition image classification system according to claim 1, wherein the average accuracy is: in the positive category of model prediction, the proportion that really belongs to the positive category is the accuracy. The average accuracy is the average of the accuracies of all the individual classesWherein: TP is actually the positive type, and the model correctly predicts the number of the positive types; FP is the actual negative class and the model mispredicts as the number of positive classes.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010222477.XA CN111444960A (en) | 2020-03-26 | 2020-03-26 | Skin disease image classification system based on multi-mode data input |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010222477.XA CN111444960A (en) | 2020-03-26 | 2020-03-26 | Skin disease image classification system based on multi-mode data input |
Publications (1)
Publication Number | Publication Date |
---|---|
CN111444960A true CN111444960A (en) | 2020-07-24 |
Family
ID=71650776
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010222477.XA Pending CN111444960A (en) | 2020-03-26 | 2020-03-26 | Skin disease image classification system based on multi-mode data input |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111444960A (en) |
Cited By (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111931717A (en) * | 2020-09-22 | 2020-11-13 | 平安科技(深圳)有限公司 | Semantic and image recognition-based electrocardiogram information extraction method and device |
CN112133441A (en) * | 2020-08-21 | 2020-12-25 | 广东省人民医院 | Establishment method and terminal of MH post-operation fissure hole state prediction model |
CN112434718A (en) * | 2020-09-02 | 2021-03-02 | 武汉大学 | New coronary pneumonia multi-modal feature extraction fusion method and system based on depth map |
CN112784801A (en) * | 2021-02-03 | 2021-05-11 | 紫东信息科技(苏州)有限公司 | Text and picture-based bimodal gastric disease classification method and device |
CN112784919A (en) * | 2021-02-03 | 2021-05-11 | 华南理工大学 | Intelligent manufacturing multi-mode data oriented classification method |
CN112949757A (en) * | 2021-03-30 | 2021-06-11 | 太原理工大学 | Multi-modal heart failure recognition and classification method and device |
CN113469049A (en) * | 2021-06-30 | 2021-10-01 | 平安科技(深圳)有限公司 | Disease information identification method, system, device and storage medium |
CN113707309A (en) * | 2021-08-31 | 2021-11-26 | 平安科技(深圳)有限公司 | Disease prediction method and device based on machine learning |
CN113705630A (en) * | 2021-08-10 | 2021-11-26 | 南京邮电大学 | Skin lesion image classification method |
CN113707312A (en) * | 2021-09-16 | 2021-11-26 | 人工智能与数字经济广东省实验室(广州) | Blood vessel quantitative identification method and device based on deep learning |
CN114662033A (en) * | 2022-04-06 | 2022-06-24 | 昆明信息港传媒有限责任公司 | Multi-modal harmful link recognition based on text and image |
CN114724030A (en) * | 2022-04-06 | 2022-07-08 | 西安电子科技大学 | Polarized SAR terrain classification method based on contrast learning |
CN114863182A (en) * | 2022-05-23 | 2022-08-05 | 北京百度网讯科技有限公司 | Image classification method, and training method and device of image classification model |
CN115022464A (en) * | 2022-05-06 | 2022-09-06 | 中国联合网络通信集团有限公司 | Number processing method, system, computing device and storage medium |
CN115083005A (en) * | 2022-06-13 | 2022-09-20 | 广东省人民医院 | ROP image classification system and method based on deep learning |
CN116246176A (en) * | 2023-05-12 | 2023-06-09 | 山东建筑大学 | Crop disease detection method and device, electronic equipment and storage medium |
CN116580282A (en) * | 2023-07-12 | 2023-08-11 | 四川大学华西医院 | Neural network model-based pressure injury staged identification system and storage medium |
WO2023178798A1 (en) * | 2022-03-25 | 2023-09-28 | 平安科技(深圳)有限公司 | Image classification method and apparatus, and device and medium |
CN112133441B (en) * | 2020-08-21 | 2024-05-03 | 广东省人民医院 | Method and terminal for establishing MH postoperative crack state prediction model |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102017010397A1 (en) * | 2017-11-09 | 2019-05-09 | FotoFinder Systems GmbH | Method for evaluating the skin pattern of a human in the context of skin cancer screening and device for its operation |
CN109840280A (en) * | 2019-03-05 | 2019-06-04 | 百度在线网络技术(北京)有限公司 | A kind of file classification method, device and computer readable storage medium |
CN109919928A (en) * | 2019-03-06 | 2019-06-21 | 腾讯科技(深圳)有限公司 | Detection method, device and the storage medium of medical image |
CN109935320A (en) * | 2019-03-21 | 2019-06-25 | 颐保医疗科技(上海)有限公司 | A kind of common skin diseases network assistance diagnostic system |
CN110298383A (en) * | 2019-05-28 | 2019-10-01 | 中国科学院计算技术研究所 | Pathological classification method and system based on multi-modal deep learning |
CN110755045A (en) * | 2019-10-30 | 2020-02-07 | 湖南财政经济学院 | Skin disease comprehensive data analysis and diagnosis auxiliary system and information processing method |
-
2020
- 2020-03-26 CN CN202010222477.XA patent/CN111444960A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102017010397A1 (en) * | 2017-11-09 | 2019-05-09 | FotoFinder Systems GmbH | Method for evaluating the skin pattern of a human in the context of skin cancer screening and device for its operation |
CN109840280A (en) * | 2019-03-05 | 2019-06-04 | 百度在线网络技术(北京)有限公司 | A kind of file classification method, device and computer readable storage medium |
CN109919928A (en) * | 2019-03-06 | 2019-06-21 | 腾讯科技(深圳)有限公司 | Detection method, device and the storage medium of medical image |
CN109935320A (en) * | 2019-03-21 | 2019-06-25 | 颐保医疗科技(上海)有限公司 | A kind of common skin diseases network assistance diagnostic system |
CN110298383A (en) * | 2019-05-28 | 2019-10-01 | 中国科学院计算技术研究所 | Pathological classification method and system based on multi-modal deep learning |
CN110755045A (en) * | 2019-10-30 | 2020-02-07 | 湖南财政经济学院 | Skin disease comprehensive data analysis and diagnosis auxiliary system and information processing method |
Non-Patent Citations (3)
Title |
---|
刘玉良: "《深度学习》", 30 November 2019, 西安电子科技大学出版社 * |
张朝阳: "《深入浅出 工业机器学习算法详解与实战》", 31 January 2020, 机械工业出版社 * |
李周平: "《网络数据爬取与分析实务》", 30 September 2018, 上海交通大学出版社 * |
Cited By (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112133441A (en) * | 2020-08-21 | 2020-12-25 | 广东省人民医院 | Establishment method and terminal of MH post-operation fissure hole state prediction model |
CN112133441B (en) * | 2020-08-21 | 2024-05-03 | 广东省人民医院 | Method and terminal for establishing MH postoperative crack state prediction model |
CN112434718A (en) * | 2020-09-02 | 2021-03-02 | 武汉大学 | New coronary pneumonia multi-modal feature extraction fusion method and system based on depth map |
CN112434718B (en) * | 2020-09-02 | 2023-04-18 | 武汉大学 | New coronary pneumonia multi-modal feature extraction fusion method and system based on depth map |
WO2021159751A1 (en) * | 2020-09-22 | 2021-08-19 | 平安科技(深圳)有限公司 | Semantic and image recognition-based electrocardiography information extraction method and apparatus, computer device, and storage medium |
CN111931717A (en) * | 2020-09-22 | 2020-11-13 | 平安科技(深圳)有限公司 | Semantic and image recognition-based electrocardiogram information extraction method and device |
CN112784919A (en) * | 2021-02-03 | 2021-05-11 | 华南理工大学 | Intelligent manufacturing multi-mode data oriented classification method |
CN112784801A (en) * | 2021-02-03 | 2021-05-11 | 紫东信息科技(苏州)有限公司 | Text and picture-based bimodal gastric disease classification method and device |
CN112784919B (en) * | 2021-02-03 | 2023-09-05 | 华南理工大学 | Classification method for intelligent manufacturing multi-mode data |
CN112949757A (en) * | 2021-03-30 | 2021-06-11 | 太原理工大学 | Multi-modal heart failure recognition and classification method and device |
CN113469049A (en) * | 2021-06-30 | 2021-10-01 | 平安科技(深圳)有限公司 | Disease information identification method, system, device and storage medium |
CN113705630A (en) * | 2021-08-10 | 2021-11-26 | 南京邮电大学 | Skin lesion image classification method |
CN113705630B (en) * | 2021-08-10 | 2023-10-13 | 南京邮电大学 | Skin lesion image classification method |
CN113707309A (en) * | 2021-08-31 | 2021-11-26 | 平安科技(深圳)有限公司 | Disease prediction method and device based on machine learning |
CN113707312A (en) * | 2021-09-16 | 2021-11-26 | 人工智能与数字经济广东省实验室(广州) | Blood vessel quantitative identification method and device based on deep learning |
WO2023178798A1 (en) * | 2022-03-25 | 2023-09-28 | 平安科技(深圳)有限公司 | Image classification method and apparatus, and device and medium |
CN114724030A (en) * | 2022-04-06 | 2022-07-08 | 西安电子科技大学 | Polarized SAR terrain classification method based on contrast learning |
CN114662033A (en) * | 2022-04-06 | 2022-06-24 | 昆明信息港传媒有限责任公司 | Multi-modal harmful link recognition based on text and image |
CN114662033B (en) * | 2022-04-06 | 2024-05-03 | 昆明信息港传媒有限责任公司 | Multi-mode harmful link identification based on text and image |
CN115022464A (en) * | 2022-05-06 | 2022-09-06 | 中国联合网络通信集团有限公司 | Number processing method, system, computing device and storage medium |
CN114863182A (en) * | 2022-05-23 | 2022-08-05 | 北京百度网讯科技有限公司 | Image classification method, and training method and device of image classification model |
CN115083005A (en) * | 2022-06-13 | 2022-09-20 | 广东省人民医院 | ROP image classification system and method based on deep learning |
CN116246176A (en) * | 2023-05-12 | 2023-06-09 | 山东建筑大学 | Crop disease detection method and device, electronic equipment and storage medium |
CN116246176B (en) * | 2023-05-12 | 2023-09-19 | 山东建筑大学 | Crop disease detection method and device, electronic equipment and storage medium |
CN116580282A (en) * | 2023-07-12 | 2023-08-11 | 四川大学华西医院 | Neural network model-based pressure injury staged identification system and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111444960A (en) | Skin disease image classification system based on multi-mode data input | |
CN111191660B (en) | Colon cancer pathology image classification method based on multi-channel collaborative capsule network | |
CN110532900B (en) | Facial expression recognition method based on U-Net and LS-CNN | |
CN110755045B (en) | Skin disease comprehensive data analysis and diagnosis auxiliary system | |
CN111144448A (en) | Video barrage emotion analysis method based on multi-scale attention convolutional coding network | |
CN113314205B (en) | Efficient medical image labeling and learning system | |
CN113158875A (en) | Image-text emotion analysis method and system based on multi-mode interactive fusion network | |
CN111738363B (en) | Alzheimer disease classification method based on improved 3D CNN network | |
CN105654141A (en) | Isomap and SVM algorithm-based overlooked herded pig individual recognition method | |
CN109145944B (en) | Classification method based on longitudinal three-dimensional image deep learning features | |
CN115100709B (en) | Feature separation image face recognition and age estimation method | |
CN113012811B (en) | Traditional Chinese medicine syndrome diagnosis and health evaluation method combining deep convolutional network and graph neural network | |
CN115761757A (en) | Multi-mode text page classification method based on decoupling feature guidance | |
CN114176607B (en) | Electroencephalogram signal classification method based on vision transducer | |
CN112668486A (en) | Method, device and carrier for identifying facial expressions of pre-activated residual depth separable convolutional network | |
CN112863650A (en) | Cardiomyopathy identification system based on convolution and long-short term memory neural network | |
CN116994295A (en) | Wild animal category identification method based on gray sample self-adaptive selection gate | |
CN116630289A (en) | Brain glioma CDKN2A/B gene state classification prediction method based on multi-mode MRI | |
CN108846327B (en) | Intelligent system and method for distinguishing pigmented nevus and melanoma | |
CN113343770B (en) | Face anti-counterfeiting method based on feature screening | |
CN113887504B (en) | Strong-generalization remote sensing image target identification method | |
CN115223021A (en) | Visual question-answering-based fruit tree full-growth period farm work decision-making method | |
CN115439683A (en) | Attention mechanism-based leukocyte fine-granularity classification method | |
CN115063374A (en) | Model training method, face image quality scoring method, electronic device and storage medium | |
Dixit et al. | An Improved Approach To Classify Plant Disease Using CNN And Random Forest |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20200724 |