CN110189305A - A kind of multitask tongue picture automatic analysis method - Google Patents

A kind of multitask tongue picture automatic analysis method Download PDF

Info

Publication number
CN110189305A
CN110189305A CN201910397988.2A CN201910397988A CN110189305A CN 110189305 A CN110189305 A CN 110189305A CN 201910397988 A CN201910397988 A CN 201910397988A CN 110189305 A CN110189305 A CN 110189305A
Authority
CN
China
Prior art keywords
multitask
tongue
deep neural
neural network
sample
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910397988.2A
Other languages
Chinese (zh)
Other versions
CN110189305B (en
Inventor
严壮志
沈婷
蒋皆恢
胡俊炜
张瑶雯
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Shanghai for Science and Technology
Original Assignee
University of Shanghai for Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Shanghai for Science and Technology filed Critical University of Shanghai for Science and Technology
Priority to CN201910397988.2A priority Critical patent/CN110189305B/en
Publication of CN110189305A publication Critical patent/CN110189305A/en
Application granted granted Critical
Publication of CN110189305B publication Critical patent/CN110189305B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/0059Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
    • A61B5/0082Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence adapted for particular medical purposes
    • A61B5/0088Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence adapted for particular medical purposes for oral or dental tissue
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/48Other medical applications
    • A61B5/4854Diagnosis based on concepts of traditional oriental medicine
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • G06F18/2155Generating training patterns; Bootstrap methods, e.g. bagging or boosting characterised by the incorporation of unlabelled data, e.g. multiple instance learning [MIL], semi-supervised techniques using expectation-maximisation [EM] or naïve labelling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/088Non-supervised learning, e.g. competitive learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/03Recognition of patterns in medical or anatomical images

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • General Physics & Mathematics (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Medical Informatics (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Surgery (AREA)
  • Computational Linguistics (AREA)
  • Veterinary Medicine (AREA)
  • Computing Systems (AREA)
  • Public Health (AREA)
  • Mathematical Physics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Animal Behavior & Ethology (AREA)
  • Pathology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Alternative & Traditional Medicine (AREA)
  • Quality & Reliability (AREA)
  • Radiology & Medical Imaging (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Dentistry (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a kind of multitask tongue picture automatic analysis methods, comprising the following steps: all images are divided into strong tally set and non label set according to the presence or absence of label by acquisition tongue picture picture;Use the multitask deep neural network of strong two kinds of the tally set training different frameworks;Use the multitask deep neural network of two kinds of the no label data alternative optimization different frameworks;Tongue picture picture is input to two trained multitask deep neural networks, the two output results analyzed averagely are last tongue picture multitask analysis result.A kind of multitask tongue picture automatic analysis method disclosed by the invention, it is for limitation and defect present in the lingual diagnosis analytical technology currently based on artificial intelligence, we attempt using semi-supervised learning method, in the case where possessing has label tongue picture data on a small quantity, using others without label data, the feature for learning tongue picture, improves the nicety of grading and generalization ability of model.

Description

A kind of multitask tongue picture automatic analysis method
Technical field
The present invention relates to intelligent diagnostics field more particularly to a kind of multitask tongue picture automatic analysis methods.
Background technique
Traditional Chinese Medicine possesses thousands of years history in China, mainly passes through the four methods of diagnosis and carries out to the illness of physical therapy person Judgement is being cured the disease and there is significant effect in the field of taking care of health.As a part of observation, TCM tongue diagnosis also becomes tcm diagnosis One important evidence.The presence or absence of the shape of tongue, color, crackle and the presence or absence of the depth, indentation can embody disease to a certain extent The property of disease, the depth of patient's condition, the prosperity and decline of qi and blood.Supply falls short of demand by traditional Chinese physician, and with based on machine learning, deep learning Image recognition technology computer field flourish, cause the intelligent analysis of TCM tongue diagnosis to have become trend.
The determination method and system of the TCM tongue diagnosis model based on convolutional Neural metanetwork have been invented (specially by University of Science & Technology, Beijing Benefit ZL 201610300494.4).This method by adjusting every layer of neuron of convolutional Neural metanetwork number, using on tongue Image is trained multiple and different convolutional Neural metanetworks adjusted as training set respectively, determines multiple and different structures The parameter of convolutional Neural metanetwork finally chooses the wherein highest network mould of accuracy rate of diagnosis using test set as input Type.
Tongue object detection method (patent No. ZL under a kind of open environment has invented in Xiamen University 201610300494.4).This method acquires image first, then carries out linear transformation to the RGB component of image and realizes color school Just.Then cut zone is sought into union after being split using distinct methods to image.Then using the texture of tongue body after segmentation Information classifies to tongue picture as feature.
A kind of tongue fur constitution discriminating conduct (publication number CN based on deep neural network has invented in South China Science & Engineering University 106683087 A).The tongue picture of acquisition is first carried out the normalization of color space and the modulation of size by this method, then sharp It is further accurately fixed with a kind of fast convolution neural network Preliminary detection tongue fur position based on region, then with neural network algorithm Position tongue fur position extracts tongue fur feature using convolutional neural networks algorithm and recurrent neural network algorithm, finally utilizes softmax Classifier or support vector machine classifier carry out tongue fur classification of TCM constitution.
It is (open that a kind of Chinese medicine tongue color coating colour automatic analysis method based on convolutional neural networks has been invented by Beijing University of Technology Number 107330889 A of CN).This method acquires tongue image data by Tongue instrument, realizes tongue fur using K mean cluster With the separation of tongue nature, and by Traditional Chinese Medicine experts demarcate tongue color coating colour.Convolutional neural networks are constructed respectively to tongue color data set and coating colour Data set is trained, and obtains tongue color disaggregated model and coating colour disaggregated model.
A kind of tongue picture constitution neural network based discrimination side has been invented by scientific and technological (Shanghai) Co., Ltd of little Wu health Method and equipment (109199334 A of publication number CN).This method trains 1 tongue picture picture by deep neural network algorithm and knows Other model, 5 tongue picture Feature Selection Models and 1 tongue picture constitution model.When predicting test set, pass through tongue picture figure first Piece identification model judges whether it is tongue picture, if tongue picture then utilizes 5 kinds of tongue picture Feature Selection Models to extract tongue nature face respectively Color, tongue as form, whether there is or not five kinds of tongue fur, thickness of the tongue coating, tongue fur color features.5 kinds of features of extraction are inputted into tongue picture constitution mould Type obtains user and corresponds to constitution.
Above based on artificial intelligence in the analysis method of tongue picture, just for some feature of tongue picture, such as tongue Shape, tongue color, coating colour etc. construct single analysis model, although in 109199334 A of publication number CN, while considering tongue Five kinds of features of elephant, but the patent is still to construct five kinds of different identification models for five kinds of features.And in practical application In, traditional Chinese physician would generally multiple features to tongue picture carry out discriminatory analysis simultaneously, so intelligently to multiple features of tongue picture It is necessary for carrying out analysis simultaneously.
Since supply falls short of demand by traditional Chinese physician, although tongue picture data is caused to be easy to obtain, there is the figure of complete diagnosis information As data, it to be more difficult acquisition that we term it strong label datas, and diagnostic message missing or no data, we are referred to as For no label data, it is often easier to acquisition.Semi-supervised learning can allow disaggregated model not depend on as a kind of learning method External world's interaction automatically promotes learning performance using unmarked sample, can make up exemplar not foot straps to a certain extent The defect come, avoids disaggregated model since the very little bring accuracy rate of training data is low, the problem of Generalization Capability difference.And it is based on Semi-supervised image classification algorithms have also been widely applied in the classification problem of natural image, but in the classification problem of tongue picture In be still to be studied.
A kind of multiclass image classification method (publication number CN based on semi-supervised extreme learning machine has invented in Southeast China University 104992184 B).The sample that unlabeled exemplars are concentrated put back to the training subset that resampling constitutes difference, then with Semi-supervised extreme learning machine model is respectively trained in marked training sample, is taken by the output summation of extreme learning machine corresponding node It is average, it takes most uncertain sample manually to be marked from unlabeled exemplars concentration and is transferred in marked training set, again Sorter model is updated, until iteration terminates, solving image classification in the related technology, there are classification accuracy rate is low and study speed Low problem is spent, has established certain basis for accurate, quick, stable image classification.
Zhongshan University is it is also proposed that a kind of image classification method (publication number CN based on active semi-supervised learning 109376796A), invention random selection part marker samples and all unmarked samples, for semi-supervised in training pattern Then dictionary learning component is introduced into a user to mark the full and accurate sample in unmarked sample, is added to the data set of label In, for the Active Learning component in training pattern, the iteration that repeats the above steps more new model is until algorithm is finally restrained or reached To a certain the number of iterations.The invention combination semi-supervised learning and Active Learning, effectively utilize all training datas, improve calculation The performance of method model.
Guangdong University of Technology discloses a kind of scene image mark side based on Active Learning and multi-tag multi-instance learning Method (105117429 A of publication number CN), the invention utilize active learning strategies, by the confidence level of computation model, select not true Qualitative maximum is without label image, and then expert manually marks the image, while guaranteeing disaggregated model accuracy, greatly Reduce the scene image quantity for needing manually to mark greatly, to reduce mark cost.Meanwhile image is converted into multi-tag More sample datas make image complexity semanteme obtain reasonable representation, improve the accuracy of image labeling.
Summary of the invention
To sum up, it has been found that existing lingual diagnosis technology has following defect: (1) being directed in the existing analysis method to tongue picture, It is to analyze some individual feature of tongue picture, that is, constructs single model of modal analysis and tongue picture feature is analyzed, But tongue picture may include tongue color, ligulate, tongue mind and coating nature etc..So individually signature analysis has limitation, and do not account for Interdependence between each feature;(2) due to there is label data amount few, lead to Analysis of Lingual Picture model generalization performance difference and standard The low problem of exactness, (utilization rate of data is low, can not largely apply in existing Analysis of Lingual Picture model without label data). It is to be solved by this invention for limitation and defect present in the above-mentioned lingual diagnosis analytical technology currently based on artificial intelligence Technical problem is: we attempt using semi-supervised learning method, in the case where possessing has label tongue picture data on a small quantity, utilize Others learn the feature of tongue picture without label data, improve the nicety of grading and generalization ability of model.Meanwhile we intend adopting The analysis of the different characteristic of tongue picture is obtained simultaneously with single disaggregated model as a result, simplified model, saves the training time of model.
To achieve the above object, the present invention provides a kind of multitask tongue picture automatic analysis methods, comprising the following steps:
Tongue picture picture is acquired, all images are divided into strong tally set and non label set according to the presence or absence of label;
Use the multitask deep neural network of strong two kinds of the tally set training different frameworks;
Use the multitask deep neural network of two kinds of the no label data alternative optimization different frameworks;
Tongue picture picture is input to two trained multitask deep neural networks, two outputs analyzed As a result averagely as last tongue picture multitask analysis result.
Using the multitask deep neural network of strong two kinds of the tally set training different frameworks, handed over using no label data For the multitask deep neural network of two kinds of optimization different frameworks, following steps are specifically included by the tongue picture in the strong tally set Sample set is randomly divided into two parts, is denoted as tally set S1 and tally set S2;Unlabeled exemplars collection is set, U is denoted as;Define two skies Collection, is denoted as A1 and A2, and A1 comes from for storing the unlabeled exemplars set from depth residual error network, A2 for storing The high unlabeled exemplars set of the confidence level of GoogleNet network;Two empty sets are defined, are denoted as B1 and B2, B1 is for storing nothing Exemplar and its confidence level generated by depth residual error network, B2 is for storing unlabeled exemplars and its passing through The confidence level that GoogleNet network generates;Define two sorter models, the respectively disaggregated model based on depth residual error network The disaggregated model for being denoted as F1 and the deep neural network based on GoogleNet is denoted as F2;Define counter i.
Further, the multitask deep neural network includes the multitask disaggregated model note based on depth residual error network For F1 and the multitask disaggregated model of the deep neural network based on GoogleNet is denoted as F2.
Further, tongue picture is input to two trained multitask deep neural networks, two analyzed A output result is averagely last tongue picture multitask analysis as a result, specifically including: more based on depth residual error network Business disaggregated model F1 is analyzed and predicted tongue color, ligulate, the tongue mind of the tongue picture picture, coating nature, and is based on GoogleNet Deep neural network multitask disaggregated model F2 to the tongue color, ligulate, tongue of the tongue picture picture mind, coating nature carry out analysis and Prediction.Further, using the multitask deep neural network of two kinds of the no label data alternative optimization different frameworks Include: select multitask disaggregated model F1 described in the confidence sample optimization that no label data is concentrated based on depth residual error network and Select the multitask disaggregated model based on GoogleNet deep neural network described in the confidence sample optimization that no label data is concentrated F2。
Further, the multitask based on depth residual error network described in the confidence sample optimization that no label data is concentrated is selected Disaggregated model F1, specifically includes:
1. obtaining disaggregated model F1 ' using tally set S1 multitask deep neural network F1;
2. integrating the sample for selecting size in U as n from unlabeled exemplars at random, it is denoted as u={ u1,u2,u3…un};
3. predicting using disaggregated model F1 ' all samples in sample u, prediction label F ' is obtained1(u);
4. enabling counter i=1;
5. using S1 ∪ (ui,F′1(ui)) data set multitask deep neural network F1 obtains F "1
6. using F '1With F "1Tally set S1 is predicted, sample u is calculated according to confidence calculations formula (1)iIt is corresponding Level of confidence △1i:
Wherein, xjTo there is label data collection S1In sample, yjTo there is label data xjCorresponding true tag.
7. recording uiConfidence level, be denoted as (ui,△1i), it stores in set B1;
8.i=i+1, judgement execute step 5 and otherwise perform the next step if i≤n;
9. selecting the highest sample of level of confidence in set B1, it is denoted as a, takes out sample a and F '1(a), it is put into empty set A2 In.
Further, the confidence sample optimization of no label data concentration is selected based on GoogleNet deep neural network Multitask disaggregated model F2, specifically includes:
1. obtaining F2 ' using tally set S2 multitask deep neural network F1;
2. integrating the sample for selecting size in U as n from unlabeled exemplars at random, it is denoted as u={ u1,u2,u3…un};
3. predicting using disaggregated model F2 ' all samples in u, prediction label F ' is obtained2(u);
4. enabling counter i=1;
5. using S2 ∪ (ui,F′2(ui)) data set multitask deep neural network F2 obtains F "2
6. using F '2With F "2Tally set S2 is predicted, calculates sample u according to confidence calculations formula (2)iIt is corresponding Level of confidence △2i:
Wherein, xjTo there is label data collection S2In sample, yjTo there is label data xjCorresponding true tag;
7. recording uiConfidence level, be denoted as (ui,△2i), it stores in set B2;
8.i=i+1, judgement, if i≤n, otherwise return step 5 performs the next step;
9. selecting the highest sample of level of confidence in set B2, it is denoted as b, takes out sample b and F '2(b), it is put into A1.
Further, using the multitask depth nerve net of two kinds of the no label data alternative optimization different frameworks Network is further comprising the steps of:
1. more new data S1=S1 ∪ A2, S2=S2 ∪ A1;
2. judging whether S1 and S2 changes, do not change, exit, otherwise, is chosen from unlabeled exemplars collection U at random again Selecting size is the sample of n, multitask disaggregated model F1 of the re -training based on depth residual error network and selects no label data collection In confidence sample optimization described in the multitask disaggregated model F2 based on GoogleNet deep neural network, until set S1 and S2 does not change.
Technical effect
1, the invention discloses a kind of analysis method for tongue picture, a kind of multitask tongue picture automatic analysis method is used Single deep neural network disaggregated model substitutes tradition sofmax function by using sigmoid activation primitive, and uses two It is worth cross entropy and realizes that ligulate, tongue color, tongue mind and coating nature are analyzed in use simultaneously as loss function, and provides analysis knot Fruit.Analysis system exports the ligulate of different tongue pictures, tongue color, tongue mind and coating nature information as a whole, it is contemplated that different special Correlation between sign.
2, the present invention uses the semi-supervised learning method based on coorinated training, constructs the neural network of two different structures, By way of screening confidence sample mutually, no label data is made full use of, is solved due to there is label data amount to lead to model less Generalization Capability difference and the low problem of accuracy.
It is described further below with reference to technical effect of the attached drawing to design of the invention, specific structure and generation, with It is fully understood from the purpose of the present invention, feature and effect.
Detailed description of the invention
Fig. 1 is a kind of flow diagram of multitask tongue picture automatic analysis method of a preferred embodiment of the invention;
Fig. 2 is that the use residual error network of a preferred embodiment of the invention carries out feature extraction and multitask point to tongue picture The network architecture diagram of class;
Fig. 3 is that a preferred embodiment of the invention using GoogleNet network carries out feature extraction and more to tongue picture The network architecture diagram of classification of task.
Specific embodiment
As shown in Figure 1, the embodiment of the invention provides a kind of multitask tongue picture automatic analysis methods, comprising the following steps:
Step 100, tongue picture picture is acquired, all images according to the presence or absence of label are divided into strong tally set and without label Collection;
Step 200, using the multitask deep neural network of strong two kinds of the tally set training different frameworks;
Step 300, using the multitask depth nerve net of two kinds of the no label data alternative optimization different frameworks Network;
Step 400, tongue picture picture is input to two trained multitask deep neural networks, analyzed The averagely as last tongue picture multitask analysis result of two output results.It is lacked for present in current Analysis of Lingual Picture technology Point, the present invention provide a kind of multitask tongue picture automatic analysis method, efficiently using no label data, for solving due to there is label Data volume leads to Analysis of Lingual Picture model generalization performance difference and the low problem of accuracy less.
The present invention realizes the multiple characteristic differentiation classification of tongue picture using improved deep neural network, by changing network end Activation primitive and loss function complete tongue picture feature multitask output, in addition, using based on the semi-supervised of coorinated training Method will be added in training pattern without label data, increase model data amount, solution conventional model generalization ability is poor, accurately The low problem of rate.That is the scheme of the embodiment of the present invention is the semi-supervised method based on coorinated training, is believed using containing multiple features The tongue picture data of breath train multitask deep neural network model end to end, obtain dividing for ligulate, tongue color, tongue mind and coating nature Class result.Wherein, using the semi-supervised learning method based on coorinated training, residual error neural network and GoogleNet are used respectively Two kinds of disaggregated models from different angles classify to tongue picture data, select think that believable unlabeled exemplars are added to mutually In training set, a kind of complementation is formed, improves the nicety of grading of entire disaggregated model.Meanwhile by introduce new activation primitive and Traditional single task neural network is changed to that ligulate, tongue color, tongue mind, the multitask of coating nature feature can be exported simultaneously by loss function Neural network.
It will be explained in detail a kind of the detailed based on semi-supervised multitasked algorithm of multitask tongue picture automatic analysis method below Step:
Initialization: the tongue picture data containing multiple characteristic informations are randomly divided into two parts, are denoted as tally set S1 and tally set S2;By unlabeled exemplars collection, it is denoted as U;Two empty sets are defined, are denoted as A1 and A2, A1 is for storing from depth residual error network Unlabeled exemplars set, A2 is for storing the high unlabeled exemplars set of the confidence level from GoogleNet network;Define two Empty set is denoted as B1 and B2, and B1 is used for the confidence level storing unlabeled exemplars and its generating by depth residual error network, and B2 is for depositing The confidence level putting unlabeled exemplars and its being generated by GoogleNet network;Two sorter models are defined, depth residual error is based on The disaggregated model of network is denoted as F1, and the deep neural network based on GoogleNet is denoted as F2;Define counter i;
Step 1 obtains F1 ' using tally set S1 multitask deep neural network F1;
Step 2 obtains F2 ' using tally set S2 multitask deep neural network F1;
Step 3 integrates the sample that size is selected in U as n from unlabeled exemplars at random, is denoted as u={ u1,u2,u3…un};
Step 4 predicts all samples in u using disaggregated model F1 ', obtains prediction label F '1(u);
Step 5, another counter i=1;
Step 6 uses S1 ∪ (ui,F′1(ui)) data set multitask deep neural network F1 obtains F "1
Step 7 uses F '1With F "1Tally set S1 is predicted, sample u is calculated according to confidence calculations formula (1)i Corresponding level of confidence △1i:
Wherein, xjTo there is label data collection S1In sample, yjTo there is label data xjCorresponding true tag.
Record uiConfidence level, be denoted as (ui,△1i), it stores in set B1;
Step 8, i=i+1, judgement, if i≤n, otherwise return step six performs the next step;
Step 9 selects the highest sample of level of confidence in set B1, is denoted as a, takes out sample a and F '1(a), it is put into In A2;
Step 10 predicts all samples in u using disaggregated model F2 ', obtains prediction label F '2(u);
Step 11, another counter i=1;
Step 12 uses S2 ∪ (ui,F′2(ui)) data set multitask deep neural network F2 obtains F "2
Step 13 uses F '2With F "2Tally set S2 is predicted, calculates sample according to confidence calculations formula (2) uiCorresponding level of confidence △2i:
Wherein, xjTo there is label data collection S2In sample, yjTo there is label data xjCorresponding true tag;
Record uiConfidence level, be denoted as (ui,△2i), it stores in set B2;
Step 14: i=i+1, judgement, if i≤n, otherwise return step 12 performs the next step;
Step 15 selects the highest sample of level of confidence in set B2, is denoted as b, takes out sample b and F '2(b), it puts Enter in A1;
Step 10 six, more new data S1=S1 ∪ A2, S2=S2 ∪ A1;
Step 10 seven, judges whether S1 and S2 changes, and does not change, and exits, otherwise, return step three.
For two sorter models F1 and F2 employed in the present invention, the present invention is respectively adopted based on depth residual error net Network and GoogleNet based on Inception module, go to train using data from two different angles, reach complementary effect Fruit improves the precision of entire model.Classifier F1 is based on depth residual error network and carries out multi-task learning and classification to tongue picture.Relatively For traditional Analysis of Lingual Picture technology, the neural network of the deep layer based on deep learning is meant to extract different stage Image information can directly input information be detoured by introducing residual block in deep layer network and pass to output, in protection letter While the integrality of breath, the disappearance of cause gradient or gradient explosion that deep layer network is easy to cause are avoided the problem that.Classifier F1 Using including 14 layer networks, the network architecture of 6 residual blocks.Whole network is all made of the convolution kernel of lesser 3x3.It specifically includes 1 input layer, 12 convolutional layers, 1 is fully connected layer and 1 output layer.Wherein first, in two residual blocks, convolution kernel Number be 32, third, in four residual blocks, the number of convolution kernel is 64, the 5th, in six residual blocks, convolution kernel Number is 128.
Comparison-of-pair sorting's device F2 carries out multitask classification to tongue picture using the GoogleNet based on Inception module.The net Network obtains the image low-level features of different dimensions type using different filter operators (convolution, convergence of various sizes etc.), and These low-level features are combined later, next layer network is allowed independently to select useful input, so that different rulers are arrived in study The characteristics of image of degree.Classifier F2 uses the GoogleNet network architecture containing 4 Inception modules.Each There is 1x1 in Inception module, tri- kinds of different filter operators of 3x3,5x5 are used to extract the characteristics of image of different scale.For The multitask output for realizing classifier F1 and F2 activates replacement to pass in the last layer of deep neural network using sigmoid The SoftMax activation of system network end-point:
Meanwhile two-value cross entropy is as classification cross entropy loss function:
Wherein C indicates cost, and x indicates that sample, y indicate physical tags value, and a indicates network output valve, and n indicates the total of sample Number.
The preferred embodiment of the present invention has been described in detail above.It should be appreciated that those skilled in the art without It needs creative work according to the present invention can conceive and makes many modifications and variations.Therefore, all technologies in the art Personnel are available by logical analysis, reasoning, or a limited experiment on the basis of existing technology under this invention's idea Technical solution, all should be within the scope of protection determined by the claims.

Claims (8)

1. a kind of multitask tongue picture automatic analysis method, which comprises the following steps:
Tongue picture picture is acquired, all images are divided into strong tally set and non label set according to the presence or absence of label;
Use the multitask deep neural network of strong two kinds of the tally set training different frameworks;
Use the multitask deep neural network of two kinds of the no label data alternative optimization different frameworks;
Tongue picture picture is input to two trained multitask deep neural networks, the two output results analyzed Be averagely last tongue picture multitask analysis result.
2. a kind of multitask tongue picture automatic analysis method as described in claim 1, which is characterized in that the multitask depth mind It include that the multitask disaggregated model based on depth residual error network is denoted as F1 and based on the deep neural network of GoogleNet through network Multitask disaggregated model be denoted as F2.
3. a kind of multitask tongue picture automatic analysis method as claimed in claim 2, which is characterized in that use the strong tally set The multitask deep neural network of two kinds of training different frameworks uses more of two kinds of no label data alternative optimization different frameworks Business deep neural network, specifically includes the following steps:
By the tongue in the strong tally set as sample set is randomly divided into two parts, it is denoted as tally set S1 and tally set S2;It is arranged without mark Sample set is signed, U is denoted as;Two empty sets are defined, are denoted as A1 and A2, A1 is for storing the unlabeled exemplars from depth residual error network Set, A2 is for storing the high unlabeled exemplars set of the confidence level from GoogleNet network;Two empty sets are defined, are denoted as B1 and B2, B1 are used for the confidence level storing unlabeled exemplars and its generating by depth residual error network, and B2 is for storing without label Sample and its confidence level generated by GoogleNet network;Two sorter models are defined, depth residual error net is respectively based on The disaggregated model of network is denoted as F1 and the disaggregated model of the deep neural network based on GoogleNet is denoted as F2;Define counter i.
4. a kind of multitask tongue picture automatic analysis method as claimed in claim 3, which is characterized in that tongue picture is input to two The trained multitask deep neural network, averagely as last tongue picture more for the two output results analyzed Business analysis is as a result, specifically include: tongue color of the multitask disaggregated model F1 based on depth residual error network to the tongue picture picture, tongue Shape, tongue mind, coating nature are analyzed and predicted, and F2 pairs of the multitask disaggregated model of the deep neural network based on GoogleNet Tongue color, ligulate, tongue mind, the coating nature of the tongue picture picture are analyzed and predicted.
5. a kind of multitask tongue picture automatic analysis method as claimed in claim 4, which is characterized in that use the no number of tags Multitask deep neural network according to two kinds of the alternative optimization different frameworks includes: the confidence sample selecting no label data and concentrating This optimization multitask disaggregated model F1 based on depth residual error network and the confidence sample for selecting no label data concentration are excellent Change the multitask disaggregated model F2 based on GoogleNet deep neural network.
6. a kind of multitask tongue picture automatic analysis method as claimed in claim 5, which is characterized in that select no label data collection In confidence sample optimization described in the multitask disaggregated model F1 based on depth residual error network, specifically include:
Disaggregated model F1 ' is obtained using tally set S1 training multitask deep neural network F1;
Integrate the sample for selecting size in U as n from unlabeled exemplars at random, is denoted as u={ u1, u2, u3...un};
All samples in sample u are predicted using disaggregated model F1 ', obtain prediction label F '1(u);
Enable counter i=1;
Use S1 ∪ (ui, F '1(ui)) data set multitask deep neural network F1 obtains F "1
Use F '1With F "1Tally set S1 is predicted, sample u is calculated according to confidence calculations formula (1)iCorresponding confidence level Horizontal Δ1i:
Wherein, xjTo there is label data collection S1In sample, yjTo there is label data xjCorresponding true tag.
Record uiConfidence level, be denoted as (ui, Δ1i), it stores in set B1;
I=i+1, judgement, if i≤n, otherwise return step six performs the next step;
The highest sample of level of confidence in set B1 is selected, a is denoted as, takes out sample a and F '1(a), it is put into empty set A2.
7. a kind of multitask tongue picture automatic analysis method as claimed in claim 5, which is characterized in that select no label data collection In multitask disaggregated model F2 of the confidence sample optimization based on GoogleNet deep neural network, specifically include:
F2 ' is obtained using tally set S2 multitask deep neural network F1;
All samples in u are predicted using disaggregated model F2 ', obtain prediction label F '2(u);
Enable counter i=1;
Use S2 ∪ (ui, F '2(ui)) data set multitask deep neural network F2 obtains F "2
Use F '2With F "2Tally set S2 is predicted, calculates sample u according to confidence calculations formula (2)iCorresponding confidence level Horizontal Δ2i:
Wherein, xjTo there is label data collection S2In sample, yjTo there is label data xjCorresponding true tag;
Record uiConfidence level, be denoted as (ui, Δ2i), it stores in set B2;
I=i+1, judgement, if i≤n, otherwise return step 12 performs the next step;
The highest sample of level of confidence in set B2 is selected, b is denoted as, takes out sample b and F '2(b), it is put into A1.
8. a kind of multitask tongue picture automatic analysis method as claimed in claims 6 or 7, which is characterized in that use the no mark The multitask deep neural network of two kinds of data alternative optimization of the label different frameworks is further comprising the steps of:
More new data S1=S1 ∪ A2, S2=S2 ∪ A1;
Judge whether S1 and S2 changes, do not change, exit, otherwise, return selects size from unlabeled exemplars collection U at random For the sample of n, it is denoted as u={ u1, u2, u3...un};
Combine the multitask disaggregated model F1 and the base based on depth residual error network described in strong exemplar collection re -training In the disaggregated model F2 of the deep neural network of GoogleNet.
CN201910397988.2A 2019-05-14 2019-05-14 Automatic analysis method for multitasking tongue picture Active CN110189305B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910397988.2A CN110189305B (en) 2019-05-14 2019-05-14 Automatic analysis method for multitasking tongue picture

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910397988.2A CN110189305B (en) 2019-05-14 2019-05-14 Automatic analysis method for multitasking tongue picture

Publications (2)

Publication Number Publication Date
CN110189305A true CN110189305A (en) 2019-08-30
CN110189305B CN110189305B (en) 2023-09-22

Family

ID=67716176

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910397988.2A Active CN110189305B (en) 2019-05-14 2019-05-14 Automatic analysis method for multitasking tongue picture

Country Status (1)

Country Link
CN (1) CN110189305B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110705425A (en) * 2019-09-25 2020-01-17 广州西思数字科技有限公司 Tongue picture multi-label classification learning method based on graph convolution network
CN111476259A (en) * 2019-11-22 2020-07-31 上海大学 Tooth mark tongue recognition algorithm based on convolutional neural network
CN111476260A (en) * 2019-11-22 2020-07-31 上海大学 Putrefaction classification algorithm based on convolutional neural network
CN113516634A (en) * 2021-06-07 2021-10-19 北京博哥科技合伙企业(有限合伙) Tongue picture teaching training device
CN113724228A (en) * 2021-08-31 2021-11-30 平安科技(深圳)有限公司 Tongue color and coating color identification method and device, computer equipment and storage medium
CN117392138A (en) * 2023-12-13 2024-01-12 四川大学 Tongue picture image processing method, storage medium and electronic equipment

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107292330A (en) * 2017-05-02 2017-10-24 南京航空航天大学 A kind of iterative label Noise Identification algorithm based on supervised learning and semi-supervised learning double-point information
CN107437096A (en) * 2017-07-28 2017-12-05 北京大学 Image classification method based on the efficient depth residual error network model of parameter
CN108764281A (en) * 2018-04-18 2018-11-06 华南理工大学 A kind of image classification method learning across task depth network based on semi-supervised step certainly
CN109034205A (en) * 2018-06-29 2018-12-18 西安交通大学 Image classification method based on the semi-supervised deep learning of direct-push
CN109522961A (en) * 2018-11-23 2019-03-26 中山大学 A kind of semi-supervision image classification method based on dictionary deep learning

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107292330A (en) * 2017-05-02 2017-10-24 南京航空航天大学 A kind of iterative label Noise Identification algorithm based on supervised learning and semi-supervised learning double-point information
CN107437096A (en) * 2017-07-28 2017-12-05 北京大学 Image classification method based on the efficient depth residual error network model of parameter
CN108764281A (en) * 2018-04-18 2018-11-06 华南理工大学 A kind of image classification method learning across task depth network based on semi-supervised step certainly
CN109034205A (en) * 2018-06-29 2018-12-18 西安交通大学 Image classification method based on the semi-supervised deep learning of direct-push
CN109522961A (en) * 2018-11-23 2019-03-26 中山大学 A kind of semi-supervision image classification method based on dictionary deep learning

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
WU HAO等: "Semi-Supervised Deep Learning Using Pseudo Labels for Hyperspectral image Classification", 《IEEE TRANSACTIONS ON IMAGE PROCESSING》 *
李艳秋: "基于集成学习的人脸识别研究", 《中国知网博士电子期刊》 *
汤一平等: "基于多任务卷积神经网络的舌象分类研究", 《计算机科学》 *
王小鹏: "基于弱监督深度学习的图像检索技术研究与实现", 《中国知网硕士电子期刊》 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110705425A (en) * 2019-09-25 2020-01-17 广州西思数字科技有限公司 Tongue picture multi-label classification learning method based on graph convolution network
CN110705425B (en) * 2019-09-25 2022-06-28 广州西思数字科技有限公司 Tongue picture multi-label classification method based on graph convolution network
CN111476259A (en) * 2019-11-22 2020-07-31 上海大学 Tooth mark tongue recognition algorithm based on convolutional neural network
CN111476260A (en) * 2019-11-22 2020-07-31 上海大学 Putrefaction classification algorithm based on convolutional neural network
CN111476260B (en) * 2019-11-22 2023-07-21 上海大学 Greasy fur classification algorithm based on convolutional neural network
CN113516634A (en) * 2021-06-07 2021-10-19 北京博哥科技合伙企业(有限合伙) Tongue picture teaching training device
CN113724228A (en) * 2021-08-31 2021-11-30 平安科技(深圳)有限公司 Tongue color and coating color identification method and device, computer equipment and storage medium
CN113724228B (en) * 2021-08-31 2024-05-10 平安科技(深圳)有限公司 Tongue color and tongue fur color identification method and device, computer equipment and storage medium
CN117392138A (en) * 2023-12-13 2024-01-12 四川大学 Tongue picture image processing method, storage medium and electronic equipment
CN117392138B (en) * 2023-12-13 2024-02-13 四川大学 Tongue picture image processing method, storage medium and electronic equipment

Also Published As

Publication number Publication date
CN110189305B (en) 2023-09-22

Similar Documents

Publication Publication Date Title
CN110189305A (en) A kind of multitask tongue picture automatic analysis method
Beevi et al. Automatic mitosis detection in breast histopathology images using convolutional neural network based deep transfer learning
Jiang et al. White blood cells classification with deep convolutional neural networks
CN106951825A (en) A kind of quality of human face image assessment system and implementation method
Ongun et al. Feature extraction and classification of blood cells for an automated differential blood count system
Colak et al. Automated McIntosh-based classification of sunspot groups using MDI images
Haehn et al. Evaluating ‘graphical perception’with CNNs
CN109815801A (en) Face identification method and device based on deep learning
CN109559300A (en) Image processing method, electronic equipment and computer readable storage medium
CN109543526A (en) True and false facial paralysis identifying system based on depth difference opposite sex feature
CN110139597A (en) The system and method for being iterated classification using neuro-physiological signals
CN109948647A (en) A kind of electrocardiogram classification method and system based on depth residual error network
WO2020224433A1 (en) Target object attribute prediction method based on machine learning and related device
Kerr et al. Collaborative deep learning models to handle class imbalance in flowcam plankton imagery
Li et al. Dating ancient paintings of Mogao Grottoes using deeply learnt visual codes
Ding et al. Multiple lesions detection of fundus images based on convolution neural network algorithm with improved SFLA
Feng et al. Fine-tuning swin transformer and multiple weights optimality-seeking for facial expression recognition
Selvia et al. Skin lesion detection using feature extraction approach
AU2021100367A4 (en) A multi-task automatic analysis method for tongue manifestation
Akbar et al. Cluster-based learning from weakly labeled bags in digital pathology
Townsend et al. Discovering visual concepts and rules in convolutional neural networks
CN113011436A (en) Traditional Chinese medicine tongue color and fur color collaborative classification method based on convolutional neural network
Sims et al. A Neural Architecture for Detecting Confusion in Eye-tracking Data
Harshini et al. Machine Learning Approach for Various Eye Diseases using Modified Voting Classifier Model
Lindsay et al. Deep Learning Networks and Visual Perception

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant