CN110070116B - Segmented selection integration image classification method based on deep tree training strategy - Google Patents

Segmented selection integration image classification method based on deep tree training strategy Download PDF

Info

Publication number
CN110070116B
CN110070116B CN201910274548.8A CN201910274548A CN110070116B CN 110070116 B CN110070116 B CN 110070116B CN 201910274548 A CN201910274548 A CN 201910274548A CN 110070116 B CN110070116 B CN 110070116B
Authority
CN
China
Prior art keywords
classifier
branch
classifiers
layer
deep
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910274548.8A
Other languages
Chinese (zh)
Other versions
CN110070116A (en
Inventor
杨云
胡媛媛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Yunnan University YNU
Original Assignee
Yunnan University YNU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Yunnan University YNU filed Critical Yunnan University YNU
Priority to CN201910274548.8A priority Critical patent/CN110070116B/en
Publication of CN110070116A publication Critical patent/CN110070116A/en
Application granted granted Critical
Publication of CN110070116B publication Critical patent/CN110070116B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention discloses a sectional type selection integration image classification method based on a deep tree training strategy, which adopts the deep tree training strategy, adds a branch classifier at each intermediate layer of a deep convolutional neural network, trains a plurality of branch classifiers which have difference and can extract the characteristics of the intermediate layer by integrating and gradient compensating the branch classifiers, and generates a candidate base classifier set; and screening the candidate base classifiers by using a sectional selection integration method based on correct rate and difference screening to obtain a selected base classifier set, and performing weighted integration on the selected base classifier set to generate a final classification model. According to the invention, image classification is carried out by a gradient compensation technology, a feature fusion technology and a mode of combining deep learning and ensemble learning, so that the training efficiency of a deep neural network is improved, the classification performance of a classifier is improved, the training cost of ensemble learning is reduced, and the generalization and the accuracy of the final classifier are enhanced to the greatest extent.

Description

Segmented selection integration image classification method based on deep tree training strategy
Technical Field
The invention relates to the field of computer vision and supervised learning, in particular to a segmented selection integration image classification method based on a deep tree training strategy.
Background
Image classification is a main research direction in the field of computer vision, and plays a very important role and influence in intelligent data processing taking images as main bodies. Because image classification has higher theoretical research and practical application value, many scholars at home and abroad put forward a plurality of image classification models based on machine learning. The traditional machine learning method is divided into three steps: feature extraction, feature selection and classification, and deep learning is one type of machine learning method. At present, the deep convolutional neural network which is widely applied is a deep learning model with the best effect.
In 2012, a.krizhevsky et al proposed an eight-layer convolutional neural network AlexNet in Imagenet classification with deep connected neural networks, and took the champion in the Imagenet image classification competition of the current year and far beyond the second ten percent, and let the convolutional neural network come back to the sight of people again. Various improved convolutional neural network structures were subsequently developed as bamboo shoots in the spring after rain, representative of which are VGG, google lenet and ResNet.
However, in deep learning, there are two main problems: (1) the problem that the gradient disappears can occur in the deep neural network along with the increase of the number of network layers, so that the model is difficult to train and converges to a good local optimal point; (2) an image classification model constructed by using the deep neural network generally only uses the deepest features of the network and a group of features with the highest image extraction degree as a final classification basis, but a single group of features is weak in robustness in the face of complex tasks, so that the single group of features is not the best features for classification for some image classification tasks, in other words, the traditional deep neural network lacks stability and robustness for the image classification tasks. For example, c.szegdy et al, in Going stripper with considerations, 2014, proposed a convolutional neural network google net with 22 layers, which employs two branch classifiers with 1 convolutional layer, 1 pooling layer, and 2 fully-connected layers, and the gradient compensation effect is lost because the gradient of compensation needs to go through more layers. In addition, the basis for final classification is only the deepest features, and the stability and robustness in the face of complex image classification problems are not ideal.
Therefore, a method combining the deep learning and the integrated learning which has flexibility and can improve stability and robustness is also developed, however, the following two main problems still exist in such a method: (1) the training cost of a single deep neural network is high, so the time cost for training an integrated member for integrated learning by adopting a traditional deep training method is very high; (2) the core idea of ensemble learning is that a plurality of ensemble members with high classification accuracy and difference are trained and integrated, so that the generalization of an ensemble learning model is improved; however, since the classification accuracy and the classification variability are not independent from each other, it is still the key point and the difficulty in the field of ensemble learning to maximize the generalization of the ensemble model by considering both the classification accuracy and the classification variability.
Disclosure of Invention
The invention aims to provide a sectional type selection integration image classification method based on a deep tree-shaped training strategy aiming at the defects and shortcomings of the prior art, and solves the problem of gradient disappearance, improves the efficiency of deep network training, and enhances the generalization and the accuracy of a final classifier to the maximum extent through technical methods such as gradient compensation, feature fusion, sectional type selection integration based on accuracy and difference screening and the like.
In order to achieve the purpose, the invention adopts the technical scheme that:
the segmented selection integration image classification method based on the deep tree training strategy comprises the following steps:
s1: a deep tree training strategy is adopted, a branch classifier is added to each middle layer of a deep convolutional neural network, each branch classifier only comprises a full connection layer, and the branch classifiers are not discarded in the training or testing process; performing gradient compensation on the weight at the shallow layer by using the branch classifier; training a deep convolutional neural network containing branch classifiers to obtain a plurality of branch classifiers which have difference and can classify on the characteristics of the middle layer, namely obtaining a candidate base classifier set with a plurality of branch classifiers;
s2: and obtaining final integration members and generating a final classification model by adopting a sectional selection integration method based on accuracy and difference screening.
Further, the construction process of the branch classifier in step S1 is as follows:
s1.11: the specific implementation method for acquiring the characteristic diagram from the intermediate convolution layer comprises the following steps:
the input picture training data set with the class mark is s { (X) i ,Y i ) I ═ 1,2, …, N }, where the picture data X i ∈R n (indicating that each picture data is n-dimensional) and picture X i Corresponding true class label Y i E {1, …, K } (where K is the total number of categories);
without considering the bias term, the convolution and activation calculation process for convolutional layer M (where M ∈ {1, …, M }, and M is the total number of convolutional layers) can be expressed as shown in formula (a1), and the pooling calculation process is shown in formula (a 2):
Q m =f(w m *Z m-1 ),and(Z 0 ≡X i ) (A1)
Z m =g(Q m ) (A2)
wherein, w m Is the weight of convolutional layer m to be learned, is the convolution operation, Z m-1 The characteristic diagram generated for layer m-1 is shown. The function f (mu) is max (0, mu) is a ReLu activation function which can accelerate network training and obtain a model with better performance, and g (Q) m ) Representing a maximal pooling function;
s1.12: flattening the characteristic diagram, wherein the specific implementation method comprises the following steps:
performing feature extraction on input data layer by the convolutional layer, wherein a feature map extracted by the last convolutional layer is a two-dimensional feature matrix, the flattening operation is to convert the two-dimensional feature matrix into a one-dimensional feature vector, then the one-dimensional feature vector is input into a classifier to perform final decision calculation, and neurons between the one-dimensional feature vector obtained by the flattening operation and the classifier are fully connected;
s1.13: the feature vector after flattening is combined with the classifier in a full-connection mode, and the specific implementation process is as follows:
s1.13.1: in the full connection layer part, V is assumed (m) And
Figure RE-GDA0002066529030000031
respectively represent the feature vectors and weight matrices input to the branch classifier m,
Figure RE-GDA0002066529030000032
representing its bias term, the final layer of output vectors O of the branch classifier m (m) Is calculated as shown in equation (A3):
Figure RE-GDA0002066529030000033
inputting the output vector into a classifier to calculate a K-dimensional probability vector, wherein the class K corresponding to the maximum probability value is the branch classifier m according to the input example X i Calculating a prediction class mark, wherein the corresponding probability value is called a score;
s1.13.2: to conveniently represent the objective function of the branch classifier m in deep tree training, convolutional layer weights (w) are used 1 ,…,w m ) Weight matrix for fully connected network with branch classifier m
Figure RE-GDA0002066529030000034
Combining as shown in formula (A4):
Figure RE-GDA0002066529030000035
s1.13.3: evaluating the learning effect of the deep learning model by measuring the distance between the prediction score and the real class mark through the cross entropy by using an objective function based on cross entropy loss, wherein the smaller the cross entropy in the training data set is, the better the fitting effect of the model on the training data set is represented; the cross-entropy loss of the branch classifier m in the training data set containing n instances is shown as equation (a 5):
Figure RE-GDA0002066529030000036
wherein, log (x) ═ ln (x);
thus, the objective function of the branch classifier m
Figure RE-GDA0002066529030000037
As shown in formula (A6):
Figure RE-GDA0002066529030000038
s1.13.4: for each branch classifier, the data characteristics of the final classification decision are different, and the contribution degree of the branch classifier to gradient compensation in the training process can also be different; therefore, the final objective function of the convolutional neural network using the deep tree training strategy performs weighted integration on the objective functions of the branch classifiers, as shown in formula (a 7):
Figure RE-GDA0002066529030000041
wherein, beta (m) Denotes the weight of the branch classifier m, W ═ W (1) ,…,W (M) ),
Figure RE-GDA0002066529030000042
Thus, the objective function of the deep convolutional neural network trained by using the deep tree training strategy is obtained, and if a set of weights W and theta are found according to the relation of the objective function full To make J (W, theta) full ) The smallest means that the classification performance of the model is the best.
Further, the process of the gradient compensation in step S1 is as follows:
s1.21: calculating the error between the prediction class mark and the real class mark of each branch classifier;
s1.22: the gradients of the branch classifiers are injected into the sharing part layer by layer in a back propagation mode, and the sharing part sums the gradients obtained from each branch classifier and is used for updating the weights. The gradient compensation is mainly expressed in a back propagation stage, and for a convolutional neural network using a deep tree training strategy, the gradient compensation is realized as follows:
weight w at layer m m Gradient that can be obtained
Figure RE-GDA0002066529030000043
From M-M branch classifiers after layer M, thus
Figure RE-GDA0002066529030000044
Is calculated as shown in equation (A8):
Figure RE-GDA0002066529030000045
wherein,
Figure RE-GDA0002066529030000046
and beta (m+i) Representing the gradient obtained from the branch classifier m + i and the assigned weight, respectively. From equation (A8), a weighted average of the gradients from the different branch classifiers is calculated at each gradient compensation node, which is also the gradient used during the parameter update of the model. Therefore, the model parameter updating is carried out by using the average gradient obtained by the gradient injected by each branch classifier, so that the stability of parameter updating can be improved, and the final objective function can quickly reach a minimum value.
Further, the specific implementation process of step S2 is as follows:
s2.1: screening the candidate base classifiers based on the classification accuracy, firstly calculating the verification accuracy of each candidate base classifier, then performing descending order arrangement on the candidate base classifiers according to the verification accuracy, and screening out the front R base classifiers with better performance in the sequence;
s2.2: adopting a double-error difference metric function, carrying out centralized screening on the first R base classifiers screened and determined in the step S2.1, and screening out a base classifier set which has the maximum internal difference and can ensure that the generalization capability of the integrated classifier is optimal as a selected base classifier set, wherein the specific implementation method comprises the following steps:
classifier d in base classifier screening based on double-error difference metric function i And d j The difference between Div (d) i ,d j ) The joint distribution between the two classifiers can be expressed as shown in formula (a 9):
Figure RE-GDA0002066529030000051
wherein N is 11 、N 10 、N 01 、N 00 Represents two classifiers d i And d j Number of instances of joint distribution between, e.g. N 11 Representation classifier d i And d j All classify the correct number of instances;
and (4) performing iterative screening on the first R base classifiers selected in the step (S2.1) one by one, and adding the classifier with the maximum difference into the selected base classifier set one by one. Wherein a certain classifier d i With a set of classifiers D * The difference between them is an average difference calculated as shown in formula (a 10):
Figure RE-GDA0002066529030000052
wherein, | D * I denotes the set of classifiers D * The number of classifiers contained;
s2.3: determining the weight of the selected base classifier set obtained by screening in the step S2.2 according to the classification error magnitude in the verification data set, considering the difference between each base classifier and the contribution degree of the base classifier set to the target task, and then determining the selected base classifier d t (where T is 1 … T) and the final classifier model is expressed by the formula (T:)A11) Shown in the figure:
Figure RE-GDA0002066529030000053
wherein,
Figure RE-GDA0002066529030000054
represented in the selected base classifier set D * The classification result (i.e. the probability vector) of the t-th classifier for input instance x,
Figure RE-GDA0002066529030000055
its corresponding weight. Therefore, the objective function of the weighted integration is as shown in equation (a 12):
Figure RE-GDA0002066529030000061
wherein, CQ ij =∫(d i (x)-f(x))(d j (x) -f (x)) p (x) dx, p (x) is the distribution of sample x.
For weighted integration, the goal is to find a set of weights such that loss (EN) takes a minimum. According to the Lagrange multiplier method, here the weight α t The calculation of (c) can be expressed as shown in equation (a 13):
Figure RE-GDA0002066529030000062
wherein,
Figure RE-GDA0002066529030000063
indicating the quality of the corresponding t-th classifier,
Figure RE-GDA0002066529030000064
is a normalization term.
Further, in the construction process of the branch classifier, the classifier used is a softmax classifier, an SVM classifier or other classifiers capable of performing multi-classification tasks.
Compared with the prior art, the invention has the following advantages and beneficial effects:
1. by a gradient compensation method, the gradient is supplemented for the deep neural network in the training process, the problem of gradient disappearance is solved, and the training efficiency of the deep neural network is improved to the greatest extent;
2. the method comprises the steps that a plurality of branches are led out from a plurality of middle hidden layers of a deep neural network, branch classifiers which adopt middle layer characteristics as classification bases are constructed, and then the branch classifiers are integrated by adopting ensemble learning, so that the middle layer characteristics and the deepest layer characteristics are fused, the middle layer characteristics are utilized to the greatest extent, and the classification performance of the final classifier is improved;
3. because the branch classifier with only one full connection layer is adopted, a plurality of classifiers with high classification accuracy and certain difference can be obtained by training the deep neural network containing a plurality of branch classifiers once, and the branch classifiers are taken as integration members, so that the integrated learning training cost of taking the deep neural network as the integration members is reduced;
4. by the method, the integrated members are screened according to the classification accuracy of the integrated members, the screened integrated members are screened again according to the classification difference of the integrated members, and the weights are determined according to the classification errors of the screened base classifiers in the verification data set for weighted integration, so that the generalization and the accuracy of the final classifier are enhanced to the greatest extent.
Drawings
FIG. 1 is a flow chart of an algorithm for segmented selection integration based on a deep tree training strategy;
FIG. 2 is a diagram of an example of CIFAR-10 image data.
Detailed Description
The present invention will be described in further detail with reference to specific embodiments, but the embodiments of the present invention are not limited thereto.
As shown in fig. 1, the present embodiment provides a segmentation type selection integrated image classification method based on a deep tree training strategy, which includes the following steps:
s1: a deep tree training strategy is adopted, a branch classifier is added to each middle layer of a deep convolutional neural network, each branch classifier only comprises a full connection layer, and the branch classifiers are not discarded in the training or testing process; performing gradient compensation on the weight at the shallow layer by using the branch classifier; training a deep convolutional neural network containing branch classifiers to obtain a plurality of branch classifiers which have difference and can classify on the characteristics of the middle layer, namely obtaining a candidate base classifier set with a plurality of branch classifiers;
s2: and obtaining final integration members and generating a final classification model by adopting a sectional selection integration method based on accuracy and difference screening.
Further, the construction process of the branch classifier in step S1 is as follows:
s1.11: the specific implementation method for acquiring the characteristic diagram from the intermediate convolution layer comprises the following steps:
the input picture training data set with the class mark is S { (X) i ,Y i ) I ═ 1,2, …, N }, where the picture data X i ∈R n (indicating that each picture data is n-dimensional) and picture X i Corresponding true class label Y i E {1, …, K } (where K is the total number of categories).
Without considering the bias term, the convolution and activation calculation process of convolutional layer M (where M ∈ {1, …, M }, and M is the total number of convolutional layers) can be expressed as shown in formula (B1), and the pooling calculation process is shown in formula (B2):
Q m =f(w m *Z m-1 ),and(Z 0 ≡X i ) (B1)
Z m =g(Q m ) (B2)
wherein, w m Is the weight of convolutional layer m to be learned, is the convolution operation, Z m-1 The characteristic diagram generated for layer m-1 is shown. The function f (mu) is max (0, mu) which is a ReLu activation function and can accelerate network training and obtain better performanceModel, g (Q) m ) The maximum pooling function is represented.
S1.12: flattening the characteristic diagram, wherein the specific implementation method comprises the following steps:
performing feature extraction on input data layer by the convolutional layer, wherein a feature map extracted by the last convolutional layer is a two-dimensional feature matrix, the flattening operation is to convert the two-dimensional feature matrix into a one-dimensional feature vector, then the one-dimensional feature vector is input into a classifier to perform final decision calculation, and neurons between the one-dimensional feature vector obtained by the flattening operation and the classifier are fully connected;
s1.13: the feature vector after flattening is combined with the classifier in a full-connection mode, and the specific implementation process is as follows:
s1.13.1: in the full connection layer part, V is assumed (m) And
Figure RE-GDA0002066529030000081
respectively represent the feature vectors and weight matrices input to the branch classifier m,
Figure RE-GDA0002066529030000082
representing its bias term, the final layer output vector o (m) of the branch classifier m is calculated as shown in equation (B3):
Figure RE-GDA0002066529030000083
inputting the output vector into a softmax classifier to calculate a K-dimensional probability vector, wherein the class K corresponding to the maximum probability value is the branch classifier m according to the input example X i The corresponding probability value is called score.
Here, the softmax classifier is represented by the formula (B4):
Figure RE-GDA0002066529030000084
wherein exp (x) e x Σ is the sum symbol, h (m) (y=k|X i ) Is a sample X i Is classified as the probability score corresponding to class k.
S1.13.2: to conveniently represent the objective function of the branch classifier m in deep tree training, convolutional layer weights (w) are used 1 ,…,w m ) Weight matrix for fully connected network with branch classifier m
Figure RE-GDA0002066529030000085
Combining as shown in formula (B5):
Figure RE-GDA0002066529030000086
s1.13.3: evaluating the learning effect of the deep learning model by measuring the distance between the prediction score and the real class mark through the cross entropy by using an objective function based on cross entropy loss, wherein the smaller the cross entropy in the training data set is, the better the fitting effect of the model on the training data set is represented; the cross-entropy loss of the branch classifier m in the training dataset containing n instances is shown as (B6):
Figure RE-GDA0002066529030000087
wherein, log (x) ═ ln (x);
thus, the objective function of the branch classifier m
Figure RE-GDA0002066529030000091
As shown in formula (B7):
Figure RE-GDA0002066529030000092
s1.13.4: for each branch classifier, the data characteristics of the final classification decision are different, and the contribution degree of the branch classifier to gradient compensation in the training process may also be different; therefore, the final objective function of the convolutional neural network using the deep tree training strategy performs weighted integration on the objective functions of the branch classifiers, as shown in formula (B8):
Figure RE-GDA0002066529030000093
wherein beta is (m) Denotes the weight of the branch classifier m, W ═ W (1) ,…,W (M) ),
Figure RE-GDA0002066529030000094
Thus, the final objective function of the deep convolutional neural network trained by using the deep tree training strategy is obtained, and if a set of weights W and theta are found according to the objective function relation full To make J (W, theta) full ) The smallest means that the classification performance of the model is the best.
Further, the process of the gradient compensation in step S1 is as follows:
s1.21: the error between the predicted and true class labels of each branch classifier is calculated.
S1.22: the gradients of the branch classifiers are injected into the sharing part layer by layer in a back propagation mode, and the sharing part sums the gradients obtained from each branch classifier and is used for updating the weights.
The gradient compensation is mainly expressed in a back propagation stage (i.e. a weight updating stage), and the effectiveness of the deep tree training strategy is illustrated by comparing the gradient of a general convolutional neural network with the gradient of a convolutional neural network using the deep tree training strategy. For a convolutional neural network that does not use a deep tree training strategy, its weight w at layer m m Gradient that can be obtained
Figure RE-GDA0002066529030000095
As shown in formula (B9), wherein
Figure RE-GDA0002066529030000096
Represents the solving function J (W, theta) full ) For w m Derivative of (a):
Figure RE-GDA0002066529030000097
and in the iterative update of the ith round, the weight w m Is shown as equation (B10), where λ is the learning rate, i.e. the step size of each weight update:
Figure RE-GDA0002066529030000098
it is obvious that
Figure RE-GDA0002066529030000101
When the factor in (1) is less than 1, the parameter w m It is difficult to get updated.
For a convolutional neural network using a deep tree training strategy, its weight w at layer m m Gradient that can be obtained
Figure RE-GDA0002066529030000102
From M-M branch classifiers after layer M, thus
Figure RE-GDA0002066529030000103
Is calculated as shown in equation (B11):
Figure RE-GDA0002066529030000104
wherein,
Figure RE-GDA0002066529030000105
and beta (m+i) Representing the gradient obtained from the branch classifier m + i and the assigned weight, respectively. Obviously, for the same convolutional neural network, there is
Figure RE-GDA0002066529030000106
That is to say using a depth treeThe training strategy can inject more gradients into the model, and the problem of gradient disappearance is solved.
From equation (B11), a weighted average of the gradients from the different branch classifiers is calculated at each gradient compensation node, which is also the gradient used during the parameter update of the model. For the classification task of a specific data set, the model that can fit the data set is not unique, so the training process of the deep neural network is also the process of searching in the solution space. However, due to factors such as data noise, there is a chance that the gradient calculated by the deep neural network in the training process (sometimes, the obtained gradient can be beneficial to model convergence, and sometimes, the obtained gradient is harmful to model convergence). In the invention, each branch classifier has the same objective function, namely, a data set is drawn to minimize the classification error. Therefore, the model parameter updating by using the average gradient obtained by the gradients injected by the branch classifiers can improve the stability of parameter updating, so that the final objective function can quickly reach a minimum value.
Further, the specific implementation process of step S2 is as follows:
s2.1: and screening the candidate base classifiers based on the classification accuracy, firstly calculating the verification accuracy of each candidate base classifier, then performing descending arrangement on the candidate base classifiers according to the verification accuracy, and screening the front R base classifiers with better performance in the sequence.
S2.2: adopting a double-error difference measurement function, then carrying out centralized screening on the first R base classifiers screened and determined in the step S2.1, and screening out a base classifier set which has the maximum internal difference and can ensure that the generalization capability of the integrated classifier is optimal as a selected base classifier set, wherein the specific implementation method comprises the following steps:
classifier d in base classifier screening based on double-error difference metric function i And d j The difference between Div (d) i ,d j ) The joint distribution between two classifiers according to table 1 can be expressed as shown in equation (B12):
Figure RE-GDA0002066529030000111
wherein N is 11 、N 10 、N 01 、N 00 Represents two classifiers d i And d j Number of instances of joint distribution between, e.g. N 11 Representation classifier d i And d j The correct number of instances is classified.
TABLE 1 Joint distribution between a pair of classifiers
Figure RE-GDA0002066529030000112
The first R base classifiers selected in the step S2.1 are iteratively screened one by one, the classifier with the maximum difference is added into a selected base classifier set one by one, and a certain classifier d is subjected to iterative screening i With a set of classifiers D * The difference between them is an average difference calculated as shown in formula (B13):
Figure RE-GDA0002066529030000113
wherein, | D * I denotes the set of classifiers D * The number of classifiers included.
Specifically, the algorithm flow based on differential screening is as follows:
inputting:
Figure RE-GDA0002066529030000114
representing a validation data set;
Figure RE-GDA0002066529030000115
represents the set of base classifiers after the first stage screening (M classifiers, d) i The ith classifier to be selected);
t indicates that T classifiers need to be selected.
And (3) outputting:
Figure RE-GDA0002066529030000116
a set of classifiers is selected (of which there are a total of T classifiers,
Figure RE-GDA0002066529030000117
for the selected ith classifier);
at S val On each candidate classifier d i The classification error of (2);
d: -at S val Classifier d for obtaining minimum classification error i
D * :={d};
D:=D/D * (differencing, i.e. removing set D from set D * All elements of (a).
The method comprises the following steps:
for each d j E, executing the following steps:
calculation of Div (d) according to equation (B13) j ,D * ),
When all classifiers D in the set D j The layer cycle is ended after the differences have been calculated;
②R t i.e. to the classifier D in the set D j According to the difference Div (d) j ,D * ) Performing descending arrangement;
(iii) d ═ in R t Classifier d being first in the middle column j
④D * :=D * U { D } indicates that the classifier D selected in the present round of the loop is added to the set D *
D ═ D/{ D } denotes that the selected classifier D in this round of rotation is removed from set D;
sixthly, executing T-1 times to finish the circulation.
S2.3: determining the weight of the selected base classifier set obtained by screening in the step S2.2 according to the classification error of the selected base classifier set in the verification data set, and considering the difference between each base classifier and the contribution degree of the base classifier set to the target task and then applying the selected base classifier set to the selected base classifier
Figure RE-GDA0002066529030000121
(where T ═ 1 … T) and the final classifier model is represented by the formula (B14):
Figure RE-GDA0002066529030000122
wherein, d t (x) Represented in the selected base classifier set D * The classification result (i.e. probability vector) of the t-th classifier for input instance x,
Figure RE-GDA0002066529030000123
its corresponding weight. Therefore, the objective function of the weighted integration is shown as (B15):
Figure RE-GDA0002066529030000124
wherein, CQ ij =∫(d i (x)-f(x))(d j (x) -f (x)) p (x) dx, p (x) is the distribution of sample x.
For weighted integration, the goal is to find a set of weights such that loss (EN) takes a minimum. According to the Lagrange multiplier method, here the weight α t The calculation of (c) can be expressed as shown in equation (B16):
Figure RE-GDA0002066529030000131
wherein,
Figure RE-GDA0002066529030000132
indicating the quality of the corresponding t-th classifier,
Figure RE-GDA0002066529030000133
is a normalization term.
The following experiments were conducted in accordance with the method of the present invention to illustrate the experimental effects of the present invention.
On the basis of the disclosed reference image classification data set, the image classification method based on the segmented selection integration technology of the deep tree training strategy disclosed by the invention is well verified.
The experimental conditions are as follows: the experiment used a CIFAR-10 image dataset containing 50000 images in the training dataset and 10000 images in the test dataset, for 10 categories, and an example of image data is shown in FIG. 2.
Preparation of the experiment: before the start of the experiment, we randomly sampled 10% from the training data set as the validation data set. The training data set is used for model training, the verification data set is used for artificially adjusted hyper-parameter selection, and the test data set is used for final model evaluation.
The method comprises the following specific implementation steps:
the experiment firstly uses a training data set and a verification data set to train according to the method of the invention to obtain a final integrated classification model, and then uses a test data set to test the classification performance of the final integrated classification model, wherein the evaluation performance mainly comprises classification accuracy, sensitivity, specificity, F1 score and CPU calculation time.
The method provided by the invention is a framework technology, namely, the method can be flexibly integrated with various basic convolutional neural networks. In this verification experiment, two specific models, respectively, a model based on an AlexNet convolutional neural Network (referred to as AlexNet-based) and a model based on a Network in Network convolutional neural Network (referred to as NIN-based), are adopted, and the specific model structures are shown in tables 2 and 3. Taking an AlexNet-based model as an example, 8 branch classifiers are led out from a hidden layer of the AlexNet-based model, then the whole model is trained, then a sectional type selection integration technology is used for screening the 8 branch classifiers, 4 branch classifiers are left after screening, and finally the 4 branch classifiers are subjected to weighted integration to obtain a final classifier model.
TABLE 2 AlexNet-based model realization by deep tree training method based on AlexNet
Figure RE-GDA0002066529030000141
TABLE 3 implementation of NIN-based model using deep tree training method on the basis of Network in Network
Figure RE-GDA0002066529030000142
The experimental results are as follows:
as can be seen from the test results in Table 4, the classification performance of the method of the invention on the CIFAR-10 data set is improved. The AlexNet-based model which takes AlexNet as a basic model and adopts the sectional selection integration based on the deep tree-shaped training strategy is improved by 2.8 percent in classification accuracy compared with the original AlexNet model, and the NIN-based model which takes Network in Network as a basic model and adopts the sectional selection integration based on the deep tree-shaped training strategy is improved by 1.64 percent in accuracy compared with the original Network in Network. In addition, the improved model is greatly improved compared with the original model on the evaluation indexes of Sensitivity (SEN), Specificity (SPE) and F1 score. The CPU Time (i.e. Time cost of per epoch) is affected by the number of layers of the model, so the improved model needs slightly more Time (mainly the calculation on the branch is Time-consuming), but the improved model performs better in the balance of performance and Time cost, and is also improved in each evaluation index compared with other classical image classification models (LENET, VGG). The method has great reference value for practical application scenes including image classification tasks, such as medical image diagnosis, biological recognition, automatic driving and the like.
TABLE 4 comparison of Classification Performance of the methods of the present invention on test sets
Figure RE-GDA0002066529030000151
The above description is only exemplary of the invention, and any modification, equivalent replacement, and improvement made within the spirit and principle of the invention should fall within the protection scope of the invention.

Claims (6)

1. The segmented selection integrated image classification method based on the deep tree training strategy is characterized by comprising the following steps of:
s1, adopting a deep tree training strategy, adding a branch classifier at each middle layer of the deep convolutional neural network, wherein each branch classifier only comprises a full connection layer, and the branch classifier is not discarded in the training or testing process; the process of gradient compensation through the branch classifiers is divided into two parts, firstly, the error between the prediction class label and the real class label of each branch classifier is calculated, then the gradient of the branch classifier is injected into a sharing part layer by layer in a back propagation mode, and the sharing part sums the gradients obtained from each branch classifier and is used for updating the weight; training a deep convolutional neural network containing branch classifiers to obtain a plurality of branch classifiers which have difference and can classify on the intermediate layer characteristics, and thus obtaining a candidate base classifier set with a plurality of branch classifiers;
s2, adopting a sectional selection integration method based on correct rate and difference screening to finally integrate members and generate a final classification model, firstly screening a base classifier set with better performance for a candidate base classifier based on classification correct rate, and then screening a base classifier set which has the maximum internal difference and can ensure the generalization capability of an integrated classifier to be optimal from the base classifier set through a double-error difference measurement function as a selected base classifier set; and finally, optimally integrating the selected base classifier into a final classifier through a weighted integration strategy.
2. The method according to claim 1, wherein the construction process of the branch classifier in step S1 is as follows:
s1.11, acquiring a characteristic diagram from the middle convolutional layer, wherein the specific implementation method comprises the following steps:
the input picture training data set with the class mark is S { (X) i ,Y i ) 1,2, a, N }, wherein picture data X i ∈R n That each picture data is n-dimensional, and picture X i Corresponding true class label Y i E { 1.., K }, where K is the total number of categories;
without considering the bias term, the convolution and activation calculation process of convolutional layer M can be expressed as shown in formula (1), where M ∈ { 1...., M }, M is the total number of convolutional layers, and the pooling calculation process is shown in formula (2):
Q m =f(w m *Z m-1 ),and(Z 0 ≡X i ) (1)
Z m =g(Q m ) (2)
wherein, w m Is the weight of convolutional layer m to be learned, is the convolution operation, Z m-1 Representing a feature map generated by a layer m-1, wherein a function f (u) ═ max (0, u) is a ReLu activation function which can accelerate network training and obtain a model with better performance, and g (·) represents a maximum pooling function;
s1.12, flattening the characteristic diagram, wherein the specific implementation method comprises the following steps:
feature extraction is carried out on input data layer by the convolutional layers, a feature map extracted by the last convolutional layer is a two-dimensional feature matrix, the flattening operation is to convert the two-dimensional feature matrix into a one-dimensional feature vector, then the one-dimensional feature vector is input into a classifier to carry out final decision calculation, and neurons between the one-dimensional feature vector obtained from the flattening operation and the classifier are in full connection;
and S1.13, combining the flattened feature vectors with a classifier in a full-connection mode.
3. The method for classifying integrated images based on segmented selection of a deep tree training strategy according to claim 2, wherein the step S1.13 is implemented by:
s1.13.1 in the full junction layer part, V is assumed (m) And
Figure FDA0003702520980000021
respectively representing the feature vectors and weight matrices input to the branch classifier m,
Figure FDA0003702520980000022
representing its bias term, the final layer of output vectors O of the branch classifier m (m) Is calculated as shown in equation (3):
Figure FDA0003702520980000023
inputting the output vector into a classifier to calculate a K-dimensional probability vector, wherein the class K corresponding to the maximum probability value is the branch classifier m according to the input example X i Calculating a prediction class mark, wherein the corresponding probability value is called a score;
s1.13.2 for convenience in representing the objective function of the branch classifier m in deep tree training, convolutional layer weights (w) 1 ,...,w m ) Weight matrix for fully connected network with branch classifier m
Figure FDA0003702520980000024
Combining, as shown in formula (4):
Figure FDA0003702520980000025
s1.13.3, evaluating the learning effect of the deep learning model by measuring the distance between the prediction score and the real class mark by using an objective function based on cross entropy loss, wherein the smaller the cross entropy in the training data set is, the better the fitting effect of the model to the training data set is; the cross-entropy loss of the branch classifier m in the training dataset containing n instances is shown as equation (5):
Figure FDA0003702520980000031
wherein log (x) ═ ln (x), h (m) (y=k|X i ) Is a sample X i The probability scores corresponding to class k are classified, M belongs to { 1., M }, and M is the total number of convolutional layers;
thus, the objective function of the branch classifier m
Figure FDA0003702520980000032
As shown in formula (6):
Figure FDA0003702520980000033
s1.13.4, for each branch classifier, the data characteristics of the final classification decision are different, and the contribution degree of the final classification decision to the gradient compensation in the training process can also be different; therefore, the final objective function of the convolutional neural network of the deep tree training strategy is used to perform weighted integration on the objective functions of the branch classifiers, as shown in formula (7):
Figure FDA0003702520980000034
wherein, beta (m) Denotes the weight of the branch classifier m, W ═ W (1) ,...,W (M) ),
Figure FDA0003702520980000035
M belongs to { 1.,. M }, wherein M is the total number of convolutional layers;
thus, the final objective function of the deep convolutional neural network trained by using the deep tree training strategy is obtained, and if a set of weights W and theta are found according to the objective function relation full To make J (W, theta) full ) The smallest means that the classification performance of the model is the best.
4. The method for classifying selective integration images based on the deep tree training strategy according to claim 1, wherein the gradient compensation process in step S1 is as follows:
s1.21: calculating the error between the prediction class mark and the real class mark of each branch classifier;
s1.22, injecting the gradients of the branch classifiers into a sharing part layer by layer in a back propagation mode, and summing the gradients obtained from each branch classifier by the sharing part and using the gradients for updating the weight; the gradient compensation is mainly expressed in a back propagation stage, and for a convolutional neural network using a deep tree training strategy, the gradient compensation is realized as follows:
weight w at layer m m Gradient that can be obtained
Figure FDA0003702520980000041
From M-M branch classifiers after layer M, thus
Figure FDA0003702520980000042
Is calculated as shown in equation (8):
Figure FDA0003702520980000043
wherein,
Figure FDA0003702520980000044
and beta (m+i) Respectively representing the gradients obtained from the branch classifier M + i and the assigned weights, M ∈ { 1., M }, M being the total number of convolutional layers, J (W, θ) full ) A final objective function for a convolutional neural network using a deep-tree training strategy, where W ═ W (1) ,...,W (M) ),
Figure FDA0003702520980000045
Expressed as convolutional layer weights (w) 1 ,...,w m ) Weight matrix for fully connected network with branch classifier m
Figure FDA0003702520980000046
In the combination of (a) and (b),
Figure FDA0003702520980000047
representing the objective function of the branch classifier m + i, W (m+i) Expressed as convolutional layer weights (w) 1 ,...,w m+i ) Weight matrix for fully connected network with branch classifier m + i
Figure FDA0003702520980000048
Combination of (a) and (b), h (m+i) Representing the probability score, O, of the branch classifier m + i (m+i) The last layer output vector, V, represented as the branch classifier m + i (m+i) Representing the feature vector, Z, input to the branch classifier m + i m+i Feature map, Q, representing the generation of layer m + i m+i Representing a characteristic diagram, Q, of the convolutional layer m + i after convolution and activation m Representing a characteristic diagram obtained after convolution and activation of the convolution layer m; as can be seen from equation (8), a weighted average of the gradients from the different branch classifiers is calculated at each gradient compensation node, which is also the gradient used during the parameter update of the model; therefore, the model parameter updating is carried out by using the average gradient obtained by the gradient injected by each branch classifier, so that the stability of parameter updating can be improved, and the final objective function can quickly reach a minimum value.
5. The method for classifying selective integration images based on the deep tree training strategy according to claim 1, wherein the step S2 is implemented by:
s2.1, screening the candidate base classifiers based on the classification accuracy, firstly calculating the verification accuracy of each candidate base classifier, then performing descending order arrangement on the candidate base classifiers according to the verification accuracy, and screening the front R base classifiers with better performance in the sequence;
s2.2, adopting a double-error difference measurement function, then carrying out centralized screening on the first R base classifiers screened and determined in the step S2.1, and screening out a base classifier set which has the maximum internal difference and can enable the generalization capability of the integrated classifier to be optimal as a selected base classifier set, wherein the specific implementation method comprises the following steps:
classifier d in base classifier screening based on double-error difference metric function i And d j The difference between the two classifiers can be expressed as the following formula (9) according to the joint distribution between the two classifiers:
Figure FDA0003702520980000051
wherein, N 11 、N 10 、N 01 、N 00 Represents two classifiers d i And d j Number of instances of joint distribution therebetween, e.g. N 11 Representation classifier d i And d j The correct number of instances are classified;
the first R base classifiers selected in the step S2.1 are iteratively screened one by one, and the classifier with the maximum difference is added to a selected base classifier set one by one, wherein a certain classifier d i With a set of classifiers D * The difference between them is the average difference, which is calculated as shown in equation (10):
Figure FDA0003702520980000052
wherein, | D * I denotes the set of classifiers D * The number of classifiers contained;
s2.3, determining the weight of the selected base classifier set obtained by screening in the step S2.2 according to the classification error of the selected base classifier set in the verification data set, considering the difference between each base classifier and the contribution degree of the base classifier set to the target task, and then applying the weight to the selected base classifier set
Figure FDA0003702520980000053
Wherein T is 1.. T, performing weighted integration, and representing a final classifier model as shown in formula (11):
Figure FDA0003702520980000054
wherein, d t (x) Represented in the selected base classifier set D * The classification result of the t-th classifier for the input instance x, i.e. the probability vector, α t Is its corresponding weight, where α t ≥0,
Figure FDA0003702520980000061
Therefore, the objective function of the weighted integration is as shown in equation (12):
Figure FDA0003702520980000062
wherein, CQ ij =∫(d i (x)-f(x))(d j (x) -f (x)) p (x) dx, p (x) is the distribution of samples x, the function f (x) max (0, x) is the ReLu activation function;
for weight integration, the goal is to find a set of weights such that loss (EN) takes a minimum, where the weight α is given by the Lagrangian multiplier method t The calculation of (c) can be expressed as shown in equation (13):
Figure FDA0003702520980000063
wherein
Figure FDA0003702520980000064
Indicating the quality of the corresponding t-th classifier,
Figure FDA0003702520980000065
is a normalization term.
6. The method for classifying the integrated image based on the segmented selection of the deep tree training strategy according to claim 2, wherein in the construction process of the branch classifier, the classifier used is a softmax classifier, an SVM classifier or other classifiers capable of performing multi-classification tasks.
CN201910274548.8A 2019-04-08 2019-04-08 Segmented selection integration image classification method based on deep tree training strategy Active CN110070116B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910274548.8A CN110070116B (en) 2019-04-08 2019-04-08 Segmented selection integration image classification method based on deep tree training strategy

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910274548.8A CN110070116B (en) 2019-04-08 2019-04-08 Segmented selection integration image classification method based on deep tree training strategy

Publications (2)

Publication Number Publication Date
CN110070116A CN110070116A (en) 2019-07-30
CN110070116B true CN110070116B (en) 2022-09-20

Family

ID=67367164

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910274548.8A Active CN110070116B (en) 2019-04-08 2019-04-08 Segmented selection integration image classification method based on deep tree training strategy

Country Status (1)

Country Link
CN (1) CN110070116B (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110956202B (en) * 2019-11-13 2023-08-01 重庆大学 Image training method, system, medium and intelligent device based on distributed learning
CN111367798B (en) * 2020-02-28 2021-05-28 南京大学 Optimization prediction method for continuous integration and deployment results
CN111523638A (en) * 2020-03-10 2020-08-11 中移(杭州)信息技术有限公司 Method, device, terminal and storage medium for measuring generalization capability of deep neural network
CN111461246A (en) * 2020-04-09 2020-07-28 北京爱笔科技有限公司 Image classification method and device
CN111553888B (en) * 2020-04-15 2021-04-27 成都飞机工业(集团)有限责任公司 Titanium alloy forging microstructure image identification method based on machine learning
CN112364926A (en) * 2020-11-17 2021-02-12 苏州大学 Gastroscope picture classification method and device based on ResNet-50 time compression and storage medium
CN114463576B (en) * 2021-12-24 2024-04-09 中国科学技术大学 Network training method based on re-weighting strategy

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107679552A (en) * 2017-09-11 2018-02-09 北京飞搜科技有限公司 A kind of scene classification method and system based on multiple-limb training
CN107784364A (en) * 2016-08-25 2018-03-09 微软技术许可有限责任公司 The asynchronous training of machine learning model
CN108154183A (en) * 2017-12-25 2018-06-12 深圳市唯特视科技有限公司 A kind of objective classification method based on part and depth characteristic set
CN108388917A (en) * 2018-02-26 2018-08-10 东北大学 A kind of hyperspectral image classification method based on improvement deep learning model
CN108445752A (en) * 2018-03-02 2018-08-24 北京工业大学 A kind of random weight Artificial neural network ensemble modeling method of adaptively selected depth characteristic
CN108596329A (en) * 2018-05-11 2018-09-28 北方民族大学 Threedimensional model sorting technique based on end-to-end Deep integrating learning network
CN108830296A (en) * 2018-05-18 2018-11-16 河海大学 A kind of improved high score Remote Image Classification based on deep learning
CN109242865A (en) * 2018-09-26 2019-01-18 上海联影智能医疗科技有限公司 Medical image auto-partition system, method, apparatus and storage medium based on multichannel chromatogram
CN109344905A (en) * 2018-10-22 2019-02-15 王子蕴 A kind of transmission facility automatic fault recognition methods based on integrated study

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7477758B2 (en) * 1992-05-05 2009-01-13 Automotive Technologies International, Inc. System and method for detecting objects in vehicular compartments
JP5950441B2 (en) * 2012-02-01 2016-07-13 日本電産エレシス株式会社 Image recognition apparatus, image recognition method, and image recognition program

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107784364A (en) * 2016-08-25 2018-03-09 微软技术许可有限责任公司 The asynchronous training of machine learning model
CN107679552A (en) * 2017-09-11 2018-02-09 北京飞搜科技有限公司 A kind of scene classification method and system based on multiple-limb training
CN108154183A (en) * 2017-12-25 2018-06-12 深圳市唯特视科技有限公司 A kind of objective classification method based on part and depth characteristic set
CN108388917A (en) * 2018-02-26 2018-08-10 东北大学 A kind of hyperspectral image classification method based on improvement deep learning model
CN108445752A (en) * 2018-03-02 2018-08-24 北京工业大学 A kind of random weight Artificial neural network ensemble modeling method of adaptively selected depth characteristic
CN108596329A (en) * 2018-05-11 2018-09-28 北方民族大学 Threedimensional model sorting technique based on end-to-end Deep integrating learning network
CN108830296A (en) * 2018-05-18 2018-11-16 河海大学 A kind of improved high score Remote Image Classification based on deep learning
CN109242865A (en) * 2018-09-26 2019-01-18 上海联影智能医疗科技有限公司 Medical image auto-partition system, method, apparatus and storage medium based on multichannel chromatogram
CN109344905A (en) * 2018-10-22 2019-02-15 王子蕴 A kind of transmission facility automatic fault recognition methods based on integrated study

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
Cost-sensitive learning of hierarchical tree classifiers for large-scale image classification and novel category detection;Fan jianping 等;《pattern recognition》;20150531;第48卷(第5期);1673-1687 *
基于差异性的分类器集成:有效性分析及优化集成;杨春 等;《自动化学报》;20131219;第40卷(第4期);660-674 *
基于随机梯度法的选择性神经网络二次集成;施彦 等;《计算机工程》;20040820(第16期);133-135,159 *
深度树状训练策略与选择集成算法的研究与应用;胡媛媛;《中国优秀博硕士学位论文全文数据库(硕士)信息科技辑》;20210815;I138-246 *

Also Published As

Publication number Publication date
CN110070116A (en) 2019-07-30

Similar Documents

Publication Publication Date Title
CN110070116B (en) Segmented selection integration image classification method based on deep tree training strategy
CN112308158B (en) Multi-source field self-adaptive model and method based on partial feature alignment
CN113378632B (en) Pseudo-label optimization-based unsupervised domain adaptive pedestrian re-identification method
Suryawati et al. Deep structured convolutional neural network for tomato diseases detection
CN111785329B (en) Single-cell RNA sequencing clustering method based on countermeasure automatic encoder
CN108537264B (en) Heterogeneous image matching method based on deep learning
CN109101938B (en) Multi-label age estimation method based on convolutional neural network
CN108764308A (en) Pedestrian re-identification method based on convolution cycle network
CN107562812A (en) A kind of cross-module state similarity-based learning method based on the modeling of modality-specific semantic space
CN108345860A (en) Personnel based on deep learning and learning distance metric recognition methods again
CN107169117B (en) Hand-drawn human motion retrieval method based on automatic encoder and DTW
CN113076927B (en) Finger vein identification method and system based on multi-source domain migration
CN111178319A (en) Video behavior identification method based on compression reward and punishment mechanism
CN110866134B (en) Image retrieval-oriented distribution consistency keeping metric learning method
CN111523367B (en) Intelligent facial expression recognition method and system based on facial attribute analysis
CN114299542A (en) Video pedestrian re-identification method based on multi-scale feature fusion
CN113011487B (en) Open set image classification method based on joint learning and knowledge migration
CN115049952A (en) Juvenile fish limb identification method based on multi-scale cascade perception deep learning network
CN112733602B (en) Relation-guided pedestrian attribute identification method
CN112633154A (en) Method and system for converting heterogeneous face feature vectors
CN114330650A (en) Small sample characteristic analysis method and device based on evolutionary element learning model training
Karthiga et al. Feature selection using multi-objective modified genetic algorithm in multimodal biometric system
Zhou et al. Multi-objective evolutionary generative adversarial network compression for image translation
Qi et al. Research on deep learning expression recognition algorithm based on multi-model fusion
CN112863650A (en) Cardiomyopathy identification system based on convolution and long-short term memory neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant