CN101231702A - Categorizer integration method - Google Patents

Categorizer integration method Download PDF

Info

Publication number
CN101231702A
CN101231702A CNA2008100467899A CN200810046789A CN101231702A CN 101231702 A CN101231702 A CN 101231702A CN A2008100467899 A CNA2008100467899 A CN A2008100467899A CN 200810046789 A CN200810046789 A CN 200810046789A CN 101231702 A CN101231702 A CN 101231702A
Authority
CN
China
Prior art keywords
classifier
sub
sample
training
weights
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CNA2008100467899A
Other languages
Chinese (zh)
Other versions
CN100587708C (en
Inventor
高常鑫
桑农
王岳环
唐奇伶
李密
高峻
笪邦友
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Streamax Technology Co Ltd
Original Assignee
Huazhong University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huazhong University of Science and Technology filed Critical Huazhong University of Science and Technology
Priority to CN200810046789A priority Critical patent/CN100587708C/en
Publication of CN101231702A publication Critical patent/CN101231702A/en
Application granted granted Critical
Publication of CN100587708C publication Critical patent/CN100587708C/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention provides a classifier integration method, which comprises the following steps that: firstly, the weight of a training sample is initialized; secondly, the classifier training is carried out to the training sample, and a plurality of the optimal sub-classifiers are selected out by adopting the classifier performance evaluation norm which is described by the error seriousness and the generalization ability; thirdly, the optimal sub-classifiers are combined. The classifier performance evaluation norm adopted by the invention can precisely select out the sub-classifiers with good performance, the better of the sub-classifier performance is, the less quantity of the sub-classifiers required by obtaining the classifiers which have the same performance in a combining way is, namely the frequency and the time of the training cycle are less. Besides, the invention also adjusts and combines the classifiers through the feedback, thereby further improving the performance of the classifier.

Description

A kind of categorizer integration method
Technical field
The invention belongs to mode identification method, be specifically related to a kind of by the integrated method that improves the performance of sorter to sub-classifier.
Background technology
Sorter is integrated to be the popular a kind of technology that improves the sorter performance of target.We know the importance of sorter performance in pattern-recognition, but sometimes the precision of single sorter is limited, and categorizer integration method is just by integrated each sorter, construct a more high performance sorter, wherein commonly used is self-service gathering (Bagging, bootstrap aggregation) method and enhancing method (Boosting) method.
Bagging independent random from training set selects a certain number of data to form self-service data set, and each self-service data set all is used to train a sub-classifier independently, and final classification results is chosen in a vote according to the court verdict of these sub-classifiers.
The Boosting method successively produces a series of sorters in training, the employed training set of each sorter all is a subclass that puts forward from total training set, whether each sample appears at the performance of depending on the sorter that produced before this in this subclass, and existing sorter judges that the sample of makeing mistakes will appear at bigger probability in the new training subclass.This makes the sorter that produces thereafter be absorbed in processing comparatively difficult sample concerning existing sorter more and distinguishes problem.
The difference of Bagging and Boosting is that the selection of the training set of Bagging is at random, and each takes turns between the training set separate, and the selection of the training set of Boosting is not independently, and learning outcome of each each wheel of the selection of taking turns training set and front is relevant; Each anticipation function of Bagging does not have weight, and Boosting has weight; The generation that can walk abreast of each anticipation function of Bagging, and each anticipation function of Boosting can only generate in proper order.
Bagging and Boosting can improve the sorter performance effectively, but in the accuracy of most data centralization Boosting than Bagging height, and in great majority are used, accuracy rate is more even more important than arithmetic speed, because it is very fast that the cost performance of computing machine improves, and the training of data is off-lines.Therefore Boosting uses widely than Bagging in the reality.
The Boosting method can strengthen the generalization ability of given algorithm, but also has two shortcomings: this method need be known the lower limit of sub-classifier study accuracy, and this is difficult to accomplish in practical problems; Secondly, this method may cause sorter afterwards to concentrate too much on the minority sample of difficulty especially, causes performance unstable.
Adaptive Boosting method (AdaBoost:Adaptive Boosting) basic idea is to utilize a plurality of sub-classifiers to stack up by certain method, constitutes the strong classifier that classification capacity is very strong.The AdaBoost algorithm is based on the Boosting algorithm.The AdaBoost algorithm no longer needs to know in advance the training error rate of sub-classifier, but comes the training error rate of rudimentary algorithm is adapted to automatically by the weights of each sub-classifier of dynamic adjustments, thereby has caused concern widely.With the Boosting class of algorithms seemingly, the AdaBoost method is adjusted each sample according to existing sorter to the classification situation of each sample in total training set and is appeared at probability in the new training subclass.Different is that AdaBoost does not need to know in advance the scope of sub-classifier predictablity rate, but sets corresponding weights according to the predictablity rate of sub-classifier automatically.When using the AdaBoost training, the training error rate upper limit is a dull function that descends, like this, as long as surpassing at random that the performance of basic sorter can be stable guessed, and the circulation long enough time, just can make the experience rate reduce to low level arbitrarily, and might guarantee that extensive error rate also is lower than an approximate upper limit.It makes the performance of sub-classifier get a promotion by setting up Combination of Multiple Classifiers, because it is exclusive to the self-adaptation of sorter performance with to crossing the immunity of learning phenomenon, has caused great concern in recent years, and has been used widely in target detection.
AdaBoost has superiority in theory very much, but AdaBoost also has a lot of fuzzy places in actual applications: how (1) selects best sub-classifier; (2) how better to make up these sub-classifiers.
Summary of the invention
The objective of the invention is to propose a kind of categorizer integration method, use more effective sorter performance evaluation criterion to select the good sub-classifier of performance, reduce sorter training time and cycle index.
A kind of categorizer integration method, concrete steps are:
(1) gives training sample weights initialize;
(2) training sample is carried out the sorter training and obtain C best sub-classifier, C is a frequency of training;
(2.1) make t=1 cycle of training;
(2.2) adopt the weights normalization method to determine the weights of each training sample when the training of t wheel;
(2.3) take predtermined category device method for designing to generate corresponding sub-classifier at each feature of training sample, therefrom select best sub-classifier, and calculate the error rate that adopts best sub-classifier that all training sample classification are produced;
(2.4) upgrade the training sample weights according to error rate;
(2.5) if t<C, t=t+1 returns step (2.2), otherwise enters step (3);
(3) C best sub-classifier of combination;
It is characterized in that,
Described step (2.3) adopts one of following dual mode to select best sub-classifier:
The sub-classifier h of j feature correspondence of i, calculation training sample jExtent of error
em j = 1 2 Σ i = 1 n w t , i | h j ( x i ) - y i | | f j ( x i ) - θ j | , Wherein n represents the number of training sample, x iThe expression training sample, w T, iThe weights of i sample when representing the training of t wheel, f j(x i) presentation class device h jTo sample x iClassification response, θ jThe presentation class threshold value;
The sub-classifier of selecting minimal error degree correspondence is as best sub-classifier.
The sub-classifier h of j feature correspondence of ii, calculation training sample jExtent of error
em j = 1 2 Σ i = 1 n w t , i | h j ( x i ) - y i | | f j ( x i ) - θ j | , Wherein n represents the number of training sample, x iThe expression training sample,
Figure S2008100467899D00044
w T, iThe weights of i sample when representing the training of t wheel, f j(x i) presentation class device h jTo sample x iClassification response, θ jThe presentation class threshold value;
Calculate the sub-classifier h of j feature correspondence jGeneralization ability
Figure S2008100467899D00045
, f wherein j(x +) expression sub-classifier h jTo separating class threshold value θ jOr the classification response of the nearest positive sample of classifying face, f j(x -) expression sub-classifier h jTo separating class threshold value θ jOr the classification response of the nearest negative sample of classifying face;
Miscount degree summation Error j=em j+ G j, select minimal error degree summation Error jCorresponding sub-classifier is as best sub-classifier.
The sorter performance evaluation criterion that the present invention adopts can be selected the good sub-classifier of performance exactly, and the sub-classifier performance is good more, and the sub-classifier quantity that makes up the sorter needs that obtain identical performance is just few more, promptly trains round-robin number of times and time all few more.In addition, the present invention also by the feedback adjusting assembled classifier, further strengthens the sorter performance.
Description of drawings
Fig. 1 is a process flow diagram of the present invention;
Fig. 2 is the sorter performance evaluation criterion synoptic diagram of hierarchical description;
Fig. 3 is the distribution situation of two groups of experimental datas, and the positive negative sample of Fig. 3 (a) is desirable evenly distribution, and the positive negative sample of Fig. 3 (b) is a Gaussian distribution;
Fig. 4 is to use sorter error rate interpretational criteria that the classification performance of experimental data is described, and Fig. 4 (a) is to classification of Data performance specification shown in Fig. 3 (a), and Fig. 4 (b) is to classification of Data performance specification shown in Fig. 3 (b);
Fig. 5 is to use the sorter performance specification of the sorter performance evaluation criterion of extent of error tolerance to experimental data, and Fig. 5 (a) is to classification of Data performance specification shown in Fig. 3 (a), and Fig. 5 (b) is to classification of Data performance specification shown in Fig. 3 (b);
Fig. 6 is the sorter performance specification of the sorter performance evaluation criterion of generalization ability to experimental data, and Fig. 6 (a) is to classification of Data performance specification shown in Fig. 3 (a), and Fig. 6 (b) is to classification of Data performance specification shown in Fig. 3 (b);
Fig. 7 is to use the sorter performance specification of the sorter performance evaluation criterion of hierarchy to experimental data, and Fig. 7 (a) is to classification of Data performance specification shown in Fig. 3 (a), and Fig. 7 (b) is to classification of Data performance specification shown in Fig. 3 (b);
Fig. 8 is an embodiment of the invention classification based training sample distribution synoptic diagram;
Fig. 9 is through assembled classifier performance of optimizing and the assembled classifier performance comparison diagram of not optimizing;
Figure 10 is through the assembled classifier performance of optimizing and being made of ten above sub-classifiers and the assembled classifier performance comparison diagram of optimization not.
Embodiment
Steps flow chart of the present invention now illustrates as shown in Figure 1:
If import n training sample (x i, y i), i=1 ..., n, wherein x iThe characteristic value collection of expression training sample, i.e. x i={ x I1, x I2..., x Iq, the feature of an element representation in the set, this feature may be that numerical value also may be vector, q is the number of feature;
Figure S2008100467899D00061
M negative sample arranged, l positive sample in the known training sample.
(1) design sub-classifier step.Determine the building method of sub-classifier according to known training sample, the sub-classifier form that j feature of training sample generates is
Figure S2008100467899D00062
H wherein jThe classification results of () expression sub-classifier, can only be ± 1, h j()=1 expression sub-classifier classification results is positive sample, h j()=-1 expression sub-classifier classification results is a negative sample; f j() is the classification response of sub-classifier to sample, i.e. f when being characterized as a numerical value j() representation feature value is when being characterized as f when vectorial j() representation feature is to the distance of this sub-classifier classifying face; θ jBe the sub-classifier threshold value, just obtained θ when obtaining a sub-classifier j, it is difference along with the difference of sub-classifier method for designing; p jGet ± 1, work as p j=1 o'clock, p jf j()<p jθ jExpression f j()<θ j, work as p j=-1 o'clock, p jf j()<p jθ jExpression f j()>θ jThe sub-classifier that obtains sorter threshold value or classifying face can have different methods for designing, as neural net method, least squares error method (MSE), support vector machine (SVM) and hidden Markov model (HMM) etc.
(2) initialization training sample weights step.The initialization error weight, w 1, iThe weights initial value of representing i sample is for y i=-1 sample w 1 , i = 1 2 m ; For y i=1 sample w 1 , i = 1 2 l , Do not adopt all samples to put the method for identical weights herein, but positive negative sample respectively account for 1/2.
(3) training classifier step.For each t=1 ... C, wherein C is the number of times of training, C macrotaxonomy device performance more is good more, can reach high precision arbitrarily, the accuracy that generalized case just can obtain near 90% for about 50 times along with C increases in theory.
(3.1) weights normalization, w t , i ← w t , i Σ v = 1 n w t , v , Make all weights and be 1, w T, iThe weights of i sample when representing the training of t wheel;
(3.2) make j=1 ... q generates corresponding sub-classifier h at j feature in the characteristic set according to the method for (1) j, calculate the sorter performance metric of current weight, use the sorter performance evaluation criterion of hierarchical description to estimate the sub-classifier quality;
(3.3) selection has the sub-classifier of top performance
Figure S2008100467899D00074
Join in the strong classifier and go, and calculate its error rate ϵ t = 1 2 = Σ i = 1 n w t , i | h ‾ t ( x i ) - y i | , The expression sub-classifier
Figure S2008100467899D00077
Classification results to i sample;
(3.4) upgrade the pairing weights of each sample:
w t + 1 , i = w t , i β t 1 - e i
Wherein, w T+1, iThe weights of representing i the sample in t wheel training renewal back; If
Figure S2008100467899D00082
To i sample x iCorrect classification, then e i=0, on the contrary e i=1, β t = ϵ t 1 - ϵ t .
(4) assembled classifier step.The strong classifier that generates is:
Figure S2008100467899D00084
Wherein best sub-classifier weights α t = log 1 β t .
(5) optimize the sorter step.By the method for feedback, adjust the sorter coefficient, strengthen the sorter performance.The present invention uses the feedforward neural network method of feedback, and the training error of tolerance network is adjusted neural network weight according to error, makes the sorter performance reach optimum.
The sub-classifier formal description of the present invention's design is as follows:
The design of sorter is actually a kind of strategy of optimizing, the least squares error method is to seek the sorter of square error minimum, the Fisher linear discriminant is a sorter of seeking divergence ratio maximum between class scatter and the class, neural network is to seek error function sorter hour, the sorter of maximization class interval during support vector machine.Equality when but each sample is considered in the design of these sorters, but under the AdaBoost framework, closely-related during the performance of the weights of sample and front sorter, those wrong samples that divides should more be absorbed in the design of sorter, be the bigger samples of weights, the weights wrong cost of dividing of sample that varies in size is different, so the present invention least squares error method of using realizes that promptly j feature generates corresponding sub-classifier h according to the least squares error method jClassification function is that weights d by each component of classifying face describes in the least squares error method.With m tRepresent the number of samples that sampling obtains in the current training, then the j of current sample is characterized as x Rjsj, 0<r<m t, 0<s j<g j, g jIt is the dimension of j feature.D={d then s, 0<s j<g jThe sub-classifier of j feature correspondence obtains by following method:
(1) the preliminary classification function d is set 0, be set to less non-zero random number, b={b r, 0<r<m tThe allowance of expression sample, the present invention is with b rAll be changed to 1;
(2) the new sample of input;
(3) u input samples are x uj = { x uj s j } , 0 < s j < g j ;
(4) revise classification function, press the following formula correction:
d←d+η(u)ω u(b u-d tx uj)x uj
(5) if the error in classification of classification function less than certain threshold value, learning process finished when perhaps iterations reached certain value, otherwise forwarded (2) to.
The adjusting range of classification function adopts η (u)=η (0)/(u+1), wherein η (0)=0.002 in the u time iteration of η (u) control among the present invention; ω uBe the weights of u input sample; d u tExpression d uTransposition; The end of learning with iterations control among the present invention, iteration stopping when iterations reaches 10000 times.
Above-mentioned algorithm has just obtained the sub-classifier of j feature correspondence, the described sub-classifier design formula of corresponding step (1)
Figure S2008100467899D00092
&theta; j = 0 , Be d tx UjThe sub-classifier of representing j feature correspondence at>0 o'clock is positive sample, d to u sample classification result tx UjThe sub-classifier of representing j feature correspondence at<0 o'clock is negative sample to u sample classification result.
Can guarantee that like this sample is identical, obtain different classification under the different situation of corresponding weights, could guarantee so can always not select identical sorter, and this algorithm is fit to the situation that is characterized as numerical value and is characterized as the vector.
The sorter performance evaluation criterion of the hierarchical description that the present invention proposes is described below:
Therefore how the quality of sorter performance is most important in pattern classification, and the performance of interpretive classification device is directly connected to the quality of classification.In the present invention, good more description criterion guarantees that the sub-classifier performance that each circulation obtains is good more, and it is just few that combination obtains the sorter quantity that the sorter of identical performance needs, and the round-robin number of times is just few, has reduced the time of training.
The structural representation of the sorter performance evaluation criterion of hierarchical description as shown in Figure 2.
During the actual enforcement of the present invention, as long as find best sorter, and do not need all sorter performances are all described in detail, for the interpretational criteria of sorter is described, we are described in detail each sorter performance, and just do not find best sub-classifier.Sorter performance criteria for different situations is described below:
(1) error rate of different under many circumstances sub-classifiers is different, uses the performance that the error rate performance criterion just can the interpretive classification device;
(2) when a plurality of sub-classifier of same error rate, original error rate performance criterion is the quality of interpretive classification device well, the performance of using the performance evaluation criterion that adds extent of error to come the interpretive classification device;
(3) when the performance evaluation criterion of the adding extent of error of a plurality of sub-classifiers is identical, add the performance that generalization ability comes the interpretive classification device, promptly use to add the sorter interpretational criteria that extent of error tolerance and generalization ability are described combination.
Last one deck of the sorter performance evaluation criterion of hierarchical description performance of descriptor sorter fully just like this.
The superiority of the sorter performance criteria that proposes among the present invention is described with simple DATA DISTRIBUTION below.Fig. 3 has shown the distribution situation of two groups of data, and positive negative sample shown in Fig. 3 (a) is that ideal uniform distributes, and the positive negative sample of Fig. 3 (b) is a Gaussian distribution.Wherein negative sample from 0.1 to 1.1 evenly distributes among Fig. 3 (a), and positive sample from 0.9 to 1.9 evenly distributes; Negative sample is that average is that 0.6 variance is 0.1 Gaussian distribution among Fig. 3 (b), and positive sample is that average is that 1.4 variances are 0.1 Gaussian distribution.
The performance criteria of sorter is the error rate of classification in the former algorithm, adopts the sub-classifier h of j feature correspondence jError rate to all sample classifications is described ε jFor:
&epsiv; j = 1 2 &Sigma; i n w t , i | h j ( x i ) - y i |
Wherein n represents the number of sample; x iThe eigenwert of expression training sample; y i=1, corresponding negative sample of 1} and positive sample; w T, iThe weights of i sample when representing the training of t wheel; h j(x i) sorter of j feature correspondence of expression is to the classification results of i sample; y iThe classification of representing i sample, promptly i sample is positive sample or negative sample.Weights with each sample in once circulating are all fixed, and are irrelevant with feature.Use this criterion as shown in Figure 4, Fig. 4 (a), Fig. 3 (a) in Fig. 4 (b) difference correspondence, Fig. 3 (b) figure to above-mentioned classification of Data device performance specification.
Only consider that as can be seen from Figure 4 the sorter performance that error rate is described is poor, may can not get best sorter at the error rate same case, the sorter of the middle minimal error rate correspondence of Fig. 4 (a) and Fig. 4 (b) all has a plurality of, and the error rate of non-minimum also has a plurality of situations among Fig. 4 (b).When the present invention then attempts using the mistake of statistics sample, add extent of error tolerance.The sub-classifier h of j feature correspondence jAdding extent of error tolerance em is described jForm is as follows:
em j = 1 2 &Sigma; i n w t , i | h j ( x i ) - y i | | f j ( x i ) - &theta; j |
Wherein n represents the number of sample; x iThe expression training sample; y i=1, corresponding negative sample of 1} and positive sample; w T, iThe weights of i sample when representing the training of t wheel; f j(x i) expression sub-classifier h jTo sample x iClassification response; θ jThe presentation class threshold value; | f j(x i)-θ j| big more expression separates the class threshold value or the classifying face distance is far away more, and then extent of error is big more.Use this criterion as shown in Figure 5 to the sorter performance specification of raw data, (a) (b) respectively (a) in the corresponding diagram 3 (b) scheme.
Found desirable sorter among Fig. 5 (a) as can be seen, but Fig. 5 (b) still can not obtain best sorter.Therefore under the identical but non-vanishing situation of sample error rate, adding extent of error tolerance is a feasible method, but the sample error rate then lost efficacy when being all zero.The present invention has considered the generalization ability description of sorter, add the tolerance of sorter generalization ability when the sample error rate is zero, and when the sample error rate was non-vanishing, generalization ability is described as constant, and the judgment criterion form is as follows:
G wherein jThe sub-classifier h that represents j feature correspondence jGeneralization ability is described; f j(x +) expression separates the response of the nearest positive sample of class threshold value or classifying face to current sub-classifier; f j(x -) expression separates the response of the nearest negative sample of class threshold value or classifying face to current sub-classifier; θ jThe presentation class threshold value.When generalization ability is described G jHour sorter performance is good more more.Use this criterion to the sorter performance specification of raw data as shown in Figure 6, Fig. 6 (a), Fig. 6 (b) be the Fig. 3 (a) in the corresponding diagram 3 respectively, Fig. 3 (b).
The sorter performance evaluation criterion form that adding extent of error tolerance of last one deck of sorter performance evaluation criterion of whole layering and generalization ability are described combination is as follows:
Error j=em j+G j
Error wherein jThe sub-classifier h that represents j feature correspondence jThe sorter performance that adding extent of error tolerance of last one deck of hierarchical description and generalization ability are described combination, em jThe performance specification that adds extent of error tolerance, G jRepresent this sorter generalization ability description.Error jMore little, its corresponding sub-classifier performance is good more.The sorter performance evaluation criterion of whole layering is described as shown in Figure 7, Fig. 7 (a), the Fig. 3 (a) in Fig. 7 (b) difference corresponding diagram 3, Fig. 3 (b).
From the above-mentioned experiment sorter performance evaluation criterion performance of interpretive classification device better of hierarchical description as can be seen.
By each sub-classifier coefficient of feedback adjusting, improve the classification performance of sorter integral body at last.The principle of passing through the better combination of feedback sub-classifier that the present invention proposes is as follows:
The method for designing that does not have theory to prove to depend in the AdaBoost classical way to obtain the sample error rate of sorter is optimum, and clearly Zui You classifiers combination method can reduce the number from sorter.Therefore need to seek optimum classifiers combination method, the present invention uses the feedforward neural network method of feedback, and the error of tolerance network is adjusted neural network weight according to error, makes the sorter performance reach optimum.
The neural network that the present invention uses adopts the learning algorithm that supervision is arranged, and it is known promptly being used for the classification of the sample learnt.When importing learning sample successively, network is revised weights with the deviation of desired output according to neuronic actual output with iterative manner, finally obtains the weights of expecting.In actual the enforcement, use the single-layer perceptron network to realize that the sub-classifier of selecting is as a neuron of perceptron.Sorter form before optimization of the present invention's combination is
Z ( ) = 1 &Sigma; t = 1 C &alpha; t h &OverBar; t ( ) &GreaterEqual; 0 0 otherwise . Assembled classifier weights α={ α t, t=1 ..., C, α tThe optimal classification device weights that expression is used to make up.Total C the neuron of the single-layer perceptron network that the present invention uses, use this method refreshing weight α specific algorithm as follows:
(1) obtains initial weight α={ α t, t=1 ..., C makes k=1;
(2) input learning sample x k
(3) calculate neuronic actual output.If the learning sample of the k time input is x k, the weights that are connected with t neuron are α t, t neuronic actual being output as For the t that selects in the training classifier step best sub-classifier to sample x kClassification results.
(4) refreshing weight.Because each sample has only two kinds of situations of positive negative sample, so C neuronic desired output is y k, weights are pressed following formula and are upgraded:
&alpha; = &alpha; + &lambda; &Sigma; t = 1 C ( y k - h &OverBar; t ( x k ) ) x k ;
(5) if adopt weights α to obtain correct classification results, finish when perhaps k is more than or equal to iterations, otherwise k=k+1 forwards step (2) to.
Wherein adjust the size that parameter lambda ∈ (0,1) control weights are adjusted, if the λ value is too big, vibration may appear in algorithm; If the λ value is too little, speed of convergence is very slow.Study end when k surpasses iterations 2000 among the present invention.
The experimental data that the present invention uses is described below: totally 5000 samples, and positive and negative each 2500, distribute as shown in Figure 8, each sample is made up of two features, and the dimension of each feature is 1.According to flow process of the present invention, use the least squares error method to design sub-classifier, sorter performance evaluation criterion with layering is selected best sorter, the present invention has compared the sorter performance of optimizing and not have to optimize then: through (the present invention λ in this experiment gets 0.002) classifiers combination performance after the Feedback Neural Network optimization and the classifiers combination performance that do not have optimization more as shown in Figure 9, the sample data of corresponding diagram 8.Figure 10 shows above the comparable situation behind ten sorters.
From the experimental result of Fig. 9 and Figure 10 as can be seen, sorter after optimization classification back error sample number little than before optimizing, and it is apparent in view in the sorter before optimization to increase situation about increasing behind the sub-classifier, sorter after the optimization has then slowed down a lot, optimizes in theory and can eliminate this situation fully.Behind feedback adjusting sorter coefficient, the performance of sorter is improved.

Claims (2)

1. categorizer integration method may further comprise the steps:
(1) gives training sample weights initialize;
(2) training sample is carried out the sorter training and obtain C best sub-classifier, C is a frequency of training;
(2.1) make t=1 cycle of training;
(2.2) adopt the weights normalization method to determine the weights of each training sample when the training of t wheel;
(2.3) take predtermined category device method for designing to generate corresponding sub-classifier at each feature of training sample, therefrom select best sub-classifier, and calculate the error rate that adopts best sub-classifier that all training sample classification are produced;
(2.4) upgrade the training sample weights according to error rate;
(2.5) if t<C, t=t+1 returns step (2.2), otherwise enters step (3);
(3) C best sub-classifier of combination;
It is characterized in that,
Described step (2.3) adopts one of following dual mode to select best sub-classifier:
The sub-classifier h of j feature correspondence of i, calculation training sample jExtent of error
em j = 1 2 &Sigma; i = 1 n w t , i | h j ( x i ) - y i | | f j ( x i ) - &theta; j | , Wherein n represents the number of training sample, x iThe expression training sample,
Figure S2008100467899C00012
w T, iThe weights of i sample when representing the training of t wheel, f j(x i) presentation class device h jTo sample x iClassification response, θ jThe presentation class threshold value;
The sub-classifier of selecting minimal error degree correspondence is as best sub-classifier.
The sub-classifier h of j feature correspondence of ii, calculation training sample jExtent of error
em j = 1 2 &Sigma; i = 1 n w t , i | h j ( x i ) - y i | | f j ( x i ) - &theta; j | , Wherein n represents the number of training sample, x iThe expression training sample,
Figure S2008100467899C00022
w T, iThe weights of i sample when representing the training of t wheel, f j(x i) presentation class device h jTo sample x iClassification response, θ jThe presentation class threshold value;
Calculate the sub-classifier h of j feature correspondence jGeneralization ability
Figure S2008100467899C00023
, f wherein j(x +) expression sub-classifier h jTo separating class threshold value θ jOr the classification response of the nearest positive sample of classifying face, f j(x -) expression sub-classifier h jTo separating class threshold value θ jOr the classification response of the nearest negative sample of classifying face;
Miscount degree summation Error j=em j+ G j, select minimal error degree summation Error jCorresponding sub-classifier is as best sub-classifier.
2. categorizer integration method according to claim 1 is characterized in that, described method also comprises step (4): adjust the assembled classifier weights, be specially,
(4.1) obtain assembled classifier weights α={ α t, t=1 ..., C, α tThe best sub-classifier weights that expression is used to make up make k=1;
(4.2) input known class training sample x k
(4.3) obtain the employing assembled classifier to training sample x kClassification results
Figure S2008100467899C00024
(4.4) refreshing weight &alpha; = &alpha; + &lambda; &Sigma; t = 1 C ( y k - h &OverBar; t ( x k ) ) x k , λ is the adjustment parameter,
Figure S2008100467899C00026
The best sub-classifier of selecting in t cycle of training for described step (2) is to sample x kClassification results, y kThe best sub-classifier of selecting in t cycle of training for described step (2) is to sample x kThe expectation classification results;
(4.5) if adopt weights α to obtain correct classification results or k, finish, otherwise k=k+1 forwards step (4.2) to more than or equal to iterations.
CN200810046789A 2008-01-25 2008-01-25 Categorizer integration method Active CN100587708C (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN200810046789A CN100587708C (en) 2008-01-25 2008-01-25 Categorizer integration method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN200810046789A CN100587708C (en) 2008-01-25 2008-01-25 Categorizer integration method

Publications (2)

Publication Number Publication Date
CN101231702A true CN101231702A (en) 2008-07-30
CN100587708C CN100587708C (en) 2010-02-03

Family

ID=39898161

Family Applications (1)

Application Number Title Priority Date Filing Date
CN200810046789A Active CN100587708C (en) 2008-01-25 2008-01-25 Categorizer integration method

Country Status (1)

Country Link
CN (1) CN100587708C (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101799925A (en) * 2010-03-05 2010-08-11 华中科技大学 Performance analysis method for automatic segmentation result of image
CN101799875A (en) * 2010-02-10 2010-08-11 华中科技大学 Target detection method
CN102231144A (en) * 2011-06-03 2011-11-02 中国电力科学研究院 Method for predicting theoretical line loss of power distribution network based on Boosting algorithm
CN102270279A (en) * 2011-07-27 2011-12-07 华北电力大学 Short-term power load predicting method
CN104102917A (en) * 2014-07-03 2014-10-15 中国石油大学(北京) Construction method of domain self-adaptive classifier, construction device for domain self-adaptive classifier, data classification method and data classification device
CN104462409A (en) * 2014-12-12 2015-03-25 重庆理工大学 Cross-language emotional resource data identification method based on AdaBoost
CN104598981A (en) * 2013-10-31 2015-05-06 国际商业机器公司 Generating device, generating method and program
CN107092878A (en) * 2017-04-13 2017-08-25 中国地质大学(武汉) It is a kind of based on hybrid classifer can autonomous learning multi-target detection method
CN107463798A (en) * 2017-08-02 2017-12-12 南京高新生物医药公共服务平台有限公司 Predict the 12 gene expressions classification device and its construction method of adenocarcinoma of colon prognosis
WO2018219016A1 (en) * 2017-06-02 2018-12-06 腾讯科技(深圳)有限公司 Facial detection training method, apparatus and electronic device
CN109945900A (en) * 2019-03-11 2019-06-28 南京智慧基础设施技术研究院有限公司 A kind of distributed optical fiber sensing method
CN109948680A (en) * 2019-03-11 2019-06-28 合肥工业大学 The classification method and system of medical record data
CN112612897A (en) * 2020-12-30 2021-04-06 湖北大学 Wikipedia concept dependency relationship identification method
CN114202524A (en) * 2021-12-10 2022-03-18 中国人民解放军陆军特色医学中心 Performance evaluation method and system of multi-modal medical image
CN116226744A (en) * 2023-03-16 2023-06-06 中金同盛数字科技有限公司 User classification method, device and equipment

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101799875A (en) * 2010-02-10 2010-08-11 华中科技大学 Target detection method
CN101799875B (en) * 2010-02-10 2011-11-30 华中科技大学 Target detection method
CN101799925A (en) * 2010-03-05 2010-08-11 华中科技大学 Performance analysis method for automatic segmentation result of image
CN102231144A (en) * 2011-06-03 2011-11-02 中国电力科学研究院 Method for predicting theoretical line loss of power distribution network based on Boosting algorithm
CN102231144B (en) * 2011-06-03 2016-02-24 中国电力科学研究院 A kind of power distribution network method for predicting theoretical line loss based on Boosting algorithm
CN102270279A (en) * 2011-07-27 2011-12-07 华北电力大学 Short-term power load predicting method
CN104598981A (en) * 2013-10-31 2015-05-06 国际商业机器公司 Generating device, generating method and program
CN104102917A (en) * 2014-07-03 2014-10-15 中国石油大学(北京) Construction method of domain self-adaptive classifier, construction device for domain self-adaptive classifier, data classification method and data classification device
CN104102917B (en) * 2014-07-03 2017-05-10 中国石油大学(北京) Construction method of domain self-adaptive classifier, construction device for domain self-adaptive classifier, data classification method and data classification device
CN104462409B (en) * 2014-12-12 2017-08-25 重庆理工大学 Across language affection resources data identification method based on AdaBoost
CN104462409A (en) * 2014-12-12 2015-03-25 重庆理工大学 Cross-language emotional resource data identification method based on AdaBoost
CN107092878A (en) * 2017-04-13 2017-08-25 中国地质大学(武汉) It is a kind of based on hybrid classifer can autonomous learning multi-target detection method
WO2018219016A1 (en) * 2017-06-02 2018-12-06 腾讯科技(深圳)有限公司 Facial detection training method, apparatus and electronic device
US10929644B2 (en) 2017-06-02 2021-02-23 Tencent Technology (Shenzhen) Company Limited Face detection training method and apparatus, and electronic device
US11594070B2 (en) 2017-06-02 2023-02-28 Tencent Technology (Shenzhen) Company Limited Face detection training method and apparatus, and electronic device
CN107463798A (en) * 2017-08-02 2017-12-12 南京高新生物医药公共服务平台有限公司 Predict the 12 gene expressions classification device and its construction method of adenocarcinoma of colon prognosis
CN109945900A (en) * 2019-03-11 2019-06-28 南京智慧基础设施技术研究院有限公司 A kind of distributed optical fiber sensing method
CN109948680A (en) * 2019-03-11 2019-06-28 合肥工业大学 The classification method and system of medical record data
CN109948680B (en) * 2019-03-11 2021-06-11 合肥工业大学 Classification method and system for medical record data
CN112612897A (en) * 2020-12-30 2021-04-06 湖北大学 Wikipedia concept dependency relationship identification method
CN114202524A (en) * 2021-12-10 2022-03-18 中国人民解放军陆军特色医学中心 Performance evaluation method and system of multi-modal medical image
CN116226744A (en) * 2023-03-16 2023-06-06 中金同盛数字科技有限公司 User classification method, device and equipment

Also Published As

Publication number Publication date
CN100587708C (en) 2010-02-03

Similar Documents

Publication Publication Date Title
CN100587708C (en) Categorizer integration method
WO2022007376A1 (en) Multi-objective multimodal particle swarm optimization method based on bayesian adaptive resonance
CN105589806A (en) SMOTE+Boosting algorithm based software defect tendency prediction method
CN104361393A (en) Method for using improved neural network model based on particle swarm optimization for data prediction
CN102567742A (en) Automatic classification method of support vector machine based on selection of self-adapting kernel function
CN106919951A (en) A kind of Weakly supervised bilinearity deep learning method merged with vision based on click
CN105868775A (en) Imbalance sample classification method based on PSO (Particle Swarm Optimization) algorithm
Wu et al. Some analysis and research of the AdaBoost algorithm
CN107451651A (en) A kind of driving fatigue detection method of the H ELM based on particle group optimizing
Schinas et al. CERTH@ MediaEval 2012 Social Event Detection Task.
CN104809230A (en) Cigarette sensory quality evaluation method based on multi-classifier integration
CN105117770A (en) Surface cooler control method based on improved fuzzy neural network
CN113762387B (en) Multi-element load prediction method for data center station based on hybrid model prediction
CN104809502A (en) Dynamic adjusting method for nodes in hidden layers of restricted Boltzmann machines
Quek et al. A novel approach to the derivation of fuzzy membership functions using the Falcon-MART architecture
CN113991751A (en) Automatic power generation control scheduling method based on hybrid algorithm
CN109934351A (en) A kind of quantum learning aid algorithm and the modified fuzzy sliding mode controlling method based on quantum learning aid algorithm
CN107578101B (en) Data stream load prediction method
CN107273922A (en) A kind of screening sample and weighing computation method learnt towards multi-source instance migration
CN105184486A (en) Power grid business classification method based on directed acyclic graphs support vector machine
CN102103691A (en) Identification method for analyzing face based on principal component
CN107341479A (en) A kind of method for tracking target based on the sparse coordination model of weighting
Li et al. Deep learning for energy efficient beamforming in MU-MISO networks: A GAT-based approach
CN105550711A (en) Firefly algorithm based selective ensemble learning method
Zhang et al. Intelligent recognition of mixture control chart pattern based on quadratic feature extraction and SVM with AMPSO

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
ASS Succession or assignment of patent right

Owner name: SHENZHEN RUIMING TECHNOLOGY CO., LTD.

Free format text: FORMER OWNER: HUAZHONG SCINECE AND TECHNOLOGY UNIV

Effective date: 20100804

C41 Transfer of patent application or patent right or utility model
COR Change of bibliographic data

Free format text: CORRECT: ADDRESS; FROM: 430074 NO.1037, LUOYU ROAD, HONGSHAN DISTRICT, WUHAN CITY, HUBEI PROVINCE TO: 518057 5/F, BUILDING 3, SHENZHEN SOFTWARE PARK, KEJIZHONGER ROAD, NEW + HIGH-TECHNOLOGY ZONE, NANSHAN DISTRICT, SHENZHEN CITY, GUANGDONG PROVINCE

TR01 Transfer of patent right

Effective date of registration: 20100804

Address after: 518057, five floor, building three, Shenzhen Software Park, two road, Nanshan District hi tech Zone, Shenzhen, Guangdong

Patentee after: Shenzhen Streaming Video Technology Co., Ltd.

Address before: 430074 Hubei Province, Wuhan city Hongshan District Luoyu Road No. 1037

Patentee before: Huazhong University of Science and Technology

PE01 Entry into force of the registration of the contract for pledge of patent right

Denomination of invention: Categorizer integration method

Effective date of registration: 20130110

Granted publication date: 20100203

Pledgee: Shenzhen SME credit financing guarantee Group Co., Ltd.

Pledgor: Shenzhen Streaming Video Technology Co., Ltd.

Registration number: 2013990000024

PLDC Enforcement, change and cancellation of contracts on pledge of patent right or utility model
PC01 Cancellation of the registration of the contract for pledge of patent right

Date of cancellation: 20140318

Granted publication date: 20100203

Pledgee: Shenzhen SME credit financing guarantee Group Co., Ltd.

Pledgor: Shenzhen Streaming Video Technology Co., Ltd.

Registration number: 2013990000024

PE01 Entry into force of the registration of the contract for pledge of patent right

Denomination of invention: Categorizer integration method

Effective date of registration: 20140318

Granted publication date: 20100203

Pledgee: Shenzhen SME credit financing guarantee Group Co., Ltd.

Pledgor: Shenzhen Streaming Video Technology Co., Ltd.

Registration number: 2014990000174

PLDC Enforcement, change and cancellation of contracts on pledge of patent right or utility model
PC01 Cancellation of the registration of the contract for pledge of patent right

Date of cancellation: 20150528

Granted publication date: 20100203

Pledgee: Shenzhen SME credit financing guarantee Group Co., Ltd.

Pledgor: Shenzhen Streaming Video Technology Co., Ltd.

Registration number: 2014990000174

PLDC Enforcement, change and cancellation of contracts on pledge of patent right or utility model
PE01 Entry into force of the registration of the contract for pledge of patent right

Denomination of invention: Categorizer integration method

Effective date of registration: 20150603

Granted publication date: 20100203

Pledgee: Shenzhen SME financing Company limited by guarantee

Pledgor: Shenzhen Streaming Video Technology Co., Ltd.

Registration number: 2015990000430

PLDC Enforcement, change and cancellation of contracts on pledge of patent right or utility model
C56 Change in the name or address of the patentee

Owner name: SHENZHEN STREAMAX TECHNOLOGY CO., LTD.

Free format text: FORMER NAME: SHENZHEN RUIMING TECHNOLOGY CO., LTD.

CP03 Change of name, title or address

Address after: Nanshan District Xueyuan Road in Shenzhen city of Guangdong province 518000 No. 1001 Nanshan Chi Park building B1 building 21-23

Patentee after: STREAMAX TECHNOLOGY CO., LTD.

Address before: 518057, five floor, building three, Shenzhen Software Park, two road, Nanshan District hi tech Zone, Shenzhen, Guangdong

Patentee before: Shenzhen Streaming Video Technology Co., Ltd.

DD01 Delivery of document by public notice

Addressee: Chen Dan

Document name: Notification of Passing Examination on Formalities

PC01 Cancellation of the registration of the contract for pledge of patent right

Date of cancellation: 20160718

Granted publication date: 20100203

Pledgee: Shenzhen SME financing Company limited by guarantee

Pledgor: STREAMAX TECHNOLOGY CO., LTD.

Registration number: 2015990000430

PLDC Enforcement, change and cancellation of contracts on pledge of patent right or utility model
PM01 Change of the registration of the contract for pledge of patent right

Change date: 20160718

Registration number: 2015990000430

Pledgor after: STREAMAX TECHNOLOGY CO., LTD.

Pledgor before: Shenzhen Streaming Video Technology Co., Ltd.