CN113449779B - SVM incremental learning method based on sample distribution density improved KKT condition - Google Patents
SVM incremental learning method based on sample distribution density improved KKT condition Download PDFInfo
- Publication number
- CN113449779B CN113449779B CN202110652246.7A CN202110652246A CN113449779B CN 113449779 B CN113449779 B CN 113449779B CN 202110652246 A CN202110652246 A CN 202110652246A CN 113449779 B CN113449779 B CN 113449779B
- Authority
- CN
- China
- Prior art keywords
- sample
- classifier
- model
- kkt condition
- samples
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 38
- 239000013598 vector Substances 0.000 claims abstract description 58
- 238000012549 training Methods 0.000 claims abstract description 33
- 230000006872 improvement Effects 0.000 claims abstract description 4
- 238000012706 support-vector machine Methods 0.000 claims description 64
- 230000008569 process Effects 0.000 claims description 8
- 238000005457 optimization Methods 0.000 claims description 5
- 230000000153 supplemental effect Effects 0.000 claims description 3
- 230000001502 supplementing effect Effects 0.000 claims description 2
- 238000004422 calculation algorithm Methods 0.000 abstract description 15
- 238000012216 screening Methods 0.000 abstract description 2
- 238000010276 construction Methods 0.000 abstract 1
- 230000006870 function Effects 0.000 description 3
- 230000008859 change Effects 0.000 description 2
- 238000010801 machine learning Methods 0.000 description 2
- 239000011159 matrix material Substances 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000004088 simulation Methods 0.000 description 2
- OYPRJOBELJOOCE-UHFFFAOYSA-N Calcium Chemical compound [Ca] OYPRJOBELJOOCE-UHFFFAOYSA-N 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000007635 classification algorithm Methods 0.000 description 1
- 238000013145 classification model Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 230000014759 maintenance of location Effects 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 239000013589 supplement Substances 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2411—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
- G06N20/20—Ensemble learning
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Physics & Mathematics (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Medical Informatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention discloses a sample distributionAn SVM incremental learning method for density improvement KKT conditions, comprising: obtaining a set of support vectors SV for a classifier 0 Standard KKT condition I of classifier; construction of SVM classifier Model old Improved KKT condition I; judging whether samples in the newly added sample set B meet the Model of the SVM classifier old Standard KKT condition i; performing secondary judgment on samples in the newly added sample set B to judge whether the Model of the SVM classifier is met old Improved KKT condition I; from a set of candidate support vectors SV 1 Training classifier Model 1 The method comprises the steps of carrying out a first treatment on the surface of the Computing a Model that satisfies a classifier 1 Improved KKT condition II; by SV 0 ∪SV 1 ∪SV add Training the classifier and outputting updated classifier Model 2 . The improved KKT condition based on the sample distribution density can realize effective screening of the newly added samples under the condition of unbalanced sample distribution, and improves the generalization capability of the SVM incremental learning algorithm.
Description
Technical Field
The invention belongs to the technical field of machine learning, and relates to an SVM incremental learning method based on sample distribution density improved KKT conditions. The SVM incremental learning method is particularly used for SVM incremental learning under the condition of uneven sample distribution, and can be used for online updating of SVM classifier under automatic incremental learning.
Background
The support vector machine (Support Vector Machine, SVM) is a machine learning pattern recognition classification algorithm proposed by Vapnik in the 90 th century, reference [ Vapnik V.statistical learning theory, new York: john Wiley & Sons, inc.,1998], which is excellent in classification tasks of small sample, high dimensional features. The conventional SVM algorithm is a batch learning mode, that is, it is assumed that all training samples can be obtained at one time before training, and the learning process is terminated after training is completed. However, in practical applications, it is generally impossible to obtain the training samples all at once, but the training samples are obtained gradually over time, and the information contained in the new training samples may change over time. Therefore, the classifier needs to have the ability to continuously learn useful knowledge from these sample data, thereby realizing online updating of the classifier under the condition of newly added samples.
How to learn useful knowledge from newly added sample data and ensure that a model updated through learning training has better classification performance becomes an important problem to be solved. This problem can be solved by incrementally learning to retain important information for the newly added sample. The incremental learning idea is summarized as follows: based on the original knowledge base, only the change caused by the newly added data is updated to the original knowledge base. This will save a significant amount of training time and memory requirements after the new sample data is added.
The SVM incremental learning algorithm proposed by Syed et al is an early more classical algorithm, and the basic idea of the algorithm is to only retain the support vector of the original classification model and perform incremental training with newly added data samples in the references [ Syed N a, liu H, sun k.increment learning with support vector machines.proc.int Joint Conference on Artificial Intelligence,1999 ]. The algorithm does not screen the newly added samples, and some newly added samples without gain effect on classification accuracy are trained, so that the efficiency of incremental learning is poor and the classification accuracy is affected. Then, researchers introduce a Karush-Kuhn-Tucker (KKT) condition to screen newly added samples, and an SVM incremental learning algorithm based on the KKT condition is provided. In 2014, zhang Canlin and the like introduce a new sample mischaracterizing error driving idea based on KKT conditions, and propose a new SVM increment learning algorithm, reference [ Zhang Canlin, yao Minghai, tong Xiaolong, and the like ]. The SVM incremental learning algorithm based on the KKT condition can screen newly added samples, but has poor generalization capability. Under the condition of uneven distribution of the newly added samples, the existing SVM increment learning method based on the KKT condition can not carry out self-adaptive screening on the newly added samples with obviously inconsistent distribution density, and under the condition of unbalanced sample distribution, the classification precision of the SVM classifier updated by the method is obviously lower.
Disclosure of Invention
Aiming at the defects of the prior art, the invention aims to provide an SVM increment learning method based on sample distribution density improvement KKT condition, so as to solve the problems of poor learning ability on newly added samples, low generalization ability of a classifier and low classification precision of an SVM increment learning algorithm under the condition of unbalanced sample distribution in the prior art.
The method of the invention can be used for updating the classifier by judging the KKT condition of the newly added sample, and pre-selecting a possible support vector set SV, but analyzing the rule of the number of the non-support vectors near the classification boundary under different sample distribution densities finds that the larger the sample distribution density is, the more the non-support vectors near the classification boundary are, the more the samples are likely to be converted into the support vectors of the new classifier, the smaller the sample distribution density is, the fewer the non-support vectors near the classification boundary are, the samples are likely to be converted into the support vectors of the new classifier are, if the bias parameters of fixed values are introduced to improve the KKT condition, and under the condition of unbalanced sample distribution, the learning degree of the sample set with low sample density on the newly added samples is obviously insufficient compared with the learning degree of the sample set with high sample density on the newly added samples.
On the basis of an SVM increment learning method based on the KKT condition, the bias parameter is calculated through sample distribution density self-adaption so as to improve the KKT condition, the improved KKT condition is adopted to automatically screen candidate support vector sets SV from positive and negative sample sets in newly added samples, sample learning of low sample density is improved, sample learning of high sample density is balanced, and under the condition of fully utilizing historical training results, quick increment learning of newly added samples is realized, so that the generalization capability and classification precision of a classifier are improved.
The invention relates to an SVM increment learning method based on sample distribution density improved KKT condition, which comprises the following steps:
1) Training an SVM classifier Model from an original sample set A old Obtaining classifier Model old Support vector set SV of (1) 0 And standard KKT condition I;
2) Calculating sample distribution density of positive and negative samples in an original sample set A, calculating bias parameters under the positive and negative samples according to the sample distribution density, adding the bias parameters on the basis of a standard KKT condition I, and constructing an SVM classifier Model old Self-adaptation to positive and negative samplesImproved KKT condition I should be optimized;
3) Judging whether all samples in the newly added sample set B meet Model of SVM (support vector machine) classifier old If all the conditions are satisfied, outputting the Model of the original SVM classifier old The model is the required model, and the process is finished; otherwise, putting samples which violate the standard KKT condition I in the newly added sample set B into a set B1, and putting samples which meet the standard KKT condition I into a set B2;
4) Judging whether samples in the newly added sample set B meet the Model of the SVM classifier old Is performed by placing samples satisfying the modified KKT condition I into the candidate support vector set SV 1 Defined as possible support vector samples;
5) According to the candidate support vector set SV 1 Training classifier Model 1 Obtaining classifier Model 1 Standard KKT condition II of (1) using support vector set SV 1 Calculating bias parameters under positive and negative samples to obtain an improved KKT condition II after the bias parameters are added;
6) Judging whether the set B2 meets the improved KKT condition II, if the set B2 is empty or all samples meet the improved KKT condition II, outputting an updated Model which is the classifier Model 1 Ending; otherwise, putting the sample which does not meet the modified KKT condition II into a supplementary support vector set SV add ;
7) By SV 0 ∪SV 1 ∪SV add Model of ensemble training classifier 2 And outputs the updated classifier as a Model 2 ,SV 0 SV is the set of support vectors for the original classifier add To supplement the support vector set.
Further, in the step 1), training an SVM classifier Model according to the original sample set old The LIBSVM kit is used in the process, and the standard KKT condition I is expressed as follows:
in the solving process of the optimal hyperplane optimal solution of the SVM classifier, the optimal hyperplane optimal solution is determinedThe meaning condition is 0.ltoreq.alpha ι The standard KKT condition is abbreviated as y i f(x i ) Is less than or equal to 1; wherein C is a penalty factor, α= (α) 1 ,α 2 ,...,α n ) T T is the transpose of the matrix, y, for Lagrangian multiplier i ∈[1,-1]For sample labels, f (x i ) For sample x i Distance from the optimal hyperplane.
Further, in the step 2), bias parameters under positive and negative samples are calculated according to the distribution density of the samples, and a classifier Model is constructed old The method for improving KKT condition I is specifically as follows:
21 Assumed original training setFor the original sample set, the positive sample is +.>Negative sample is +.>The class centers of positive and negative samples in the sample set are c respectively + And c - ,/>Representing the distance between the positive and negative samples and the center of the class;
22 Calculating the sample distribution density xi of the positive and negative sample sets + With xi - :
23 Calculating the bias parameter delta under positive and negative samples + ,δ - :
δ + =1-ξ + ,δ - =1-ξ - ;
24 Calculating Model of SVM classifier old Is defined in KKT condition I:
|y i f(x i )|≤1+δ=1+(1-ξ),ξ∈[ξ + ,ξ - ]
wherein N is + Represent the positive number of samples, N - Represents the negative number of samples, and max (·) represents the distance maximum.
Further, in the step 3), for the newly added sample set b= { s 1 ,s 2 ,...,s j ,...,s M Sample vector s in } j Judging if s j All meet the standard KKT condition I, output classifier Model old Ending; otherwise, dividing the newly added sample set B into two sets of a set B1 and a set B2, wherein the Model of the SVM classifier is satisfied old Sample vectors of standard KKT condition i are placed in set B2 and sample vectors that violate standard KKT condition i are placed in set B1.
Further, in the step 5), the LIBSVM tool box is adopted to implement the method according to the candidate support vector set SV 1 Training classifier Model 1 Calculating a candidate support vector set SV 1 Sample distribution density of the positive and negative samples, calculating bias parameters under the positive and negative samples according to the sample distribution density, adding the bias parameters on the basis of standard KKT condition II, and constructing a Model of the SVM classifier 1 Improved KKT condition II aiming at the self-adaptive optimization of positive and negative samples is provided.
Further, in the step 7), the support vector set SV of the original classifier is obtained 0 A set of possible support vectors SV selected from the newly added sample set B 1 Supplemental support vector set SV add Merging and combining SV by using LIBSVM tool box 0 ∪SV 1 ∪SV add Training classifier Model 2 The classifier Model 2 Is the classifier model found.
The invention has the beneficial effects that:
1. the invention considers that the existing SVM increment learning method based on the improved KKT condition often adopts fixed offset parameters to improve the KKT condition to realize the selection of the newly added sample, and does not consider the special condition that the sample distribution is obviously unbalanced or the sample set is smaller to influence the selection of the newly added sample, so that the classification precision of the classifier after increment learning updating is lower.
2. The invention provides a new improved KKT condition SVM increment learning method, which introduces sample distribution density parameter self-adaptive calculation of bias parameters of improved KKT conditions, so that the improved KKT conditions can more effectively select candidate support vector sets SV from newly added samples under the condition of unbalanced sample distribution or smaller sample set, and the classification precision of a classifier is improved.
3. The invention improves the learning efficiency of the increment sample, and under the condition of obtaining the same target classification precision, the increment learning frequency is less, and the time consumption of online updating of the classifier is reduced.
Drawings
FIG. 1 is a schematic diagram of the method of the present invention.
Fig. 2 is a ROC graph of the 10-time classification result for the abalone dataset increment.
FIG. 3 is a ROC graph of the 10-degree classification result for a Waveform dataset increment.
FIG. 4 is a graph showing the classification accuracy of a wave form dataset as a function of the number of increments.
Detailed Description
The invention will be further described with reference to examples and drawings, to which reference is made, but which are not intended to limit the scope of the invention.
Referring to fig. 1, the SVM incremental learning method based on sample distribution density improving KKT condition of the present invention comprises the following steps:
1) Training an SVM classifier Model from an original sample set A old Obtaining classifier Model old Support vector set SV of (1) 0 And standard KKT condition I;
training an SVM classifier Model according to the original sample set in the step 1) old The LIBSVM kit is used in the process, and the standard KKT condition I is expressed as follows:
optimal hyper-flatness of SVM classifierIn the process of solving the face optimal solution, defining a condition of 0-alpha ι The standard KKT condition is abbreviated as y i f(x i ) Is less than or equal to 1; wherein C is a penalty factor, α= (α) 1 ,α 2 ,...,α n ) T T is the transpose of the matrix, y, for Lagrangian multiplier i ∈[1,-1]For sample labels, f (x i ) For sample x i Distance from the optimal hyperplane.
2) Calculating sample distribution density of positive and negative samples in an original sample set A, calculating bias parameters under the positive and negative samples according to the sample distribution density, adding the bias parameters on the basis of a standard KKT condition I, and constructing an SVM classifier Model old The improved KKT condition I aiming at the self-adaptive optimization of positive and negative samples is adopted;
in the step 2), the bias parameters of the positive and negative samples are calculated according to the distribution density of the samples, and a classifier Model is constructed old The method for improving KKT condition I is specifically as follows:
21 Assumed original training setFor the original sample set, the positive sample is +.>Negative sample is +.>The class centers of positive and negative samples in the sample set are c respectively + And c - ,/>Representing the distance between the positive and negative samples and the center of the class;
22 Calculating the sample distribution density xi of the positive and negative sample sets + With xi - :
23 Calculating the bias parameter delta under positive and negative samples + ,δ - :
δ + =1-ξ + ,δ - =1-ξ _ ;
24 Calculating Model of SVM classifier old Is defined in KKT condition I:
|y i f(x i )|≤1+δ=1+(1-ξ),ξ∈[ξ + ,ξ _ ]
wherein N is + Represent the positive number of samples, N _ Represents the negative number of samples, and max (·) represents the distance maximum.
3) Judging whether all samples in the newly added sample set B meet Model of SVM (support vector machine) classifier old If all the conditions are satisfied, outputting the Model of the original SVM classifier old The model is the required model, and the process is finished; otherwise, putting samples which violate the standard KKT condition I in the newly added sample set B into a set B1, and putting samples which meet the standard KKT condition I into a set B2;
in the step 3), for the newly added sample set b= { s 1 ,s 2 ,...,s j ,...,s M Sample vector s in } j Judging if s j All meet the standard KKT condition I, output classifier Model old Ending; otherwise, dividing the newly added sample set B into two sets of a set B1 and a set B2, wherein the Model of the SVM classifier is satisfied old Sample vectors of standard KKT condition i are placed in set B2 and sample vectors that violate standard KKT condition i are placed in set B1.
4) Judging whether samples in the newly added sample set B meet the Model of the SVM classifier old Is performed by placing samples satisfying the modified KKT condition I into the candidate support vector set SV 1 Defined as possible support vector samples.
5) According to the candidate support vector set SV 1 Training classifier Model 1 Obtaining classifier Model 1 Standard KKT condition II of (1) using support vector set SV 1 Calculating bias parameters under positive and negative samples to obtain an improved KKT condition II after the bias parameters are added;
l is adopted in the step 5)The IBSVM tool box is based on the candidate support vector set SV 1 Training classifier Model 1 Calculating a candidate support vector set SV 1 Sample distribution density of the positive and negative samples, calculating bias parameters under the positive and negative samples according to the sample distribution density, adding the bias parameters on the basis of standard KKT condition II, and constructing a Model of the SVM classifier 1 Improved KKT condition II aiming at the self-adaptive optimization of positive and negative samples is provided.
6) Judging whether the set B2 meets the improved KKT condition II, if the set B2 is empty or all samples meet the improved KKT condition II, outputting an updated Model which is the classifier Model 1 Ending; otherwise, putting the sample which does not meet the modified KKT condition II into a supplementary support vector set SV add 。
7) By SV 0 ∪SV 1 ∪SV add Model of ensemble training classifier 2 And outputs the updated classifier as a Model 2 ,SV 0 SV is the set of support vectors for the original classifier add Supplementing a support vector set;
the support vector set SV of the original classifier in the step 7) 0 A set of possible support vectors SV selected from the newly added sample set B 1 Supplemental support vector set SV add Merging and combining SV by using LIBSVM tool box 0 ∪SV 1 ∪SV add Training classifier Model 2 The classifier Model 2 Is the classifier model found.
The data adopted by the embodiment of the invention are an abalone data set and a Waveform data set in UCI standard classification data set. Wherein, the abalone classification data set contains 4177 examples, and the attribute quantity of each example is 8. The Waveform classification dataset contained 5000 instances, each with a number of attributes of 21. The preparation work in the early stage is as follows: 4177 examples in the abalone classification dataset were randomly aliquoted into 10, 1 of which was the original sample set, 1 of which was the test sample set, and 8 of which was the incremental sample set at 8 increments. The same segmentation method was applied to split the Waveform classification dataset equally into 10 aliquots. The SVM training function employed in the operating program employs the libsvmtrain function of the LIBSVM toolbox. Comparing the incremental learning SVM algorithm of the present invention with other methods, the simulation results are shown in FIG. 2 and FIG. 3 below.
As shown in fig. 2 and 3, a subject operating characteristic curve (Receiver Operating characteristic Curve, ROC) curve of classification results on UCI standard dataset abalone and Waveform, wherein the ordinate true positive rate (True Positive Rate, TPR) of the ROC curve indicates the ratio of correctly judged positive samples among all the actually positive samples; the abscissa false positive rate (False Positive Rate, FPR) represents the rate at which positive samples are erroneously judged to be among all the samples that are actually negative; the Area Under the Curve is the Area Under Curve (AUC) index, and the higher the AUC value is, the better the classifier performance is. The algorithm of the invention is shown in the figure by a modified KKT-ISVM curve, and the other two methods are respectively a KKT-based incremental SVM algorithm (KKT-ISVM) and a combined retention set-based incremental SVM algorithm (Combined Reserved Set Incremental Support Vector Machine, CRS-ISVM). FIG. 4 shows the classification accuracy of different methods on the Waveform dataset varying with the number of incremental learning, the higher the classification accuracy, the better the classifier performance.
According to the ROC curves of multiple incremental learning under two simulation comparison experiments and the AUC index comparison results of the area representation under the curves, the ROC curves of the improved KKT incremental SVM algorithm (improved KKT-ISVM) in the four methods are closer to the upper left, the area AUC indexes under the curves are the highest, the higher the AUC is used for indicating higher classification accuracy, the higher the classifier performance is, and the method provided by the invention is proved to be effective and has advanced performance.
The present invention has been described in terms of the preferred embodiments thereof, and it should be understood by those skilled in the art that various modifications can be made without departing from the principles of the invention, and such modifications should also be considered as being within the scope of the invention.
Claims (4)
1. An SVM incremental learning method for improving KKT conditions based on sample distribution density is characterized by comprising the following steps:
1) Training an SVM classifier Model from the raw sample set A old Obtaining classifier Model old Support vector set SV of (1) 0 And standard KKT condition I;
2) Calculating sample distribution densities of positive and negative samples in an original sample set A, calculating bias parameters under the positive and negative samples according to the sample distribution densities, adding the bias parameters on the basis of a standard KKT condition I, and constructing an SVM classifier Model old The improved KKT condition I aiming at the self-adaptive optimization of positive and negative samples is adopted;
3) Judging whether all samples in the newly added sample set B meet Model of SVM (support vector machine) classifier old If all the conditions are satisfied, outputting the Model of the original SVM classifier old The model is the required model, and the process is finished; otherwise, putting samples which violate the standard KKT condition I in the newly added sample set B into a set B1, and putting samples which meet the standard KKT condition I into a set B2;
4) Judging whether samples in the newly added sample set B meet the Model of the SVM classifier old Is performed by placing samples satisfying the modified KKT condition I into the candidate support vector set SV 1 ;
5) According to the candidate support vector set SV 1 Training classifier Model 1 Obtaining classifier Model 1 Standard KKT condition II of (1) using support vector set SV 1 Calculating bias parameters under positive and negative samples to obtain an improved KKT condition II after the bias parameters are added;
6) Judging whether the set B2 meets the improved KKT condition II, if the set B2 is empty or all samples meet the improved KKT condition II, outputting an updated Model which is the classifier Model 1 Ending; otherwise, putting the sample which does not meet the modified KKT condition II into a supplementary support vector set SV add ;
7) By SV 0 ∪SV 1 SV add Model of ensemble training classifier 2 And outputs the updated classifier as a Model 2 ,SV 0 SV is the set of support vectors for the original classifier add Supplementing a support vector set;
the step 2) is based on a sample distribution densitometerCalculating bias parameters under positive and negative samples, and constructing classifier Model old The method for improving KKT condition I is specifically as follows:
21 Assumed original training setFor the original sample set, the positive sample is +.>Negative sample is +.>The class centers of positive and negative samples in the sample set are c respectively + And c - ,/>And->Representing the distance between the positive and negative samples and the center of the class;
22 Calculating the sample distribution density xi of the positive and negative sample sets + With xi - :
23 Calculating the bias parameter delta under positive and negative samples + ,δ - :
δ + =1-ξ + ,δ - =1-ξ - ;
24 Calculating Model of SVM classifier old Is defined in KKT condition I:
|y i f(x i )|≤1+δ=1+(1-ξ),ξ∈[ξ + ,ξ - ]
wherein N is + Represents the positive number of samples, N-represents the negative number of samples, and max (·) represents the distance maximum.
2. The SVM incremental learning method based on sample distribution density improvement KKT condition of claim 1, wherein in said step 3), for a new added sample set b= { s 1 ,s 2 ,...,s j ,...,s M Sample vector s in } j Judging if s j All meet the standard KKT condition I, output classifier Model old Ending; otherwise, dividing the newly added sample set B into two sets of a set B1 and a set B2, wherein the Model of the SVM classifier is satisfied old Sample vectors of standard KKT condition i are placed in set B2 and sample vectors that violate standard KKT condition i are placed in set B1.
3. The method for learning SVM increments based on the sample distribution density improvement KKT condition according to claim 1, wherein the step 5) uses LIBSVM tool kit according to the candidate support vector set SV 1 Training classifier Model 1 Calculating a candidate support vector set SV 1 Sample distribution density of the positive and negative samples, calculating bias parameters under the positive and negative samples according to the sample distribution density, adding the bias parameters on the basis of standard KKT condition II, and constructing a Model of the SVM classifier 1 Improved KKT condition II aiming at the self-adaptive optimization of positive and negative samples is provided.
4. The method for SVM delta learning based on sample distribution density refinement KKT conditions of claim 1, wherein in said step 7) the support vector set SV of the original classifier is 0 A set of possible support vectors SV selected from the newly added sample set B 1 Supplemental support vector set SV add Merging and combining SV by using LIBSVM tool box 0 ∪SV 1 SV add Training classifier Model 2 The classifier Model 2 Is the classifier model found.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110652246.7A CN113449779B (en) | 2021-06-11 | 2021-06-11 | SVM incremental learning method based on sample distribution density improved KKT condition |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110652246.7A CN113449779B (en) | 2021-06-11 | 2021-06-11 | SVM incremental learning method based on sample distribution density improved KKT condition |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113449779A CN113449779A (en) | 2021-09-28 |
CN113449779B true CN113449779B (en) | 2024-04-16 |
Family
ID=77811441
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110652246.7A Active CN113449779B (en) | 2021-06-11 | 2021-06-11 | SVM incremental learning method based on sample distribution density improved KKT condition |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113449779B (en) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101944122A (en) * | 2010-09-17 | 2011-01-12 | 浙江工商大学 | Incremental learning-fused support vector machine multi-class classification method |
CN109190719A (en) * | 2018-11-30 | 2019-01-11 | 长沙理工大学 | Support vector machines learning method, device, equipment and computer readable storage medium |
CN111160457A (en) * | 2019-12-27 | 2020-05-15 | 南京航空航天大学 | Turboshaft engine fault detection method based on soft class extreme learning machine |
US10970650B1 (en) * | 2020-05-18 | 2021-04-06 | King Abdulaziz University | AUC-maximized high-accuracy classifier for imbalanced datasets |
-
2021
- 2021-06-11 CN CN202110652246.7A patent/CN113449779B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101944122A (en) * | 2010-09-17 | 2011-01-12 | 浙江工商大学 | Incremental learning-fused support vector machine multi-class classification method |
CN109190719A (en) * | 2018-11-30 | 2019-01-11 | 长沙理工大学 | Support vector machines learning method, device, equipment and computer readable storage medium |
CN111160457A (en) * | 2019-12-27 | 2020-05-15 | 南京航空航天大学 | Turboshaft engine fault detection method based on soft class extreme learning machine |
US10970650B1 (en) * | 2020-05-18 | 2021-04-06 | King Abdulaziz University | AUC-maximized high-accuracy classifier for imbalanced datasets |
Also Published As
Publication number | Publication date |
---|---|
CN113449779A (en) | 2021-09-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110378799B (en) | Alumina comprehensive production index decision method based on multi-scale deep convolution network | |
CN107506865B (en) | Load prediction method and system based on LSSVM optimization | |
CN107392919B (en) | Adaptive genetic algorithm-based gray threshold acquisition method and image segmentation method | |
CN101968853B (en) | Improved immune algorithm based expression recognition method for optimizing support vector machine parameters | |
CN109801295B (en) | Image segmentation method based on secondary limited region growing method | |
CN116503676B (en) | Picture classification method and system based on knowledge distillation small sample increment learning | |
CN112836735B (en) | Method for processing unbalanced data set by optimized random forest | |
CN115578248B (en) | Generalized enhanced image classification algorithm based on style guidance | |
CN102800051A (en) | Self-adaptation flotation froth image enhancement method based on fractional order differential | |
CN116452862A (en) | Image classification method based on domain generalization learning | |
CN117290721A (en) | Digital twin modeling method, device, equipment and medium | |
CN110399917B (en) | Image classification method based on hyper-parameter optimization CNN | |
CN113449779B (en) | SVM incremental learning method based on sample distribution density improved KKT condition | |
CN111930484B (en) | Power grid information communication server thread pool performance optimization method and system | |
CN116805389A (en) | Open world target detection method based on decoupling cascade region generation network | |
CN110070184A (en) | Merge the data sampling method of sample losses and optimal speed constraint | |
CN115690476A (en) | Automatic data clustering method based on improved harmony search algorithm | |
CN104298987A (en) | Handwritten numeral recognition method based on point density weighting online FCM clustering | |
CN114972261A (en) | Method for identifying surface quality defects of plate strip steel | |
CN114444614A (en) | Self-adaptive multi-mean two-step clustering method | |
Feng | Automatic Density Peaks Clustering based on the Cosine Similarity | |
Maoxiang et al. | Steel Plate Surface Defects Classification Method using Multiple Hyper-planes Twin Support Vector Machine with Additional Information. | |
CN117975444B (en) | Food material image recognition method for food crusher | |
CN116363420A (en) | Rapid K neighbor classifier method for large-scale data | |
CN114821206B (en) | Multi-modal image fusion classification method and system based on confrontation complementary features |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |