CN104991974A - Particle swarm algorithm-based multi-label classification method - Google Patents

Particle swarm algorithm-based multi-label classification method Download PDF

Info

Publication number
CN104991974A
CN104991974A CN201510464344.2A CN201510464344A CN104991974A CN 104991974 A CN104991974 A CN 104991974A CN 201510464344 A CN201510464344 A CN 201510464344A CN 104991974 A CN104991974 A CN 104991974A
Authority
CN
China
Prior art keywords
particle
sample
algorithm
distance
label
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201510464344.2A
Other languages
Chinese (zh)
Inventor
梁庆中
樊媛媛
姚宏
颜雪松
胡成玉
曾德泽
刘超
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China University of Geosciences
Original Assignee
China University of Geosciences
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China University of Geosciences filed Critical China University of Geosciences
Priority to CN201510464344.2A priority Critical patent/CN104991974A/en
Publication of CN104991974A publication Critical patent/CN104991974A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification

Abstract

The invention provides a particle swarm algorithm-based multi-label classification method. The particle swarm algorithm-based multi-label classification method includes an optimization stage and a classification stage. According to the optimization stage, a particle swarm algorithm is adopted to optimize the feature weight of a feature weighting KNN algorithm. According to the classification stage, the feature weight obtained in the optimization stage is applied to the feature weighting KNN algorithm, so that a test sample X can be classified, and finally, the labels of all samples in a test set can be outputted, namely, classification is completed. With the particle swarm algorithm-based multi-label classification method of the invention adopted, an optimal feature weight can be found out to eliminate the redundancy or irrelevance of the features in a data set which are attribute values when a distance is calculated, and therefore, distance deviation can be decreased, and the accuracy of classification can be improved.

Description

A kind of method of many labelings based on particle cluster algorithm
Technical field
The invention belongs to many labelings technical field, be specifically related to a kind of many labelings based on particle cluster algorithm method.
Background technology
The research of many labelings problem is promoted by text classification, and modern, many practical applications are all many labelings problems, and such as scene classification, protein functional assays, separated film and music are sorted out.The sample that many label datas are concentrated has multiple label, and how setting up and solve such optimization problem is the major issue that will solve.Though the realization of algorithm has certain difficulty, its advantage is that it does not change the structure of data set, does not destroy the incidence relation between classification, reflects the special nature of many labelings.According to the distinct methods setting up optimization problem, this algorithm also can be divided into multiple different form, as: maximize entropy algorithm based on many labelings algorithm of Adaboost algorithm, the many labelings algorithm using traditional decision-tree expansion, many labels algorithm of support vector machine, many labels k nearest neighbor algorithm (KNN algorithm), many labels.But these algorithms can due to the redundancy of eigenwert or uncorrelated and cause the error of calculation.
Summary of the invention
One of the object of the invention is the defect for overcoming prior art, provides a kind of method of the many labelings based on particle cluster algorithm that degree of accuracy is high.
A kind of method of many labelings based on particle cluster algorithm provided by the invention, comprises optimizing phase and sorting phase:
S10: the optimizing phase is the feature weight adopting particle cluster algorithm to optimize characteristic weighing KNN algorithm, specifically comprises the steps:
S11: adopt random device initialization population, the position of each particle and the dimension of speed are n, the feature weight vector w=(w that its position corresponding data concentrates to record 1, w 2, w n): Qi Zhongyou
Σ i - 1 n w i = 1 ;
S12: calculate adaptive value, and then obtain locally optimal solution and globally optimal solution:
When calculating adaptive value, the position of particle and feature weight are applied in characteristic weighing KNN algorithm, former training sample is concentrated front 70% as new training sample, rear 30% as new forecast sample collection, this forecast sample collection is classified, calculate the accuracy of classification, accuracy is more high more meets adaptive value;
The original tag that forecast sample integrates every bar record is as li=(li1, li2, lin), and the prediction label after sorted is lj=(lj1, lj2, ljn), li and l jbe sum, so accuracy rate Accruay=sum/n with the number that overlaps;
S20: sorting phase:
The feature weight drawn optimizing phase is applied in characteristic weighing KNN algorithm classifies to test sample book X, and the final label exporting all samples in test set, has namely classified.
Further, described particle cluster algorithm comprises the steps:
SA1: initialization Particle Swarm, comprising the position xi=(x of the whole population of initialization i1, x i2, x id) twith speed v i=(v i1, v i2, v id) tand local optimum and total optimization, wherein id represents d particle in the i-th generation.
SA2: calculate the fitness value fitness of each particle in current position id=f (x id).Then according to the size of fitness value, initialization locally optimal solution pbest i=fitness iwith total optimization solution gbest=min (fintess 1, fitness 2, fitness n), i=1,2, N;
SA3: in each iterative process, each particle upgrades position and the speed of oneself according to following criterion
v id(t+1)=wv id(t)+c 1r 1(p ld-x id(t))+c 2r 2(p gd-x id(t))
x id(t+1)=x id(t)+v id(t+1)
Wherein v idfor the speed of particle, x idfor the position of particle, w is inertia weight, c1 and c2 is acceleration factor, r 1and r 2random number, P ldfor globally optimal solution and P gdfor locally optimal solution;
SA4: upgrade locally optimal solution pbest iwith total optimization solution gbest;
SA5: if total optimization solution gbest reaches the threshold value of setting or reached maximum iteration time, algorithm can stop calculating; Otherwise jump to step SA3.
Further, described characteristic weighing KNN algorithm specifically comprises the steps:
SB1: input m training sample, and set k value size;
SB2: the A [1] in first Stochastic choice training set ~ A [k] sample is as the initial most adjacent node of the k of sample X to be predicted;
SB3: the weighting Euclidean distance wd (X, A [i]) calculating the most adjacent node of sample X to be predicted and each initial k, i=1,2 ... .., k), calculating range formula is:
w d ( X , A [ i ] ) = Σ l = 1 n w l ( X l - A [ i ] l ) 2 ,
Wherein n represents sample A [i] attribute number, i.e. A [i]=(A [i] 1, A [i] 2, A [i] 3, A [i] n);
SB4: will the distance wd (X, A [i]) obtained be asked in described step SB3 by ascending sort, try to achieve the maximum distance maxD=max{d (X, A [i]) of distance wd (X, A [i]) | i=1,2 ... .., k};
SB5: calculation training concentrates the distance of remaining record and sample to be tested X successively, and compared with the maximum distance maxD tried to achieve in described step SB4, if less than maximum distance maxD, then maximum distance maxD is updated to the distance value of this record and sample to be tested X, and again by ascending order adjust the distance wd (X, A [i]) sequence;
SB6: the occurrence number calculating the label of the every bar record in present range wd (X, A [i]) sequence, and sort according to the height of occurrence number;
SB7: using the label of L label before sequence in described step SB6 obtains as sample X.
Beneficial effect of the present invention is, this method can find optimum feature weight to eliminate the redundancy of the feature (referring to property value when calculating distance) of data centralization or uncorrelated, thus decreases range deviation, improves the accuracy of classification.
Accompanying drawing explanation
Figure 1 shows that the many labelings method flow diagram that the present invention is based on particle cluster algorithm.
Embodiment
Hereafter will describe the present invention in detail in conjunction with specific embodiments.It should be noted that the combination of technical characteristic or the technical characteristic described in following embodiment should not be considered to isolated, they can mutually be combined thus be reached better technique effect.
As shown in Figure 1, a kind of method of many labelings based on particle cluster algorithm provided by the invention comprises optimizing phase and sorting phase:
Optimizing phase is the feature weight adopting population (Particle Swarm Optimization, PSO) algorithm optimization characteristic weighing KNN algorithm, and concrete steps are as follows:
S10: the optimizing phase is the feature weight adopting particle cluster algorithm to optimize characteristic weighing KNN algorithm, specifically comprises the steps:
S11: adopt random device initialization population, the position of each particle and the dimension of speed are n, the feature weight vector w=(w that its position corresponding data concentrates to record 1, w 2, w n): Qi Zhongyou
Σ i - 1 n w i = 1 ;
S12: calculate adaptive value, and then obtain locally optimal solution and globally optimal solution:
When calculating adaptive value, the position of particle and feature weight are applied in characteristic weighing KNN algorithm, former training sample is concentrated front 70% as new training sample, rear 30% as new forecast sample collection, this forecast sample collection is classified, calculate the accuracy of classification, accuracy is more high more meets adaptive value;
The original tag that forecast sample integrates every bar record is as li=(li1, li2, lin), and the prediction label after sorted is lj=(lj1, lj2, ljn), li and l jbe sum, so accuracy rate Accruay=sum/n with the number that overlaps;
S20: sorting phase:
The feature weight drawn optimizing phase is applied in characteristic weighing KNN algorithm classifies to test sample book X, and the final label exporting all samples in test set, completes classification.
Particle cluster algorithm is the one belonging to evolution algorithmic, and be a kind of optimized algorithm based on iteration, system initialization is one group of RANDOM SOLUTION, by iterated search optimal value.But it does not use intersection (crossover) and variation (mutation), but the particle that particle follows optimum in solution space is searched for, the advantage of PSO is simple easily realization and does not have many parameters to need adjustment.Have a lot of particle in its each population, each particle has its position x and speed v Two Variables, and often produce in the new population of a generation and have the position of a particle best, this particle is exactly the locally optimal solution pbest of this generation i, from locally optimal solution, produce globally optimal solution gbest.
Particle cluster algorithm comprises the steps:
SA1: initialization Particle Swarm, comprising the position xi=(x of the whole population of initialization i1, x i2, x id) twith speed v i=(v i1, v i2, v id) tand local optimum and total optimization, wherein id represents d particle in the i-th generation.
SA2: calculate the fitness value fitness of each particle in current position id=f (x id).Then according to the size of fitness value, initialization locally optimal solution pbest i=fitness iwith total optimization solution gbest=min (fintess 1, fitness 2, fitness n), i=1,2, N.
SA3: in each iterative process, each particle upgrades position and the speed of oneself according to following criterion:
v id(t+1)=wv id(t)+c 1r 1(p ld-x id(t))+c 2r 2(p gd-x id(t))
x id(t+1)=x id(t)+v id(t+1)
Wherein v idfor the speed of particle, x idfor the position of particle, w is inertia weight, c1 and c2 is acceleration factor, r 1and r 2random number, P ldfor globally optimal solution and P gdfor locally optimal solution;
SA4: upgrade locally optimal solution pbest iwith total optimization solution gbest;
SA5: if total optimization solution gbest reaches the threshold value of setting or reached maximum iteration time, algorithm can stop calculating; Otherwise jump to step SA3.
Characteristic weighing KNN algorithm specifically comprises the steps:
SB1: input m training sample, and set k value size;
SB2: the A [1] in first Stochastic choice training set ~ A [k] sample is as the initial most adjacent node of the k of sample X to be predicted;
SB3: the weighting Euclidean distance wd (X, A [i]) calculating the most adjacent node of sample X to be predicted and each initial k, i=1,2 ... .., k), calculating range formula is:
w d ( X , A [ i ] ) = Σ l = 1 n w l ( X l - A [ i ] l ) 2 ,
Wherein n represents sample A [i] attribute number, i.e. A [i]=(A [i] 1, A [i] 2, A [i] 3, A [i] n);
SB4: will the distance wd (X, A [i]) obtained be asked in described step SB3 by ascending sort, try to achieve the maximum distance maxD=max{d (X, A [i]) of distance wd (X, A [i]) | i=1,2 ... .., k};
SB5: calculation training concentrates the distance of remaining record and sample to be tested X successively, and compared with the maximum distance maxD tried to achieve in described step SB4, if less than maximum distance maxD, then maximum distance maxD is updated to the distance value of this record and sample to be tested X, and again by ascending order adjust the distance wd (X, A [i]) sequence;
SB6: the occurrence number calculating the label of the every bar record in present range wd (X, A [i]) sequence, and sort according to the height of occurrence number;
SB7: using the label of L label before sequence in described step SB6 obtains as sample X.
For verifying validity of the present invention, test as follows:
Test independent operating 10 times, population is 50, and iterations is 100, inertia weight w=1, Studying factors c1=c2=2, for saving time, most adjacent node number K=1.
Four data sets of this test that what table 1 was listed is, they are all data sets that machine learning is commonly used, in order to make to have comparability between data characteristics, carried out standardization to data set, training set example and test set example respectively account for 70% and 30% of total example:
Table 1: data set
Data set Attribute number Class number Training set example Test set example
CAL500 68 174 351 151
Emotions 72 6 391 202
Scene 294 6 1211 1196
Yeast 8 10 1039 445
Experimental result and analysis:
Table 2 gives without the KNN algorithm of weights, WKNN-DIS and the Performance comparision of the many labels based on particle cluster algorithm provided by the invention PSOKNN algorithm on different test set, wherein WKNN-DIS is the KNN method based on distance, characteristic weighing KNN algorithm discusses because different characteristic is different to the influence degree of label, can cause the error of classification.And WKNN-DIS is the algorithm based on Euclidean distance, because the difference of distance is also different to the influence degree of label, in general, larger for sample to be sorted impact apart from nearer training sample, shared weight is also larger.WKNN-DIS and PSOKNN is similar, and the weight of its distance is also optimized by optimized algorithm, selects the particle cluster algorithm the same with PSOKNN in this test.
Table 2: the accuracy rate of algorithm on test set compares
Wherein Average Accuracy is the mean value of PSOKNN (or WKNN-DIS) algorithm 10 experimental results, the accuracy rate of first 10,20,30 refers to and the size of the individuality in last generation according to adaptive value is sorted, then by before rank 10,20,30 individuality correspondence weight combination be applied to respectively in classification, accuracy rate is averaged, after carrying out 10 tests, each experiment is averaged and average again, the origin of front 10,20,30 accuracys rate that Here it is.
Result can find out that WKNN-DIS is generally high in the accuracy rate of each data set than original KNN method by experiment, and PSOKNN method is generally good than WKNN-DIS method.Because PSOKNN method is main is find optimum feature weight to eliminate the redundancy of the feature (referring to property value when calculating distance) of data centralization or uncorrelated as far as possible.Other two kinds of methods all do not have this function, so other two kinds of methods have error when calculating minimum distance, because minimum distance is calculated with reasonably range formula by attribute, so higher to the value dependence of attribute, when redundance is higher, deviation will affect the accuracy of classification.So the method for proposition EVOLUTIONARY COMPUTATION herein draws the feature weight of optimization, reduce range deviation, improve the accuracy of classification, it is also proposed feasible adaptive classifier effectively.
Although give some embodiments of the present invention, it will be understood by those of skill in the art that without departing from the spirit of the invention herein, can change embodiment herein.Above-described embodiment is exemplary, should using embodiment herein as the restriction of interest field of the present invention.

Claims (3)

1. based on many labelings method of particle cluster algorithm, it is characterized in that, comprise optimizing phase and sorting phase:
S10: the optimizing phase is the feature weight adopting particle cluster algorithm to optimize characteristic weighing KNN algorithm, specifically comprises the steps:
S11: adopt random device initialization population, the position of each particle and the dimension of speed are n, the vectorial w=(w of feature weight that the position corresponding data of each particle concentrates record 1, w 2, w n): Qi Zhongyou
Σ i - 1 n w i = 1 ;
S12: calculate adaptive value, and then obtain locally optimal solution and globally optimal solution:
When calculating adaptive value, the position of particle and feature weight are applied in characteristic weighing KNN algorithm, former training sample is concentrated front 70% as new training sample, rear 30% as new forecast sample collection, this forecast sample collection is classified, calculate the accuracy of classification, accuracy is more high more meets adaptive value;
The original tag that forecast sample integrates every bar record is as li=(li1, li2, lin), and the prediction label after sorted is lj=(lj1, lj2, ljn), li and l jbe sum, so accuracy rate Accruay=sum/n with the number that overlaps;
S20: sorting phase:
The feature weight drawn optimizing phase is applied in characteristic weighing KNN algorithm classifies to test sample book X, and the final label exporting all samples in test set, completes classification.
2. a kind of method of many labelings based on particle cluster algorithm as claimed in claim 1, it is characterized in that, described particle cluster algorithm comprises the steps:
SA1: initialization Particle Swarm, comprising the position xi=(x of the whole population of initialization i1, x i2, x id) twith speed v i=(v i1, v i2, v id) tand local optimum and total optimization, wherein id represents d particle in the i-th generation;
SA2: calculate the fitness value fitness of each particle in current position id=f (x id).Then according to the size of fitness value, initialization locally optimal solution pbest i=fitness iwith total optimization solution gbest=min (fintess 1, fitness 2, fitness n), i=1,2, N;
SA3: in each iterative process, each particle upgrades position and the speed of oneself according to following criterion:
v id(t+1)=wv id(t)+c 1r 1(p ld-x id(t))+c 2r 2(p gd-x id(t))
x id(t+1)=x id(t)+v id(t+1)
Wherein v idfor the speed of particle, x idfor the position of particle, w is inertia weight, c1 and c2 is acceleration factor, r 1and r 2random number, P ldfor globally optimal solution and P gdfor locally optimal solution;
SA4: upgrade locally optimal solution pbest iwith total optimization solution gbest;
SA5: if total optimization solution gbest reaches the threshold value of setting or reached maximum iteration time, algorithm can stop calculating; Otherwise jump to step SA3.
3. a kind of method of many labelings based on particle cluster algorithm as claimed in claim 1, it is characterized in that, described characteristic weighing KNN algorithm specifically comprises the steps:
SB1: input m training sample, and set k value size;
SB2: the A [1] in first Stochastic choice training set ~ A [k] sample is as the initial most adjacent node of the k of sample X to be predicted;
SB3: the weighting Euclidean distance wd (X, A [i]) calculating the most adjacent node of sample X to be predicted and each initial k, i=1,2 ... .., k), calculating range formula is:
w d ( X , A [ i ] ) = Σ l = 1 n w l ( X l - A [ i ] l ) 2 ,
Wherein n represents sample A [i] attribute number, i.e. A [i]=(A [i] 1, A [i] 2, A [i] 3, A [i] n);
SB4: will the distance wd (X, A [i]) obtained be asked in described step SB3 by ascending sort, try to achieve the maximum distance maxD=max{d (X, A [i]) of distance wd (X, A [i]) | i=1,2 ... .., k};
SB5: calculation training concentrates the distance of remaining record and sample to be tested X successively, and compared with the maximum distance maxD tried to achieve in described step SB4, if less than maximum distance maxD, maximum distance maxD is updated to the distance value of this record and sample to be tested X, again by ascending order adjust the distance wd (X, A [i]) sequence;
SB6: the occurrence number calculating the label of the every bar record in present range wd (X, A [i]) sequence, and sort according to the height of occurrence number;
SB7: using the label of L label before sequence in described step SB6 obtains as sample X.
CN201510464344.2A 2015-07-31 2015-07-31 Particle swarm algorithm-based multi-label classification method Pending CN104991974A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510464344.2A CN104991974A (en) 2015-07-31 2015-07-31 Particle swarm algorithm-based multi-label classification method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510464344.2A CN104991974A (en) 2015-07-31 2015-07-31 Particle swarm algorithm-based multi-label classification method

Publications (1)

Publication Number Publication Date
CN104991974A true CN104991974A (en) 2015-10-21

Family

ID=54303789

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510464344.2A Pending CN104991974A (en) 2015-07-31 2015-07-31 Particle swarm algorithm-based multi-label classification method

Country Status (1)

Country Link
CN (1) CN104991974A (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105824961A (en) * 2016-03-31 2016-08-03 北京奇艺世纪科技有限公司 Tag determining method and device
CN106959608A (en) * 2017-02-27 2017-07-18 同济大学 A kind of water supply network seepage optimal control method based on cluster particle cluster algorithm
CN108399267A (en) * 2018-03-27 2018-08-14 东北大学 A kind of reaction type clustering method based on cluster analysis of semantic characteristics
CN108445537A (en) * 2018-02-07 2018-08-24 中国地质大学(武汉) Earthquake data before superposition AVO elastic parameter inversion methods based on Spark and system
CN109032889A (en) * 2018-07-11 2018-12-18 广东水利电力职业技术学院(广东省水利电力技工学校) A kind of New cold type server system and management method, computer program
CN109062290A (en) * 2018-07-13 2018-12-21 山东工业职业学院 A kind of reading intelligent agriculture environmental monitoring system and monitoring method based on big data
CN109559797A (en) * 2018-10-31 2019-04-02 青岛大学附属医院 A kind of haemodialysis rehabilitation exercise training method and system, terminal
CN109581987A (en) * 2018-12-29 2019-04-05 广东飞库科技有限公司 A kind of AGV scheduling paths planning method and system based on particle swarm algorithm
CN110493718A (en) * 2019-08-28 2019-11-22 奇点新源国际技术开发(北京)有限公司 A kind of localization method and device
CN110575530A (en) * 2019-09-26 2019-12-17 杜运升 Medicine for promoting sow to enhance maternal performance and preparation method thereof
CN110704624A (en) * 2019-09-30 2020-01-17 武汉大学 Geographic information service metadata text multi-level multi-label classification method
CN114836823A (en) * 2022-06-08 2022-08-02 连城凯克斯科技有限公司 Method for predicting crystal growth diameter of monocrystalline silicon smelting furnace
CN116451099A (en) * 2023-06-19 2023-07-18 浪潮通用软件有限公司 High-entropy KNN clustering method, equipment and medium based on random traversal

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102880872A (en) * 2012-08-28 2013-01-16 中国科学院东北地理与农业生态研究所 Classification and construction method for semi-supervised support vector machine (SVM) remote sensing image
CN103714354A (en) * 2014-01-16 2014-04-09 西安电子科技大学 Hyperspectral image wave band selection method based on quantum-behaved particle swarm optimization algorithm
CN104361393A (en) * 2014-09-06 2015-02-18 华北电力大学 Method for using improved neural network model based on particle swarm optimization for data prediction

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102880872A (en) * 2012-08-28 2013-01-16 中国科学院东北地理与农业生态研究所 Classification and construction method for semi-supervised support vector machine (SVM) remote sensing image
CN103714354A (en) * 2014-01-16 2014-04-09 西安电子科技大学 Hyperspectral image wave band selection method based on quantum-behaved particle swarm optimization algorithm
CN104361393A (en) * 2014-09-06 2015-02-18 华北电力大学 Method for using improved neural network model based on particle swarm optimization for data prediction

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
任江涛等: "基于PSO面向K近邻分类的特征权重学习算法", 《计算机科学》 *

Cited By (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105824961B (en) * 2016-03-31 2019-06-14 北京奇艺世纪科技有限公司 A kind of label determines method and device
CN105824961A (en) * 2016-03-31 2016-08-03 北京奇艺世纪科技有限公司 Tag determining method and device
CN106959608A (en) * 2017-02-27 2017-07-18 同济大学 A kind of water supply network seepage optimal control method based on cluster particle cluster algorithm
CN106959608B (en) * 2017-02-27 2019-10-01 同济大学 A kind of water supply network leakage optimal control method based on cluster particle swarm algorithm
CN108445537A (en) * 2018-02-07 2018-08-24 中国地质大学(武汉) Earthquake data before superposition AVO elastic parameter inversion methods based on Spark and system
CN108445537B (en) * 2018-02-07 2019-05-31 中国地质大学(武汉) Earthquake data before superposition AVO elastic parameter inversion method and system based on Spark
CN108399267A (en) * 2018-03-27 2018-08-14 东北大学 A kind of reaction type clustering method based on cluster analysis of semantic characteristics
CN108399267B (en) * 2018-03-27 2020-04-14 东北大学 Feedback clustering method based on cluster semantic feature analysis
CN109032889A (en) * 2018-07-11 2018-12-18 广东水利电力职业技术学院(广东省水利电力技工学校) A kind of New cold type server system and management method, computer program
CN109062290A (en) * 2018-07-13 2018-12-21 山东工业职业学院 A kind of reading intelligent agriculture environmental monitoring system and monitoring method based on big data
CN109559797A (en) * 2018-10-31 2019-04-02 青岛大学附属医院 A kind of haemodialysis rehabilitation exercise training method and system, terminal
CN109581987A (en) * 2018-12-29 2019-04-05 广东飞库科技有限公司 A kind of AGV scheduling paths planning method and system based on particle swarm algorithm
CN110493718A (en) * 2019-08-28 2019-11-22 奇点新源国际技术开发(北京)有限公司 A kind of localization method and device
CN110493718B (en) * 2019-08-28 2021-02-26 奇点新源国际技术开发(北京)有限公司 Positioning method and device
CN110575530A (en) * 2019-09-26 2019-12-17 杜运升 Medicine for promoting sow to enhance maternal performance and preparation method thereof
CN110704624A (en) * 2019-09-30 2020-01-17 武汉大学 Geographic information service metadata text multi-level multi-label classification method
CN110704624B (en) * 2019-09-30 2021-08-10 武汉大学 Geographic information service metadata text multi-level multi-label classification method
CN114836823A (en) * 2022-06-08 2022-08-02 连城凯克斯科技有限公司 Method for predicting crystal growth diameter of monocrystalline silicon smelting furnace
CN114836823B (en) * 2022-06-08 2024-03-19 连城凯克斯科技有限公司 Crystal growth diameter prediction method of monocrystalline silicon melting furnace
CN116451099A (en) * 2023-06-19 2023-07-18 浪潮通用软件有限公司 High-entropy KNN clustering method, equipment and medium based on random traversal
CN116451099B (en) * 2023-06-19 2023-09-01 浪潮通用软件有限公司 High-entropy KNN clustering method, equipment and medium based on random traversal

Similar Documents

Publication Publication Date Title
CN104991974A (en) Particle swarm algorithm-based multi-label classification method
CN107256245B (en) Offline model improvement and selection method for spam message classification
CN106845717B (en) Energy efficiency evaluation method based on multi-model fusion strategy
CN106971091B (en) Tumor identification method based on deterministic particle swarm optimization and support vector machine
Ghanem et al. Multi-class pattern classification in imbalanced data
CN103425996B (en) A kind of large-scale image recognition methods of parallel distributed
CN104766098A (en) Construction method for classifier
CN113378913B (en) Semi-supervised node classification method based on self-supervised learning
CN105389583A (en) Image classifier generation method, and image classification method and device
CN110942091A (en) Semi-supervised few-sample image classification method for searching reliable abnormal data center
CN113887643B (en) New dialogue intention recognition method based on pseudo tag self-training and source domain retraining
CN113554100B (en) Web service classification method for enhancing attention network of special composition picture
CN109522544A (en) Sentence vector calculation, file classification method and system based on Chi-square Test
CN110738232A (en) grid voltage out-of-limit cause diagnosis method based on data mining technology
CN111695011B (en) Tensor expression-based dynamic hypergraph structure learning classification method and system
CN105512675A (en) Memory multi-point crossover gravitational search-based feature selection method
CN109670687A (en) A kind of mass analysis method based on particle group optimizing support vector machines
CN111652478A (en) Electric power system voltage stability evaluation misclassification constraint method based on umbrella algorithm
CN105069485A (en) Extreme-learning-machine-based mode identification method in tensor mode
CN102929977B (en) Event tracing method aiming at news website
Guo et al. Reducing evaluation cost for circuit synthesis using active learning
CN117076871B (en) Battery fault classification method based on unbalanced semi-supervised countermeasure training framework
Jingbiao et al. Research and improvement of clustering algorithm in data mining
CN104881688A (en) Two-stage clustering algorithm based on difference evolution and fuzzy C-means
Liu et al. A weight-incorporated similarity-based clustering ensemble method

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20151021

RJ01 Rejection of invention patent application after publication