CN111931829B - Classifier screening method, system, storage medium and computer equipment - Google Patents

Classifier screening method, system, storage medium and computer equipment Download PDF

Info

Publication number
CN111931829B
CN111931829B CN202010722712.XA CN202010722712A CN111931829B CN 111931829 B CN111931829 B CN 111931829B CN 202010722712 A CN202010722712 A CN 202010722712A CN 111931829 B CN111931829 B CN 111931829B
Authority
CN
China
Prior art keywords
classifier
auc
data set
dynamic programming
expression
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010722712.XA
Other languages
Chinese (zh)
Other versions
CN111931829A (en
Inventor
陈泽鹏
徐维超
陈昌润
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong University of Technology
Original Assignee
Guangdong University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong University of Technology filed Critical Guangdong University of Technology
Priority to CN202010722712.XA priority Critical patent/CN111931829B/en
Publication of CN111931829A publication Critical patent/CN111931829A/en
Application granted granted Critical
Publication of CN111931829B publication Critical patent/CN111931829B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/285Selection of pattern recognition techniques, e.g. of classifiers in a multi-classifier system
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/10Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The utility model provides a classifier screening method, a classifier screening system, a storage medium and computer equipment, wherein the method comprises the following steps: preprocessing a data set, and converting the text of the data set into a vectorization matrix; acquiring a variance expression of the AUC of the classifier under the corresponding classification problem according to the vectorization matrix; reconstructing the variance expression by using a dynamic programming term to obtain an AUC expression based on dynamic programming; and acquiring an AUC sample estimated value of the classifier to be selected according to the AUC expression based on dynamic programming, and taking the classifier to be selected with the maximum AUC sample estimated value as the optimal classifier under the corresponding classification problem.

Description

Classifier screening method, system, storage medium and computer equipment
Technical Field
The present utility model relates to the field of machine learning technologies, and in particular, to a classifier screening method, a classifier screening system, a storage medium, and a computer device.
Background
Machine learning is generally divided into two types: supervised learning and unsupervised learning. Classification is a supervised learning process, which categories are known in the target database, and all that is required to do is to assign each record to a corresponding category, and the output of the classification problem is no longer a continuous value, but a discrete value, which is used to specify which category it belongs to. For probability classifiers, the corresponding discrete (class II) classifier can be obtained by setting a threshold value: when the output of the classifier exceeds the threshold, a result of 1 is output, otherwise 0. Each threshold may produce a different point in the plane of the receiver operating characteristic (Receiver Operating Characteristic, hereinafter ROC). Conceptually, we can generate a curve by plotting the corresponding points of each threshold in ROC space, which is a ROC curve essentially representing the trade-off relationship between true and false positive rates at different decision thresholds. The area under the curve (Area Under the Curve, AUC) can be used to evaluate whether a classifier can effectively distinguish positive and negative samples in a particular problem. The classifier is evaluated through an AUC algorithm, so that a more suitable classifier can be screened.
As shown in AUC-based classifier performance evaluation problem study (Jilin university, jiang Shuai), the existing method has the advantages of narrow applicability, poor efficiency, complex treatment process, deviation of results and obvious limitation.
Disclosure of Invention
Aiming at the limitation of the prior art, the utility model provides a classifier screening method, a classifier screening system, a storage medium and computer equipment, and the technical scheme adopted by the utility model is as follows:
a classifier screening method comprising the steps of:
preprocessing a data set, and converting the text of the data set into a vectorization matrix;
acquiring a variance expression of the AUC of the classifier under the corresponding classification problem according to the vectorization matrix;
reconstructing the variance expression by using a dynamic programming term to obtain an AUC expression based on dynamic programming;
and acquiring an AUC sample estimated value of the classifier to be selected according to the AUC expression based on dynamic programming, and taking the classifier to be selected with the maximum AUC sample estimated value as the optimal classifier under the corresponding classification problem.
Compared with the prior art, the method can be used as an unbiased estimation of the MWUS variance, the time complexity is linear and far lower than that of the conventional method, and is equivalent to the most advanced rank-based method; meanwhile, the processing speed is far higher than that of other similar methods, and the most suitable classifier can be quickly selected. In addition to these advantages, the structure of the algorithm can be easily expanded to the three-classification or multi-classification condition, and the application range of the method is wide. In addition, the method provided by the utility model can also be used for improving the current cell detection method.
As a preferred solution, under the classification problem, the AUC expression based on dynamic programming is as follows:
wherein ,representing an AUC sample estimate; m represents the positive sample number of the data set, and n represents the negative sample number of the data set; x is X i A positive sample sequence representing the dataset, Y j A negative sample sequence representing the dataset; />Indicating an exponential function, indicating that the function takes a value of 1 when the statement in the bracket is true, and indicating that the function takes a value of 0 when the statement in the bracket is false; dynamic programming item S 1 The satisfied range is epsilon (x=y); dynamic programming item S 2 The satisfied range is epsilon (X>Y)。
As a preferred solution, under the classification problem, the dynamic programming term S 1 ,S 2 ,....S 9 ,S 10 The corresponding satisfaction ranges are as follows:
S 1 ,ε(X=Y);S 2 ,ε(X>Y);S 3 ,ε(X>Y>Y′)orε(X>Y>Y);
S 4 ,ε(X>Y=Y′);S 5 ,ε(X=Y>Y)orε(X=Y>Y′);S 6 ,ε(X=Y=Y′);
S 7 ,ε(X=X>Y)orε(X′=X>Y);S 8 ,ε(X=X>Y);
S 9 ,ε(X=X>Y)orε(X>X=Y);S 10 ,ε(X=X=Y);
wherein X and Y respectively represent independent identical distribution sample sets selected from two parent distributions, and X 'and Y' respectively correspond to the independent identical distribution sample sets of which X and Y have the same element.
Further, the expression of the dynamic programming term is as follows:
m represents the positive sample number of the data set, and n represents the negative sample number of the data set; x is X i A positive sample sequence representing the dataset, Y j A negative sample sequence representing the dataset, Z k X represents i And Y is equal to j Is arranged according to a non-descending order, K=m+n, K is more than or equal to 1 and less than or equal to K; a, a i X represents i At Z k Is a sequence of (a); c i Represents Y j At Z k Is a sequence of (a) and (b).
Further, under the classification problem, according to the vectorization matrix, the obtained variance expression of the classifier AUC is as follows:
wherein ,
representing the expectation of a random variable, m representing the number of positive samples of the data set, n representing the number of negative samples of the data set; />Representing an AUC sample estimate; />Representing an exponential function
Further, under a classification problem, after the variance expression is reconstructed by using a dynamic programming term, the variance expression is as follows:
wherein ,
further, preprocessing the data set to convert the text of the data set into a vectorization matrix, including the following steps:
segmenting the text of the data set into a vocabulary, and converting the text of the data set into a vectorization matrix according to the vocabulary;
and automatically filling null values in the vectorization matrix, and carrying out Laplace smoothing on the vectorization matrix.
A classifier screening system comprising:
the preprocessing module is used for preprocessing the data set and converting the data of the data set into a vectorization matrix;
the variance expression acquisition module acquires a variance expression of the classifier AUC under the corresponding classification problem according to the vectorization matrix;
the AUC expression acquisition module based on dynamic programming acquires the AUC expression based on dynamic programming under the two classification problems by reconstructing the variance expression by using a dynamic programming term;
the classification module is used for acquiring an AUC sample estimated value of the classifier to be selected according to the AUC expression based on dynamic programming under the classification problem, and taking the classifier to be selected with the maximum AUC sample estimated value as the optimal classifier under the corresponding classification problem
The utility model also provides the following:
a storage medium having stored thereon a computer program which, when executed by a processor, performs the steps of a classifier screening method as described above.
A computer device comprising a storage medium, a processor and a computer program stored in the storage medium and executable by the processor, which when executed by the processor performs the steps of a classifier screening method as described above.
Drawings
Fig. 1 is a flowchart of a classifier screening method provided in embodiment 1 of the present utility model;
fig. 2 is a flowchart of a classifier screening method provided in embodiment 2 of the present utility model;
fig. 3 is a schematic diagram of a classifier screening system according to an embodiment of the present utility model.
Detailed Description
The drawings are for illustrative purposes only and are not to be construed as limiting the present patent;
it should be understood that the described embodiments are merely some, but not all embodiments of the present utility model. All other embodiments, which can be made by one of ordinary skill in the art without undue burden from the utility model, are intended to be within the scope of the embodiments of the present utility model.
The terminology used in the embodiments of the utility model is for the purpose of describing particular embodiments only and is not intended to be limiting of embodiments of the utility model. As used in this application and the appended claims, the singular forms "a," "an," and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It should also be understood that the term "and/or" as used herein refers to and encompasses any or all possible combinations of one or more of the associated listed items.
When the following description refers to the accompanying drawings, the same numbers in different drawings refer to the same or similar elements, unless otherwise indicated. The implementations described in the following exemplary examples do not represent all implementations consistent with the utility model. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the utility model as detailed in the accompanying claims. In the description of the present utility model, it should be understood that the terms "first," "second," "third," and the like are used merely to distinguish between similar objects and are not necessarily used to describe a particular order or sequence, nor should they be construed to indicate or imply relative importance. The specific meaning of the above terms in the present utility model can be understood by those of ordinary skill in the art according to the specific circumstances.
Furthermore, in the description of the present utility model, unless otherwise indicated, "a plurality" means two or more. "and/or", describes an association relationship of an association object, and indicates that there may be three relationships, for example, a and/or B, and may indicate: a exists alone, A and B exist together, and B exists alone. The character "/" generally indicates that the context-dependent object is an "or" relationship. The utility model is further illustrated in the following figures and examples.
In order to solve the limitations of the prior art, the present embodiment provides a technical solution, and the technical solution of the present utility model is further described below with reference to the drawings and the embodiments.
Referring to fig. 1, a classifier screening method includes the following steps:
s01, preprocessing a data set, and converting the text of the data set into a vectorization matrix;
s02, obtaining a variance expression of the AUC of the classifier under the corresponding classification problem according to the vectorization matrix;
s03, reconstructing the variance expression by using a dynamic programming term to obtain an AUC expression based on dynamic programming;
s04, obtaining an AUC sample estimated value of the classifier to be selected according to the AUC expression based on dynamic programming, and taking the classifier to be selected with the maximum AUC sample estimated value as the optimal classifier under the corresponding classification problem.
Compared with the prior art, the method can be used as an unbiased estimation of the MWUS variance, the time complexity is linear and far lower than that of the conventional method, and is equivalent to the most advanced rank-based method; meanwhile, the processing speed is far higher than that of other similar methods, and the most suitable classifier can be quickly selected. In addition to these advantages, the structure of the algorithm can be easily expanded to the three-classification or multi-classification condition, and the application range of the method is wide. In addition, the method provided by the utility model can also be used for improving the current cell detection method.
In particular, in an alternative embodiment, the text of the dataset may take the form of news text.
Classification problems are very widely used in reality, such as spam recognition, handwritten number recognition, face recognition, speech recognition, etc. Wherein, a classified problem, such as a microblog message data set, contains a plurality of words and numbers, and in order to shield the profoundly speaking, a rapid classification can be constructed: if a message is in a negative or insulting language, then the message is marked as being miscontained; for the problem of three classifications, such as in heart detection, three alternative diagnostic results can be obtained by measuring the heart beat frequency signal: normal heart rate, too fast heart beat, too slow heart beat; one more variable is included in the expression in the analysis process, which is equivalent to three samples; the AUC plane in the previous two classification problem also becomes curved. For the three-classification or multi-classification problem, the variance expression can be reconstructed by using a dynamic programming term to obtain a consistent and predictable result based on the principle of the patent, so that the description is omitted.
As a preferred solution, under the classification problem, the AUC expression based on dynamic programming is as follows:
wherein ,representing an AUC sample estimate; m represents the positive sample number of the data set, and n represents the negative sample number of the data set; x is X i A positive sample sequence representing the dataset, Y j A negative sample sequence representing the dataset; />Indicating an exponential function, indicating that the function takes a value of 1 when the statement in the bracket is true, and indicating that the function takes a value of 0 when the statement in the bracket is false; dynamic programming item S 1 The satisfied range is epsilon (x=y); dynamic programming item S 2 The satisfied range is epsilon (X>Y)。
As a preferred solution, under the classification problem, the dynamic programming term S 1 ,S 2 ,....S 9 ,S 10 The corresponding satisfaction ranges are as follows:
S 1 ,ε(X=Y);S 2 ,ε(X>Y);S 3 ,ε(X>Y>Y′)orε(X>Y>Y);
S 4 ,ε(X>Y=Y′);S 5 ,ε(X=Y>Y)orε(X=Y>Y′);S 6 ,ε(X=Y=Y′);
S 7 ,ε(X=X>Y)orε(X=X>Y);S 8 ,ε(X=X>Y);
S 9 ,ε(X=X>Y)orε(X>X=Y);S 10 ,ε(X=X=Y);
wherein X and Y respectively represent independent identical distribution sample sets selected from two parent distributions, and X 'and Y' respectively correspond to the independent identical distribution sample sets of which X and Y have the same element.
In the present embodiment, Z is k X represents i And Y is equal to j Is arranged in a non-decreasing order, K=m+n, 1.ltoreq.k.ltoreq.K,
two technical vectors are readily available:
the dynamic programming term can thus be obtained as follows:
i and j each represent two independent identical sample numbers from i=1, …, n; j=1, …, m, l represents X i and Yj A value sequence in the sequencing of the samples; c l 、t l Respectively representing the positions and the number of the data in the data sequence of two independent and respectively sampled samples;m represents the positive sample number of the data set, and n represents the negative sample number of the data set; x is X i A positive sample sequence representing the dataset, Y j A negative sample sequence representing the dataset, Z k X represents i And Y is equal to j Is arranged according to a non-descending order, K=m+n, K is more than or equal to 1 and less than or equal to K; a, a i X represents i At Z k Is a sequence of (a); c i Represents Y j At Z k Is a sequence of (a) and (b).
Further, under the classification problem, according to the vectorization matrix, the obtained variance expression of the classifier AUC is as follows:
wherein ,
representing the expectation of a random variable, m representing the number of positive samples of the data set, n representing the number of negative samples of the data set; />Representing an AUC sample estimate; />Representing an exponential function
Specifically, for the above variance expression, it can be converted into a compact form of variance:
wherein ,
the above expression, although unbiased, is and />Is a cubic order, which is very inefficient for large m and n; after the dynamic programming term is introduced, the variance expression is further converted into:
wherein ,
in particular, the method comprises the steps of,indicating an exponential function, indicating that the function takes a value of 1 when the statement in the bracket is true, and indicating that the function takes a value of 0 when the statement in the bracket is false; n is n 2 For the dynamic programming-based AUC expression, there is:
the method for obtaining the AUC based on the AUC expression of dynamic programming is unbiased.
Further, preprocessing the data set to convert the text of the data set into a vectorization matrix, including the following steps:
s011, segmenting the text of the data set into a vocabulary, and converting the text of the data set into a vectorization matrix according to the vocabulary;
s012, automatically filling null values in the vectorization matrix, and carrying out Laplacian smoothing on the vectorization matrix.
Specifically, for English text, split function can be used to split with non-letters and non-numbers as symbols; for Chinese sentences, the jieba function is used for segmentation. Laplacian smoothing of the processed vocabulary is to solve the 0 probability problem, and avoid errors caused by underflow or floating point rounding in a logarithmic form. Meanwhile, the natural logarithm is adopted for processing without any loss.
A classifier screening system comprising:
the preprocessing module 1 is used for preprocessing the data set and converting the data of the data set into a vectorization matrix;
the variance expression acquisition module 2 acquires a variance expression of the classifier AUC under the corresponding classification problem according to the vectorization matrix;
the AUC expression acquisition module 3 based on dynamic programming acquires the AUC expression based on dynamic programming under the two classification problems by reconstructing the variance expression by using a dynamic programming term;
and the classification module 4 is used for acquiring an AUC sample estimated value of the classifier to be selected according to the AUC expression based on dynamic programming under the classification problem, and taking the classifier to be selected with the maximum AUC sample estimated value as the optimal classifier under the corresponding classification problem.
The utility model also provides the following:
a storage medium having stored thereon a computer program which, when executed by a processor, performs the steps of a classifier screening method as described above.
A computer device comprising a storage medium, a processor and a computer program stored in the storage medium and executable by the processor, which when executed by the processor performs the steps of a classifier screening method as described above.
It is to be understood that the above examples of the present utility model are provided by way of illustration only and not by way of limitation of the embodiments of the present utility model. Other variations or modifications of the above teachings will be apparent to those of ordinary skill in the art. It is not necessary here nor is it exhaustive of all embodiments. Any modification, equivalent replacement, improvement, etc. which come within the spirit and principles of the utility model are desired to be protected by the following claims.

Claims (8)

1. A classifier screening method, comprising the steps of:
preprocessing a data set, and converting the text of the data set into a vectorization matrix;
acquiring a variance expression of the AUC of the classifier under the corresponding classification problem according to the vectorization matrix;
reconstructing the variance expression by using a dynamic programming term to obtain an AUC expression based on dynamic programming;
acquiring an AUC sample estimated value of the classifier to be selected according to the AUC expression based on dynamic programming, and taking the classifier to be selected with the maximum AUC sample estimated value as the optimal classifier under the corresponding classification problem;
under the two classification problems, the dynamic programming item S 1 ,S 2 ,....S 9 ,S 10 The corresponding satisfaction ranges are as follows:
S 1 ,ε(X=Y);S 2 ,ε(X>Y);S 3 ,ε(X>Y>Y′)orε(X>Y′>Y);
S 4 ,ε(X>Y=Y′);S 5 ,ε(X=Y′>Y)orε(X=Y>Y′);S 6 ,ε(X=Y=Y′);
S 7 ,ε(X=X′>Y)orε(X′=X>Y);S 8 ,ε(X=X′>Y);
S 9 ,ε(X′=X>Y)orε(X>X′=Y);S 10 ,ε(X=X′=Y);
wherein X and Y respectively represent independent identical distribution sample sets selected from two parent distributions, and X 'and Y' respectively correspond to the independent identical distribution sample sets of which X and Y have the same element;
the expression of the dynamic programming term is as follows:
i and j each represent two independent identical sample numbers from i=1, …, n; j=1, …, m, l represents X i and Yj A value sequence in the sequencing of the samples; c l 、t l Respectively representing the positions and the number of the data in the data sequence of two independent and respectively sampled samples;indicating an exponential function, indicating that the function takes a value of 1 when the statement in the bracket is true, and indicating that the function takes a value of 0 when the statement in the bracket is false; m represents the positive sample number of the data set, and n represents the negative sample number of the data set; x is X i A positive sample sequence representing the dataset, Y j A negative sample sequence representing the dataset, Z k X represents i And Y is equal to j Is arranged according to a non-descending order, K=m+n, K is more than or equal to 1 and less than or equal to K; a, a i X represents i At Z k Is a sequence of (a); c i Represents Y j At Z k Is a sequence of (a) and (b).
2. The classifier screening method according to claim 1, wherein under a classification problem, the AUC expression based on dynamic programming is as follows:
wherein ,representing an AUC sample estimate; m represents the positive sample number of the data set, and n represents the negative sample number of the data set; x is X i A positive sample sequence representing the dataset, Y j A negative sample sequence representing the dataset; />Indicating an exponential function, indicating that the function takes a value of 1 when the statement in the bracket is true, and indicating that the function takes a value of 0 when the statement in the bracket is false; dynamic programming item S 1 The satisfied range is epsilon (x=y); dynamic programming item S 2 The satisfied range is epsilon (X>Y)。
3. The classifier screening method according to claim 1, wherein under the classification problem, according to the vectorization matrix, the obtained variance expression of the classifier AUC is as follows:
wherein ,
representing the expectation of a random variable, m representing the number of positive samples of the data set, n representing the number of negative samples of the data set; />Representing an AUC sample estimate; />Representing an exponential function
4. The classifier screening method according to claim 1, wherein under a classification problem, after the variance expression is reconstructed by using a dynamic programming term, the variance expression is as follows:
wherein ,
5. the classifier screening method of claim 1, wherein preprocessing the data set to convert text of the data set into a vectorization matrix comprises the steps of:
segmenting the text of the data set into a vocabulary, and converting the text of the data set into a vectorization matrix according to the vocabulary;
and automatically filling null values in the vectorization matrix, and carrying out Laplace smoothing on the vectorization matrix.
6. A classifier screening system, characterized by being applied to the classifier screening method of claim 1, comprising:
the preprocessing module is used for preprocessing the data set and converting the data of the data set into a vectorization matrix;
the variance expression acquisition module acquires a variance expression of the classifier AUC under the corresponding classification problem according to the vectorization matrix;
the AUC expression acquisition module based on dynamic programming acquires the AUC expression based on dynamic programming under the two classification problems by reconstructing the variance expression by using a dynamic programming term;
and the classification module is used for acquiring an AUC sample estimated value of the classifier to be selected according to the AUC expression based on dynamic programming under the classification problem, and taking the classifier to be selected with the maximum AUC sample estimated value as the optimal classifier under the corresponding classification problem.
7. A storage medium having a computer program stored thereon, characterized by: the computer program, when executed by a processor, implements the steps of the classifier screening method of any one of claims 1 to 5.
8. A computer device, characterized by: comprising a storage medium, a processor and a computer program stored in the storage medium and executable by the processor, which computer program, when executed by the processor, implements the steps of the classifier screening method of any one of claims 1 to 5.
CN202010722712.XA 2020-07-24 2020-07-24 Classifier screening method, system, storage medium and computer equipment Active CN111931829B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010722712.XA CN111931829B (en) 2020-07-24 2020-07-24 Classifier screening method, system, storage medium and computer equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010722712.XA CN111931829B (en) 2020-07-24 2020-07-24 Classifier screening method, system, storage medium and computer equipment

Publications (2)

Publication Number Publication Date
CN111931829A CN111931829A (en) 2020-11-13
CN111931829B true CN111931829B (en) 2023-09-01

Family

ID=73315453

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010722712.XA Active CN111931829B (en) 2020-07-24 2020-07-24 Classifier screening method, system, storage medium and computer equipment

Country Status (1)

Country Link
CN (1) CN111931829B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114330468A (en) * 2021-07-14 2022-04-12 广东工业大学 Classifier screening method and system based on dynamic programming and computer equipment

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102129568A (en) * 2011-04-29 2011-07-20 南京邮电大学 Method for detecting image-based spam email by utilizing improved gauss hybrid model classifier
CN102631194A (en) * 2012-04-13 2012-08-15 西南大学 Taboo searching method used for electrocardial characteristic selection
CN102955946A (en) * 2011-08-18 2013-03-06 刘军 Two-stage fast classifier based on linear classification tree and neural network
CN106528771A (en) * 2016-11-07 2017-03-22 中山大学 Fast structural SVM text classification optimization algorithm
CN106845156A (en) * 2017-01-11 2017-06-13 张渠 Sorting technique, apparatus and system based on blood platelet difference expression gene mark
CN110399490A (en) * 2019-07-17 2019-11-01 武汉斗鱼网络科技有限公司 A kind of barrage file classification method, device, equipment and storage medium

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102129568A (en) * 2011-04-29 2011-07-20 南京邮电大学 Method for detecting image-based spam email by utilizing improved gauss hybrid model classifier
CN102955946A (en) * 2011-08-18 2013-03-06 刘军 Two-stage fast classifier based on linear classification tree and neural network
CN102631194A (en) * 2012-04-13 2012-08-15 西南大学 Taboo searching method used for electrocardial characteristic selection
CN106528771A (en) * 2016-11-07 2017-03-22 中山大学 Fast structural SVM text classification optimization algorithm
CN106845156A (en) * 2017-01-11 2017-06-13 张渠 Sorting technique, apparatus and system based on blood platelet difference expression gene mark
CN110399490A (en) * 2019-07-17 2019-11-01 武汉斗鱼网络科技有限公司 A kind of barrage file classification method, device, equipment and storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于关节信息和极限学习机的人体动作识别;张孙培;孙怀江;;现代电子技术(第10期);第1-4页 *

Also Published As

Publication number Publication date
CN111931829A (en) 2020-11-13

Similar Documents

Publication Publication Date Title
CN107256245B (en) Offline model improvement and selection method for spam message classification
CN111126386B (en) Sequence domain adaptation method based on countermeasure learning in scene text recognition
CN109065028B (en) Speaker clustering method, speaker clustering device, computer equipment and storage medium
CN110633725B (en) Method and device for training classification model and classification method and device
CN108694346B (en) Ship radiation noise signal identification method based on two-stage CNN
JP2003526142A (en) Unsupervised adaptation and classification of multi-source data using generalized Gaussian mixture model
JP2005202932A (en) Method of classifying data into a plurality of classes
CN113850281B (en) MEANSHIFT optimization-based data processing method and device
CN107729520B (en) File classification method and device, computer equipment and computer readable medium
JP2012042990A (en) Image identification information adding program and image identification information adding apparatus
CN111046183A (en) Method and device for constructing neural network model for text classification
CN112437053B (en) Intrusion detection method and device
CN108681532B (en) Sentiment analysis method for Chinese microblog
CN111125469A (en) User clustering method and device for social network and computer equipment
CN111079427A (en) Junk mail identification method and system
CN111931829B (en) Classifier screening method, system, storage medium and computer equipment
CN104077598A (en) Emotion recognition method based on speech fuzzy clustering
Liu et al. A method of plant classification based on wavelet transforms and support vector machines
CN111401440B (en) Target classification recognition method and device, computer equipment and storage medium
CN116467141A (en) Log recognition model training, log clustering method, related system and equipment
CN107563287B (en) Face recognition method and device
WO2021017736A1 (en) Image analysis apparatus
CN112906804A (en) Hash sample balance cancer labeling method for histopathology image
CN114463574A (en) Scene classification method and device for remote sensing image
CN107341485B (en) Face recognition method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant