CN107391433A - A kind of feature selection approach based on composite character KDE conditional entropies - Google Patents
A kind of feature selection approach based on composite character KDE conditional entropies Download PDFInfo
- Publication number
- CN107391433A CN107391433A CN201710526050.7A CN201710526050A CN107391433A CN 107391433 A CN107391433 A CN 107391433A CN 201710526050 A CN201710526050 A CN 201710526050A CN 107391433 A CN107391433 A CN 107391433A
- Authority
- CN
- China
- Prior art keywords
- mrow
- msub
- mfrac
- kde
- mtr
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Pure & Applied Mathematics (AREA)
- Mathematical Optimization (AREA)
- Algebra (AREA)
- Computational Mathematics (AREA)
- Databases & Information Systems (AREA)
- Software Systems (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Analysis (AREA)
- Character Discrimination (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The present invention provides a kind of new feature selection approach based on composite character KDE conditional entropies, it this method propose composite character KDE probability and composite character KDE entropys, discrete features and continuous feature are effectively unified in KDE entropys in the case of discrete not by continuous data, information theory has been expanded and has proposed the greedy feature selection approach based on composite character KDE conditional entropies.
Description
Technical field
The present invention relates to feature selection approach, in particular to a kind of feature selecting side based on composite character KDE conditional entropies
Method.
Background technology
With the raising of the storage capacity and computing capability of data, size of data and data dimension are increasing, give
Data mining or machine learning task bring bigger pressure.Feature selecting is as data mining, pattern-recognition, machine learning
The important pre-treatment step of task, from substantial amounts of attribute, redundancy, unrelated attribute are eliminated, data dimension is reduced, improves
The efficiency of algorithm.
The concepts such as entropy and mutual information in information theory occupy consequence in feature selecting, possess and know without priori
The advantages that knowing detection non-linear relation, anti-noise jamming.But based on the feature selection approach of information theory primarily directed to from
Attribute is dissipated, for continuous feature, takes the mode of discretization mostly to adapt to traditional feature selection approach.Density Estimator
(KDE) it is that a kind of probability density function to stochastic variable carries out the method without ginseng estimation.By the entropy knot in KDE and information theory
Close, being currently based on the feature selection approach of KDE entropys has preferable effect, but existing method is just for continuous feature.
Against this problem, the present invention has expanded information theory so that can be applied to composite character based on KDE entropys.
The content of the invention
The invention aims to handle the feature selecting of composite character, and propose a kind of new based on composite character
The feature selection approach of KDE conditional entropies.Composite character KDE entropys are this method propose, information theory has been expanded and has proposed and be based on
The greedy feature selection approach of composite character KDE conditional entropies.
It is of the invention to be using technical scheme:
A kind of feature selection approach based on composite character KDE conditional entropies, comprises the following steps:
Step 1, input include decision-making feature D data set U, wherein, data set U has n sample, decision-making feature D=
{ 1,2 ..., N }, discrete features vector Α={ A1,A2,...,Am, continuous characteristic vector X={ X1,X2,...,Xt, window width ginseng
Number h, outage threshold T;
Step 2, if the feature set selected is B, non-selected feature set is E, and initial value is set toE=A ∪ X,
The often difference of the conditional entropy before and after one feature of selection
Step 3, temporal aspect collection B ' is established by all properties in each the attribute S and feature set B in feature set E;
Step 4, for each value x of continuous feature set X ' in B ', and each value d in decision kind set D
And the middle discrete features collection A ' of B ' each value a, calculate KDE probability With
Step 5, composite character KDE conditional entropies are based on by obtaining KDE probability calculations in step 4WithAnd based on composite character KDE combination entropiesWherein rememberFor Category Attributes collection A ' codomain,For
Connection attribute collection X ' codomain,For decision set D codomain;
Step 6, the minimum attribute of alternative condition entropyIt is added in feature set B, obtains
Attribute B=B ∪ { S* } have been selected, and E=E- { S* } is deleted from non-selected feature set;
Step 7, the difference of the conditional entropy added before and after new attribute is obtained by B=B ∪ { S* } in step 6, i.e.,
Step 8, the difference of the conditional entropy of judgment step sevenThe Characteristic Number whether being more than in threshold value T and feature set B
Less than the characteristic of data lump, i.e.,If meeting condition, return to step three;Otherwise export
Feature set B.
KDE probability in the step 4Generated by formula (1):
KDE probability in the step 4Generated by formula (2):
KDE probability in the step 4Generated by formula (3):
KDE probability in the step 4Generated by formula (4):
The step 5 conditional entropyGenerated by formula (5):
The step 5 conditional entropyGenerated by formula (6):
The step 5 conditional entropyGenerated by formula (7):
Wherein, due to the sample probability of continuous random variable,
The beneficial effects of the invention are as follows:
1st, the present invention proposes composite character KDE probability and composite character KDE entropys, not by the discrete situation of continuous data
It is lower that discrete features and continuous feature are effectively unified in KDE entropys.
2nd, the present invention is used based on standard of the composite character KDE conditional entropies as evaluating characteristic, is carried out using greedy algorithm
Feature selecting.
3rd, the method that the present invention passes through Experimental comparison's discretization continuous data, it was demonstrated that this algorithm has in various classification experiments
More preferable effect.
Brief description of the drawings
Fig. 1 is flow chart of the method for the present invention;
Fig. 2 is the implementing procedure figure of the present invention.
Embodiment
The present invention is further analyzed with reference to specific embodiment.
The method flow of the present invention is shown in Fig. 1, and based on the definition of above-mentioned composite character KDE entropys, the present invention is based on composite character
The feature selection approach of KDE conditional entropies is described in detail below:
Step 1 101, input include decision-making feature D data set U;Wherein, data set U has n sample, decision-making feature D
={ 1,2 ..., N }, discrete features vector Α={ A1,A2,...,Am, continuous characteristic vector X={ X1,X2,...,Xt, window width
Parameter h, outage threshold T;
Step 2 102, if the feature set selected is B, non-selected feature set is E, and initial value is set toE=A
∪ X, the often conditional entropy before and after one feature of selection difference
Step 3 103, temporal aspect collection is established by all properties in each the attribute S and feature set B in feature set E
B ', perform following steps;
Step 4 104, for each value x of continuous feature set X ' in B ', and it is every in decision kind set D
A kind of middle discrete features collection A ' of value d and B ' each value a, calculate KDE probability With
The KDE probabilityGenerated by formula (1):
The KDE probabilityGenerated by formula (2):
The KDE probabilityGenerated by formula (3):
The KDE probabilityGenerated by formula (4):
Step 5 105, composite character KDE conditional entropies are based on by obtaining KDE probability calculations in step 4WithAnd based on composite character KDE combination entropiesWherein rememberFor Category Attributes collection A ' codomain,
For connection attribute collection X ' codomain,For decision set D codomain;
The step 5 conditional entropyGenerated by formula (5):
The step 5 conditional entropyGenerated by formula (6):
The step 5 conditional entropyGenerated by formula (7):
Wherein, due to the sample probability of continuous random variable,
Step 6 106, the minimum attribute of alternative condition entropyIt is added in feature set B,
Attribute B=B ∪ { S* } have been selected in acquisition, and E=E- { S* } is deleted from non-selected feature set;
Step 7 107, the difference of the conditional entropy added before and after new attribute is obtained by B=B ∪ { S* } in step 6, i.e.,
Step 8 108, the conditional entropy of judgment step sevenThe Characteristic Number whether being more than in threshold value T and feature set B is small
In the characteristic of data lump, i.e.,If meeting condition, return to step three;Otherwise output is special
Collect B.
Implementing procedure is shown in Fig. 2, is specifically:
(1) input data set U, window width h, outage threshold T
(2) character subset is obtained by the feature selection approach proposed by the present invention based on composite character KDE conditional entropies
(3) output result
Experimental example 1:
By the operation by the inventive method (abbreviation GS_KDE) on real data collection hepatitis, two kinds are compared for
Using the method for sliding-model control, one kind is wide discrete (abbreviation GS_eqW, section number parameter take 2,4,6), and another kind is
Its validity is shown Deng frequency discrete (abbreviation GS_eqF, section number parameter take 2,4,6).Wherein, every kind of method all selects best
Parameter.The result of operation is as shown in table 1:Wherein, data set derives from disclosed UCI data warehouses (http://
archive.ics.uci.edu/ml);Outage threshold T=0.01, h=k/log2N (k takes 1,2,3), wherein n are data sample
Quantity.Classification accuracy is the average value of five folding cross validations, and the grader used is KNN (k=3), C4.5, PART.
Test result indicates that of the invention classified based on composite character KDE conditional entropies feature selection approach (GS_KDE)
Accuracy rate on better than GS_eqW and GS_eqF, be also better than feature complete or collected works.
The characteristic of table 1 and classification accuracy
It is that the present invention is not limited only to above-described embodiment, as long as meeting for limitation of the invention that above-described embodiment, which is not,
Application claims, belong to protection scope of the present invention.
Claims (8)
1. a kind of feature selection approach based on composite character KDE conditional entropies, it is characterised in that comprise the following steps:
Step 1, input include decision-making feature D data set U, wherein, data set U has a n sample, decision-making feature D=1,
2 ..., N }, discrete features vector Α={ A1,A2,...,Am, continuous characteristic vector X={ X1,X2,...,Xt, window width
H, outage threshold T;
Step 2, if the feature set selected is B, non-selected feature set is E, and initial value is set toE=A ∪ X, are often selected
Select the difference of the conditional entropy before and after a feature
Step 3, temporal aspect collection B ' is established by all properties in each the attribute S and feature set B in feature set E, performed
Following steps;
Step 4, for each value x of continuous feature set X ' in B ', and each value d in decision kind set D and
The middle discrete features collection A ' of B ' each value a, calculate KDE probability With
Step 5, composite character KDE conditional entropies are based on by obtaining KDE probability calculations in step 4WithAnd based on composite character KDE combination entropiesWherein rememberFor Category Attributes collection A ' codomain,
For connection attribute collection X ' codomain,For decision set D codomain;
Step 6, the minimum attribute of alternative condition entropyIt is added in feature set B, has been selected
Attribute B=B ∪ { S* }, and E=E- { S* } is deleted from non-selected feature set;
Step 7, the difference of the conditional entropy added before and after new attribute is obtained by B=B ∪ { S* } in step 6, i.e.,
Step 8, the difference of the conditional entropy of judgment step sevenWhether the Characteristic Number being more than in threshold value T and feature set B is less than
The characteristic of data lump, i.e.,If meeting condition, return to step three;Otherwise output characteristic
Collect B.
A kind of 2. feature selection approach based on composite character KDE conditional entropies according to claim 1, it is characterised in that
KDE probability in the step 4Generated by formula (1):
<mrow>
<mtable>
<mtr>
<mtd>
<mrow>
<mover>
<mi>p</mi>
<mo>^</mo>
</mover>
<mrow>
<mo>(</mo>
<mi>a</mi>
<mo>,</mo>
<mi>x</mi>
<mo>,</mo>
<mi>d</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<mover>
<mi>p</mi>
<mo>^</mo>
</mover>
<mrow>
<mo>(</mo>
<mi>d</mi>
<mo>,</mo>
<mi>a</mi>
<mo>)</mo>
</mrow>
<mover>
<mi>p</mi>
<mo>^</mo>
</mover>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>|</mo>
<mi>d</mi>
<mo>,</mo>
<mi>a</mi>
<mo>)</mo>
</mrow>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<mo>=</mo>
<mfrac>
<msub>
<mi>n</mi>
<mrow>
<mi>d</mi>
<mo>,</mo>
<mi>a</mi>
</mrow>
</msub>
<mi>n</mi>
</mfrac>
<mfrac>
<mn>1</mn>
<msub>
<mi>n</mi>
<mrow>
<mi>d</mi>
<mo>,</mo>
<mi>a</mi>
</mrow>
</msub>
</mfrac>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>I</mi>
<mrow>
<mi>d</mi>
<mo>,</mo>
<mi>a</mi>
</mrow>
</msub>
</mrow>
</munder>
<mi>&phi;</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>,</mo>
<mi>h</mi>
<mo>)</mo>
</mrow>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<mo>=</mo>
<mfrac>
<mn>1</mn>
<mi>n</mi>
</mfrac>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>I</mi>
<mrow>
<mi>d</mi>
<mo>,</mo>
<mi>a</mi>
</mrow>
</msub>
</mrow>
</munder>
<mi>&phi;</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>,</mo>
<mi>h</mi>
<mo>)</mo>
</mrow>
</mrow>
</mtd>
</mtr>
</mtable>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>1</mn>
<mo>)</mo>
</mrow>
</mrow>
A kind of 3. feature selection approach based on composite character KDE conditional entropies according to claim 1, it is characterised in that
The step 4 KDE probabilityGenerated by formula (2):
<mrow>
<mtable>
<mtr>
<mtd>
<mrow>
<mover>
<mi>p</mi>
<mo>^</mo>
</mover>
<mrow>
<mo>(</mo>
<mi>d</mi>
<mo>|</mo>
<mi>x</mi>
<mo>,</mo>
<mi>a</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<mfrac>
<mrow>
<mover>
<mi>p</mi>
<mo>^</mo>
</mover>
<mrow>
<mo>(</mo>
<mi>d</mi>
<mo>,</mo>
<mi>x</mi>
<mo>,</mo>
<mi>a</mi>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<mover>
<mi>p</mi>
<mo>^</mo>
</mover>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>a</mi>
<mo>)</mo>
</mrow>
</mrow>
</mfrac>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<mo>=</mo>
<mfrac>
<mrow>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>I</mi>
<mrow>
<mi>d</mi>
<mo>,</mo>
<mi>a</mi>
</mrow>
</msub>
</mrow>
</munder>
<mi>&phi;</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>,</mo>
<mi>h</mi>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>I</mi>
<mi>a</mi>
</msub>
</mrow>
</munder>
<mi>&phi;</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>,</mo>
<mi>h</mi>
<mo>)</mo>
</mrow>
</mrow>
</mfrac>
</mrow>
</mtd>
</mtr>
</mtable>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>2</mn>
<mo>)</mo>
</mrow>
</mrow>
1
A kind of 4. feature selection approach based on composite character KDE conditional entropies according to claim 1, it is characterised in that
The step 4 KDE probabilityGenerated by formula (3):
<mrow>
<mtable>
<mtr>
<mtd>
<mrow>
<mover>
<mi>p</mi>
<mo>^</mo>
</mover>
<mrow>
<mo>(</mo>
<mi>d</mi>
<mo>,</mo>
<mi>a</mi>
<mo>|</mo>
<mi>x</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<mfrac>
<mrow>
<mover>
<mi>p</mi>
<mo>^</mo>
</mover>
<mrow>
<mo>(</mo>
<mi>d</mi>
<mo>,</mo>
<mi>x</mi>
<mo>,</mo>
<mi>a</mi>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<mover>
<mi>p</mi>
<mo>^</mo>
</mover>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>)</mo>
</mrow>
</mrow>
</mfrac>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<mo>=</mo>
<mfrac>
<mrow>
<mfrac>
<mn>1</mn>
<mi>n</mi>
</mfrac>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>I</mi>
<mrow>
<mi>d</mi>
<mo>,</mo>
<mi>a</mi>
</mrow>
</msub>
</mrow>
</munder>
<mi>&phi;</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>,</mo>
<mi>h</mi>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<mfrac>
<mn>1</mn>
<mi>n</mi>
</mfrac>
<munderover>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mi>n</mi>
</munderover>
<mi>&phi;</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>,</mo>
<mi>h</mi>
<mo>)</mo>
</mrow>
</mrow>
</mfrac>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<mo>=</mo>
<mfrac>
<mrow>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>I</mi>
<mrow>
<mi>d</mi>
<mo>,</mo>
<mi>a</mi>
</mrow>
</msub>
</mrow>
</munder>
<mi>&phi;</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>,</mo>
<mi>h</mi>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<munderover>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mi>n</mi>
</munderover>
<mi>&phi;</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>,</mo>
<mi>h</mi>
<mo>)</mo>
</mrow>
</mrow>
</mfrac>
</mrow>
</mtd>
</mtr>
</mtable>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>3</mn>
<mo>)</mo>
</mrow>
</mrow>
A kind of 5. feature selection approach based on composite character KDE conditional entropies according to claim 1, it is characterised in that
The step 4 KDE probabilityGenerated by formula (4):
<mrow>
<mtable>
<mtr>
<mtd>
<mrow>
<mover>
<mi>p</mi>
<mo>^</mo>
</mover>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>a</mi>
<mo>|</mo>
<mi>d</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<mfrac>
<mrow>
<mover>
<mi>p</mi>
<mo>^</mo>
</mover>
<mrow>
<mo>(</mo>
<mi>d</mi>
<mo>,</mo>
<mi>x</mi>
<mo>,</mo>
<mi>a</mi>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<mi>p</mi>
<mrow>
<mo>(</mo>
<mi>d</mi>
<mo>)</mo>
</mrow>
</mrow>
</mfrac>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<mo>=</mo>
<mfrac>
<mrow>
<mfrac>
<mn>1</mn>
<mi>n</mi>
</mfrac>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>I</mi>
<mrow>
<mi>d</mi>
<mo>,</mo>
<mi>a</mi>
</mrow>
</msub>
</mrow>
</munder>
<mi>&phi;</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>,</mo>
<mi>h</mi>
<mo>)</mo>
</mrow>
</mrow>
<mfrac>
<msub>
<mi>n</mi>
<mi>d</mi>
</msub>
<mi>n</mi>
</mfrac>
</mfrac>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<mo>=</mo>
<mfrac>
<mn>1</mn>
<msub>
<mi>n</mi>
<mi>d</mi>
</msub>
</mfrac>
<munder>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>&Element;</mo>
<msub>
<mi>I</mi>
<mrow>
<mi>d</mi>
<mo>,</mo>
<mi>a</mi>
</mrow>
</msub>
</mrow>
</munder>
<mi>&phi;</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<msub>
<mi>x</mi>
<mi>i</mi>
</msub>
<mo>,</mo>
<mi>h</mi>
<mo>)</mo>
</mrow>
</mrow>
</mtd>
</mtr>
</mtable>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>4</mn>
<mo>)</mo>
</mrow>
</mrow>
A kind of 6. feature selection approach based on composite character KDE conditional entropies according to claim 1, it is characterised in that
The step 5 conditional entropyGenerated by formula (5):
A kind of 7. feature selection approach based on composite character KDE conditional entropies according to claim 1, it is characterised in that
The step 5 conditional entropyGenerated by formula (6):
A kind of 8. feature selection approach based on composite character KDE conditional entropies according to claim 1, it is characterised in that
The step 5 conditional entropyGenerated by formula (7):
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710526050.7A CN107391433B (en) | 2017-06-30 | 2017-06-30 | Feature selection method based on KDE conditional entropy of mixed features |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710526050.7A CN107391433B (en) | 2017-06-30 | 2017-06-30 | Feature selection method based on KDE conditional entropy of mixed features |
Publications (2)
Publication Number | Publication Date |
---|---|
CN107391433A true CN107391433A (en) | 2017-11-24 |
CN107391433B CN107391433B (en) | 2021-04-13 |
Family
ID=60334870
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710526050.7A Active CN107391433B (en) | 2017-06-30 | 2017-06-30 | Feature selection method based on KDE conditional entropy of mixed features |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107391433B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115238148A (en) * | 2022-09-21 | 2022-10-25 | 杭州衡泰技术股份有限公司 | Characteristic combination screening method for multi-party enterprise joint credit rating and application |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6317517B1 (en) * | 1998-11-30 | 2001-11-13 | Regents Of The University Of California | Statistical pattern recognition |
WO2009067655A2 (en) * | 2007-11-21 | 2009-05-28 | University Of Florida Research Foundation, Inc. | Methods of feature selection through local learning; breast and prostate cancer prognostic markers |
CN106570887A (en) * | 2016-11-04 | 2017-04-19 | 天津大学 | Adaptive Mean Shift target tracking method based on LBP features |
-
2017
- 2017-06-30 CN CN201710526050.7A patent/CN107391433B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6317517B1 (en) * | 1998-11-30 | 2001-11-13 | Regents Of The University Of California | Statistical pattern recognition |
WO2009067655A2 (en) * | 2007-11-21 | 2009-05-28 | University Of Florida Research Foundation, Inc. | Methods of feature selection through local learning; breast and prostate cancer prognostic markers |
CN106570887A (en) * | 2016-11-04 | 2017-04-19 | 天津大学 | Adaptive Mean Shift target tracking method based on LBP features |
Non-Patent Citations (5)
Title |
---|
MIN HAN,ET AL: "Sparse kerneldensityestimationsanditsapplicationinvariableselection", 《NEUROCOMPUTING》 * |
NOJUN KWAK,ET AL: "Input Feature Selection by MutualInformation Based on Parzen Window", 《IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE》 * |
PABLO A. ESTÉVEZ,ET AL: "Normalized Mutual Information Feature Selection", 《IEEE TRANSACTIONS ON NEURAL NETWORKS》 * |
ZHIHONG ZHANG,ET AL: "KERNEL ENTROPY-BASED UNSUPERVISED SPECTRAL FEATURE SELECTION", 《INTERNATIONAL JOURNAL OF PATTERN RECOGNITION》 * |
刘真勃,等: "期待序信息系统的优势粗糙集模型", 《数码设计》 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115238148A (en) * | 2022-09-21 | 2022-10-25 | 杭州衡泰技术股份有限公司 | Characteristic combination screening method for multi-party enterprise joint credit rating and application |
Also Published As
Publication number | Publication date |
---|---|
CN107391433B (en) | 2021-04-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Negash et al. | Artificial neural network based production forecasting for a hydrocarbon reservoir under water injection | |
Zheng et al. | A rolling bearing fault diagnosis method based on multi-scale fuzzy entropy and variable predictive model-based class discrimination | |
Si et al. | A novel approach for coal seam terrain prediction through information fusion of improved D–S evidence theory and neural network | |
Esmael et al. | Multivariate time series classification by combining trend-based and value-based approximations | |
Boufoussi et al. | Functional differential equations in Hilbert spaces driven by a fractional Brownian motion | |
CN103995237A (en) | Satellite power supply system online fault diagnosis method | |
CN104933444A (en) | Design method of multi-dimension attribute data oriented multi-layered clustering fusion mechanism | |
CN113705099B (en) | Social platform rumor detection model construction method and detection method based on contrast learning | |
CN104915679A (en) | Large-scale high-dimensional data classification method based on random forest weighted distance | |
Singh et al. | An ensemble approach for feature selection of Cyber Attack Dataset | |
CN106549675A (en) | A kind of average dependent quadrature matching pursuit algorithm based on compressed sensing | |
Dorj et al. | Anomaly detection approach using hidden Markov model | |
CN107391433A (en) | A kind of feature selection approach based on composite character KDE conditional entropies | |
Bogdanov et al. | Sktr: Trace recovery from stochastically known logs | |
CN108985462A (en) | Unsupervised feature selection approach based on mutual information and fractal dimension | |
CN105224954A (en) | A kind of topic discover method removing the impact of little topic based on Single-pass | |
CN106919650A (en) | A kind of textural anomaly detection method of increment parallel type Dynamic Graph | |
Mahoney et al. | Trajectory boundary modeling of time series for anomaly detection | |
CN104657473A (en) | Large-scale data mining method capable of guaranteeing quality monotony | |
Chiuso et al. | Learning sparse dynamic linear systems using stable spline kernels and exponential hyperpriors | |
CN111159961A (en) | Abnormity detection method and system based on curve data | |
Cho et al. | Development of locally specified soil stratification method with CPT data based on machine learning techniques | |
Liu et al. | Experimental assessment of gradual deformation method | |
Wei et al. | A symbolic tree model for oil and gas production prediction using time-series production data | |
KR101093521B1 (en) | Pattern analysis method for continuously generating data |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |