CN107391433A - A kind of feature selection approach based on composite character KDE conditional entropies - Google Patents

A kind of feature selection approach based on composite character KDE conditional entropies Download PDF

Info

Publication number
CN107391433A
CN107391433A CN201710526050.7A CN201710526050A CN107391433A CN 107391433 A CN107391433 A CN 107391433A CN 201710526050 A CN201710526050 A CN 201710526050A CN 107391433 A CN107391433 A CN 107391433A
Authority
CN
China
Prior art keywords
mrow
msub
mfrac
kde
mtr
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710526050.7A
Other languages
Chinese (zh)
Other versions
CN107391433B (en
Inventor
代建华
徐思琪
高帅超
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianjin University
Original Assignee
Tianjin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University filed Critical Tianjin University
Priority to CN201710526050.7A priority Critical patent/CN107391433B/en
Publication of CN107391433A publication Critical patent/CN107391433A/en
Application granted granted Critical
Publication of CN107391433B publication Critical patent/CN107391433B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Pure & Applied Mathematics (AREA)
  • Mathematical Optimization (AREA)
  • Algebra (AREA)
  • Computational Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • Software Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Analysis (AREA)
  • Character Discrimination (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The present invention provides a kind of new feature selection approach based on composite character KDE conditional entropies, it this method propose composite character KDE probability and composite character KDE entropys, discrete features and continuous feature are effectively unified in KDE entropys in the case of discrete not by continuous data, information theory has been expanded and has proposed the greedy feature selection approach based on composite character KDE conditional entropies.

Description

A kind of feature selection approach based on composite character KDE conditional entropies
Technical field
The present invention relates to feature selection approach, in particular to a kind of feature selecting side based on composite character KDE conditional entropies Method.
Background technology
With the raising of the storage capacity and computing capability of data, size of data and data dimension are increasing, give Data mining or machine learning task bring bigger pressure.Feature selecting is as data mining, pattern-recognition, machine learning The important pre-treatment step of task, from substantial amounts of attribute, redundancy, unrelated attribute are eliminated, data dimension is reduced, improves The efficiency of algorithm.
The concepts such as entropy and mutual information in information theory occupy consequence in feature selecting, possess and know without priori The advantages that knowing detection non-linear relation, anti-noise jamming.But based on the feature selection approach of information theory primarily directed to from Attribute is dissipated, for continuous feature, takes the mode of discretization mostly to adapt to traditional feature selection approach.Density Estimator (KDE) it is that a kind of probability density function to stochastic variable carries out the method without ginseng estimation.By the entropy knot in KDE and information theory Close, being currently based on the feature selection approach of KDE entropys has preferable effect, but existing method is just for continuous feature. Against this problem, the present invention has expanded information theory so that can be applied to composite character based on KDE entropys.
The content of the invention
The invention aims to handle the feature selecting of composite character, and propose a kind of new based on composite character The feature selection approach of KDE conditional entropies.Composite character KDE entropys are this method propose, information theory has been expanded and has proposed and be based on The greedy feature selection approach of composite character KDE conditional entropies.
It is of the invention to be using technical scheme:
A kind of feature selection approach based on composite character KDE conditional entropies, comprises the following steps:
Step 1, input include decision-making feature D data set U, wherein, data set U has n sample, decision-making feature D= { 1,2 ..., N }, discrete features vector Α={ A1,A2,...,Am, continuous characteristic vector X={ X1,X2,...,Xt, window width ginseng Number h, outage threshold T;
Step 2, if the feature set selected is B, non-selected feature set is E, and initial value is set toE=A ∪ X, The often difference of the conditional entropy before and after one feature of selection
Step 3, temporal aspect collection B ' is established by all properties in each the attribute S and feature set B in feature set E;
Step 4, for each value x of continuous feature set X ' in B ', and each value d in decision kind set D And the middle discrete features collection A ' of B ' each value a, calculate KDE probability With
Step 5, composite character KDE conditional entropies are based on by obtaining KDE probability calculations in step 4WithAnd based on composite character KDE combination entropiesWherein rememberFor Category Attributes collection A ' codomain,For Connection attribute collection X ' codomain,For decision set D codomain;
Step 6, the minimum attribute of alternative condition entropyIt is added in feature set B, obtains Attribute B=B ∪ { S* } have been selected, and E=E- { S* } is deleted from non-selected feature set;
Step 7, the difference of the conditional entropy added before and after new attribute is obtained by B=B ∪ { S* } in step 6, i.e.,
Step 8, the difference of the conditional entropy of judgment step sevenThe Characteristic Number whether being more than in threshold value T and feature set B Less than the characteristic of data lump, i.e.,If meeting condition, return to step three;Otherwise export Feature set B.
KDE probability in the step 4Generated by formula (1):
KDE probability in the step 4Generated by formula (2):
KDE probability in the step 4Generated by formula (3):
KDE probability in the step 4Generated by formula (4):
The step 5 conditional entropyGenerated by formula (5):
The step 5 conditional entropyGenerated by formula (6):
The step 5 conditional entropyGenerated by formula (7):
Wherein, due to the sample probability of continuous random variable,
The beneficial effects of the invention are as follows:
1st, the present invention proposes composite character KDE probability and composite character KDE entropys, not by the discrete situation of continuous data It is lower that discrete features and continuous feature are effectively unified in KDE entropys.
2nd, the present invention is used based on standard of the composite character KDE conditional entropies as evaluating characteristic, is carried out using greedy algorithm Feature selecting.
3rd, the method that the present invention passes through Experimental comparison's discretization continuous data, it was demonstrated that this algorithm has in various classification experiments More preferable effect.
Brief description of the drawings
Fig. 1 is flow chart of the method for the present invention;
Fig. 2 is the implementing procedure figure of the present invention.
Embodiment
The present invention is further analyzed with reference to specific embodiment.
The method flow of the present invention is shown in Fig. 1, and based on the definition of above-mentioned composite character KDE entropys, the present invention is based on composite character The feature selection approach of KDE conditional entropies is described in detail below:
Step 1 101, input include decision-making feature D data set U;Wherein, data set U has n sample, decision-making feature D ={ 1,2 ..., N }, discrete features vector Α={ A1,A2,...,Am, continuous characteristic vector X={ X1,X2,...,Xt, window width Parameter h, outage threshold T;
Step 2 102, if the feature set selected is B, non-selected feature set is E, and initial value is set toE=A ∪ X, the often conditional entropy before and after one feature of selection difference
Step 3 103, temporal aspect collection is established by all properties in each the attribute S and feature set B in feature set E B ', perform following steps;
Step 4 104, for each value x of continuous feature set X ' in B ', and it is every in decision kind set D A kind of middle discrete features collection A ' of value d and B ' each value a, calculate KDE probability With
The KDE probabilityGenerated by formula (1):
The KDE probabilityGenerated by formula (2):
The KDE probabilityGenerated by formula (3):
The KDE probabilityGenerated by formula (4):
Step 5 105, composite character KDE conditional entropies are based on by obtaining KDE probability calculations in step 4WithAnd based on composite character KDE combination entropiesWherein rememberFor Category Attributes collection A ' codomain, For connection attribute collection X ' codomain,For decision set D codomain;
The step 5 conditional entropyGenerated by formula (5):
The step 5 conditional entropyGenerated by formula (6):
The step 5 conditional entropyGenerated by formula (7):
Wherein, due to the sample probability of continuous random variable,
Step 6 106, the minimum attribute of alternative condition entropyIt is added in feature set B, Attribute B=B ∪ { S* } have been selected in acquisition, and E=E- { S* } is deleted from non-selected feature set;
Step 7 107, the difference of the conditional entropy added before and after new attribute is obtained by B=B ∪ { S* } in step 6, i.e.,
Step 8 108, the conditional entropy of judgment step sevenThe Characteristic Number whether being more than in threshold value T and feature set B is small In the characteristic of data lump, i.e.,If meeting condition, return to step three;Otherwise output is special Collect B.
Implementing procedure is shown in Fig. 2, is specifically:
(1) input data set U, window width h, outage threshold T
(2) character subset is obtained by the feature selection approach proposed by the present invention based on composite character KDE conditional entropies
(3) output result
Experimental example 1:
By the operation by the inventive method (abbreviation GS_KDE) on real data collection hepatitis, two kinds are compared for Using the method for sliding-model control, one kind is wide discrete (abbreviation GS_eqW, section number parameter take 2,4,6), and another kind is Its validity is shown Deng frequency discrete (abbreviation GS_eqF, section number parameter take 2,4,6).Wherein, every kind of method all selects best Parameter.The result of operation is as shown in table 1:Wherein, data set derives from disclosed UCI data warehouses (http:// archive.ics.uci.edu/ml);Outage threshold T=0.01, h=k/log2N (k takes 1,2,3), wherein n are data sample Quantity.Classification accuracy is the average value of five folding cross validations, and the grader used is KNN (k=3), C4.5, PART.
Test result indicates that of the invention classified based on composite character KDE conditional entropies feature selection approach (GS_KDE) Accuracy rate on better than GS_eqW and GS_eqF, be also better than feature complete or collected works.
The characteristic of table 1 and classification accuracy
It is that the present invention is not limited only to above-described embodiment, as long as meeting for limitation of the invention that above-described embodiment, which is not, Application claims, belong to protection scope of the present invention.

Claims (8)

1. a kind of feature selection approach based on composite character KDE conditional entropies, it is characterised in that comprise the following steps:
Step 1, input include decision-making feature D data set U, wherein, data set U has a n sample, decision-making feature D=1, 2 ..., N }, discrete features vector Α={ A1,A2,...,Am, continuous characteristic vector X={ X1,X2,...,Xt, window width H, outage threshold T;
Step 2, if the feature set selected is B, non-selected feature set is E, and initial value is set toE=A ∪ X, are often selected Select the difference of the conditional entropy before and after a feature
Step 3, temporal aspect collection B ' is established by all properties in each the attribute S and feature set B in feature set E, performed Following steps;
Step 4, for each value x of continuous feature set X ' in B ', and each value d in decision kind set D and The middle discrete features collection A ' of B ' each value a, calculate KDE probability With
Step 5, composite character KDE conditional entropies are based on by obtaining KDE probability calculations in step 4WithAnd based on composite character KDE combination entropiesWherein rememberFor Category Attributes collection A ' codomain, For connection attribute collection X ' codomain,For decision set D codomain;
Step 6, the minimum attribute of alternative condition entropyIt is added in feature set B, has been selected Attribute B=B ∪ { S* }, and E=E- { S* } is deleted from non-selected feature set;
Step 7, the difference of the conditional entropy added before and after new attribute is obtained by B=B ∪ { S* } in step 6, i.e.,
Step 8, the difference of the conditional entropy of judgment step sevenWhether the Characteristic Number being more than in threshold value T and feature set B is less than The characteristic of data lump, i.e.,If meeting condition, return to step three;Otherwise output characteristic Collect B.
A kind of 2. feature selection approach based on composite character KDE conditional entropies according to claim 1, it is characterised in that KDE probability in the step 4Generated by formula (1):
<mrow> <mtable> <mtr> <mtd> <mrow> <mover> <mi>p</mi> <mo>^</mo> </mover> <mrow> <mo>(</mo> <mi>a</mi> <mo>,</mo> <mi>x</mi> <mo>,</mo> <mi>d</mi> <mo>)</mo> </mrow> <mo>=</mo> <mover> <mi>p</mi> <mo>^</mo> </mover> <mrow> <mo>(</mo> <mi>d</mi> <mo>,</mo> <mi>a</mi> <mo>)</mo> </mrow> <mover> <mi>p</mi> <mo>^</mo> </mover> <mrow> <mo>(</mo> <mi>x</mi> <mo>|</mo> <mi>d</mi> <mo>,</mo> <mi>a</mi> <mo>)</mo> </mrow> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mo>=</mo> <mfrac> <msub> <mi>n</mi> <mrow> <mi>d</mi> <mo>,</mo> <mi>a</mi> </mrow> </msub> <mi>n</mi> </mfrac> <mfrac> <mn>1</mn> <msub> <mi>n</mi> <mrow> <mi>d</mi> <mo>,</mo> <mi>a</mi> </mrow> </msub> </mfrac> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>I</mi> <mrow> <mi>d</mi> <mo>,</mo> <mi>a</mi> </mrow> </msub> </mrow> </munder> <mi>&amp;phi;</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>,</mo> <mi>h</mi> <mo>)</mo> </mrow> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mo>=</mo> <mfrac> <mn>1</mn> <mi>n</mi> </mfrac> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>I</mi> <mrow> <mi>d</mi> <mo>,</mo> <mi>a</mi> </mrow> </msub> </mrow> </munder> <mi>&amp;phi;</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>,</mo> <mi>h</mi> <mo>)</mo> </mrow> </mrow> </mtd> </mtr> </mtable> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>)</mo> </mrow> </mrow>
A kind of 3. feature selection approach based on composite character KDE conditional entropies according to claim 1, it is characterised in that The step 4 KDE probabilityGenerated by formula (2):
<mrow> <mtable> <mtr> <mtd> <mrow> <mover> <mi>p</mi> <mo>^</mo> </mover> <mrow> <mo>(</mo> <mi>d</mi> <mo>|</mo> <mi>x</mi> <mo>,</mo> <mi>a</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mrow> <mover> <mi>p</mi> <mo>^</mo> </mover> <mrow> <mo>(</mo> <mi>d</mi> <mo>,</mo> <mi>x</mi> <mo>,</mo> <mi>a</mi> <mo>)</mo> </mrow> </mrow> <mrow> <mover> <mi>p</mi> <mo>^</mo> </mover> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>a</mi> <mo>)</mo> </mrow> </mrow> </mfrac> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mo>=</mo> <mfrac> <mrow> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>I</mi> <mrow> <mi>d</mi> <mo>,</mo> <mi>a</mi> </mrow> </msub> </mrow> </munder> <mi>&amp;phi;</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>,</mo> <mi>h</mi> <mo>)</mo> </mrow> </mrow> <mrow> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>I</mi> <mi>a</mi> </msub> </mrow> </munder> <mi>&amp;phi;</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>,</mo> <mi>h</mi> <mo>)</mo> </mrow> </mrow> </mfrac> </mrow> </mtd> </mtr> </mtable> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>2</mn> <mo>)</mo> </mrow> </mrow> 1
A kind of 4. feature selection approach based on composite character KDE conditional entropies according to claim 1, it is characterised in that The step 4 KDE probabilityGenerated by formula (3):
<mrow> <mtable> <mtr> <mtd> <mrow> <mover> <mi>p</mi> <mo>^</mo> </mover> <mrow> <mo>(</mo> <mi>d</mi> <mo>,</mo> <mi>a</mi> <mo>|</mo> <mi>x</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mrow> <mover> <mi>p</mi> <mo>^</mo> </mover> <mrow> <mo>(</mo> <mi>d</mi> <mo>,</mo> <mi>x</mi> <mo>,</mo> <mi>a</mi> <mo>)</mo> </mrow> </mrow> <mrow> <mover> <mi>p</mi> <mo>^</mo> </mover> <mrow> <mo>(</mo> <mi>x</mi> <mo>)</mo> </mrow> </mrow> </mfrac> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mo>=</mo> <mfrac> <mrow> <mfrac> <mn>1</mn> <mi>n</mi> </mfrac> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>I</mi> <mrow> <mi>d</mi> <mo>,</mo> <mi>a</mi> </mrow> </msub> </mrow> </munder> <mi>&amp;phi;</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>,</mo> <mi>h</mi> <mo>)</mo> </mrow> </mrow> <mrow> <mfrac> <mn>1</mn> <mi>n</mi> </mfrac> <munderover> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>n</mi> </munderover> <mi>&amp;phi;</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>,</mo> <mi>h</mi> <mo>)</mo> </mrow> </mrow> </mfrac> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mo>=</mo> <mfrac> <mrow> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>I</mi> <mrow> <mi>d</mi> <mo>,</mo> <mi>a</mi> </mrow> </msub> </mrow> </munder> <mi>&amp;phi;</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>,</mo> <mi>h</mi> <mo>)</mo> </mrow> </mrow> <mrow> <munderover> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>n</mi> </munderover> <mi>&amp;phi;</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>,</mo> <mi>h</mi> <mo>)</mo> </mrow> </mrow> </mfrac> </mrow> </mtd> </mtr> </mtable> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>3</mn> <mo>)</mo> </mrow> </mrow>
A kind of 5. feature selection approach based on composite character KDE conditional entropies according to claim 1, it is characterised in that The step 4 KDE probabilityGenerated by formula (4):
<mrow> <mtable> <mtr> <mtd> <mrow> <mover> <mi>p</mi> <mo>^</mo> </mover> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>a</mi> <mo>|</mo> <mi>d</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mrow> <mover> <mi>p</mi> <mo>^</mo> </mover> <mrow> <mo>(</mo> <mi>d</mi> <mo>,</mo> <mi>x</mi> <mo>,</mo> <mi>a</mi> <mo>)</mo> </mrow> </mrow> <mrow> <mi>p</mi> <mrow> <mo>(</mo> <mi>d</mi> <mo>)</mo> </mrow> </mrow> </mfrac> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mo>=</mo> <mfrac> <mrow> <mfrac> <mn>1</mn> <mi>n</mi> </mfrac> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>I</mi> <mrow> <mi>d</mi> <mo>,</mo> <mi>a</mi> </mrow> </msub> </mrow> </munder> <mi>&amp;phi;</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>,</mo> <mi>h</mi> <mo>)</mo> </mrow> </mrow> <mfrac> <msub> <mi>n</mi> <mi>d</mi> </msub> <mi>n</mi> </mfrac> </mfrac> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mo>=</mo> <mfrac> <mn>1</mn> <msub> <mi>n</mi> <mi>d</mi> </msub> </mfrac> <munder> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>&amp;Element;</mo> <msub> <mi>I</mi> <mrow> <mi>d</mi> <mo>,</mo> <mi>a</mi> </mrow> </msub> </mrow> </munder> <mi>&amp;phi;</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>,</mo> <mi>h</mi> <mo>)</mo> </mrow> </mrow> </mtd> </mtr> </mtable> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>4</mn> <mo>)</mo> </mrow> </mrow>
A kind of 6. feature selection approach based on composite character KDE conditional entropies according to claim 1, it is characterised in that The step 5 conditional entropyGenerated by formula (5):
A kind of 7. feature selection approach based on composite character KDE conditional entropies according to claim 1, it is characterised in that The step 5 conditional entropyGenerated by formula (6):
A kind of 8. feature selection approach based on composite character KDE conditional entropies according to claim 1, it is characterised in that The step 5 conditional entropyGenerated by formula (7):
CN201710526050.7A 2017-06-30 2017-06-30 Feature selection method based on KDE conditional entropy of mixed features Active CN107391433B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710526050.7A CN107391433B (en) 2017-06-30 2017-06-30 Feature selection method based on KDE conditional entropy of mixed features

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710526050.7A CN107391433B (en) 2017-06-30 2017-06-30 Feature selection method based on KDE conditional entropy of mixed features

Publications (2)

Publication Number Publication Date
CN107391433A true CN107391433A (en) 2017-11-24
CN107391433B CN107391433B (en) 2021-04-13

Family

ID=60334870

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710526050.7A Active CN107391433B (en) 2017-06-30 2017-06-30 Feature selection method based on KDE conditional entropy of mixed features

Country Status (1)

Country Link
CN (1) CN107391433B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115238148A (en) * 2022-09-21 2022-10-25 杭州衡泰技术股份有限公司 Characteristic combination screening method for multi-party enterprise joint credit rating and application

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6317517B1 (en) * 1998-11-30 2001-11-13 Regents Of The University Of California Statistical pattern recognition
WO2009067655A2 (en) * 2007-11-21 2009-05-28 University Of Florida Research Foundation, Inc. Methods of feature selection through local learning; breast and prostate cancer prognostic markers
CN106570887A (en) * 2016-11-04 2017-04-19 天津大学 Adaptive Mean Shift target tracking method based on LBP features

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6317517B1 (en) * 1998-11-30 2001-11-13 Regents Of The University Of California Statistical pattern recognition
WO2009067655A2 (en) * 2007-11-21 2009-05-28 University Of Florida Research Foundation, Inc. Methods of feature selection through local learning; breast and prostate cancer prognostic markers
CN106570887A (en) * 2016-11-04 2017-04-19 天津大学 Adaptive Mean Shift target tracking method based on LBP features

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
MIN HAN,ET AL: "Sparse kerneldensityestimationsanditsapplicationinvariableselection", 《NEUROCOMPUTING》 *
NOJUN KWAK,ET AL: "Input Feature Selection by MutualInformation Based on Parzen Window", 《IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE》 *
PABLO A. ESTÉVEZ,ET AL: "Normalized Mutual Information Feature Selection", 《IEEE TRANSACTIONS ON NEURAL NETWORKS》 *
ZHIHONG ZHANG,ET AL: "KERNEL ENTROPY-BASED UNSUPERVISED SPECTRAL FEATURE SELECTION", 《INTERNATIONAL JOURNAL OF PATTERN RECOGNITION》 *
刘真勃,等: "期待序信息系统的优势粗糙集模型", 《数码设计》 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115238148A (en) * 2022-09-21 2022-10-25 杭州衡泰技术股份有限公司 Characteristic combination screening method for multi-party enterprise joint credit rating and application

Also Published As

Publication number Publication date
CN107391433B (en) 2021-04-13

Similar Documents

Publication Publication Date Title
Negash et al. Artificial neural network based production forecasting for a hydrocarbon reservoir under water injection
Zheng et al. A rolling bearing fault diagnosis method based on multi-scale fuzzy entropy and variable predictive model-based class discrimination
Si et al. A novel approach for coal seam terrain prediction through information fusion of improved D–S evidence theory and neural network
Esmael et al. Multivariate time series classification by combining trend-based and value-based approximations
Boufoussi et al. Functional differential equations in Hilbert spaces driven by a fractional Brownian motion
CN103995237A (en) Satellite power supply system online fault diagnosis method
CN104933444A (en) Design method of multi-dimension attribute data oriented multi-layered clustering fusion mechanism
CN113705099B (en) Social platform rumor detection model construction method and detection method based on contrast learning
CN104915679A (en) Large-scale high-dimensional data classification method based on random forest weighted distance
Singh et al. An ensemble approach for feature selection of Cyber Attack Dataset
CN106549675A (en) A kind of average dependent quadrature matching pursuit algorithm based on compressed sensing
Dorj et al. Anomaly detection approach using hidden Markov model
CN107391433A (en) A kind of feature selection approach based on composite character KDE conditional entropies
Bogdanov et al. Sktr: Trace recovery from stochastically known logs
CN108985462A (en) Unsupervised feature selection approach based on mutual information and fractal dimension
CN105224954A (en) A kind of topic discover method removing the impact of little topic based on Single-pass
CN106919650A (en) A kind of textural anomaly detection method of increment parallel type Dynamic Graph
Mahoney et al. Trajectory boundary modeling of time series for anomaly detection
CN104657473A (en) Large-scale data mining method capable of guaranteeing quality monotony
Chiuso et al. Learning sparse dynamic linear systems using stable spline kernels and exponential hyperpriors
CN111159961A (en) Abnormity detection method and system based on curve data
Cho et al. Development of locally specified soil stratification method with CPT data based on machine learning techniques
Liu et al. Experimental assessment of gradual deformation method
Wei et al. A symbolic tree model for oil and gas production prediction using time-series production data
KR101093521B1 (en) Pattern analysis method for continuously generating data

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant