CN111507365A - Confidence rule automatic generation method based on fuzzy clustering - Google Patents

Confidence rule automatic generation method based on fuzzy clustering Download PDF

Info

Publication number
CN111507365A
CN111507365A CN201910821869.5A CN201910821869A CN111507365A CN 111507365 A CN111507365 A CN 111507365A CN 201910821869 A CN201910821869 A CN 201910821869A CN 111507365 A CN111507365 A CN 111507365A
Authority
CN
China
Prior art keywords
rule
confidence
attribute
weight
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910821869.5A
Other languages
Chinese (zh)
Inventor
王晓丽
吕星晓
阳春华
周佳怡
桂卫华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Central South University
Original Assignee
Central South University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Central South University filed Critical Central South University
Priority to CN201910821869.5A priority Critical patent/CN111507365A/en
Publication of CN111507365A publication Critical patent/CN111507365A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a confidence rule automatic generation method based on fuzzy clustering, and belongs to the technical field of data mining. Aiming at the problem of time and labor consumption of establishing a confidence rule base completely depending on manual experience, on the basis of the method for expanding the confidence rule base, the method firstly carries out fuzzy clustering on sample data, and takes a fuzzy clustering center of the sample and the sample which takes the clustering center as a circle center and takes the distance from the clustering center as a certain set value as basic data for generating the expanded confidence rule base; and calculating the precondition attribute weight and the rule weight according to the variable incidence relation and the sample membership degree matrix. The invention aims to solve the problems that sample data is difficult to select and is not suitable for large sample data modeling in the method for expanding the confidence rule base. The invention reduces the dependence degree of experience setting precondition attribute weight and initial rule weight on expert experience in the existing method for expanding the confidence rule base. By adopting the method and the device, the confidence rule is extracted from the actual production data, and compared with the manual setting of the initial confidence rule, the generation efficiency and the reasoning precision of the confidence rule base are improved.

Description

Confidence rule automatic generation method based on fuzzy clustering
Technical Field
The invention relates to a confidence rule automatic generation method based on fuzzy clustering, and belongs to the technical field of data mining.
Background
Automatic data transfer to knowledge is a technology necessary for intelligent manufacturing development. The confidence rule base method is a knowledge-based method, is developed on the basis of a D-S evidence theory, a decision theory, a fuzzy theory and a traditional IF-THEN rule base, and has the capability of modeling various uncertain and nonlinear data. Compared with the traditional IF-THEN generation type rule, the confidence rule adds weight parameters such as rule weight, premise attribute weight, result confidence coefficient and the like, so that the rule description range is expanded, the randomness and the fuzziness of knowledge can be fully combined, and a knowledge expression mode which is closer to reality and can contain more information is provided. The confidence rule base method can utilize quantitative production data with large amount of uncertainty and qualitative expert knowledge in the production process, has strong model modeling capacity on data with fuzzy uncertainty or probability uncertainty, or collected incomplete data, or data with remarkable nonlinear characteristics, and can better deal with the situations. The confidence rule base method has been successfully applied to the detection, evaluation and prediction of problems in the fields of engineering management, financial decision, military, medical treatment and the like. However, the value of the weight parameter of the confidence rule base system is initially given by an expert, and the accuracy of the weight parameter directly affects the prediction performance of the system. Because the personal experience and judgment of experts have certain subjectivity and are not accurate enough, especially when modeling industrial complex problems, the initial confidence rule base established by the experts often cannot meet the precision requirement and cannot adapt to the change of the situation. Therefore, a parameter optimization learning model of the confidence rule base system is usually required to be established, parameters such as rule weight, premise attribute weight and result confidence coefficient are optimized, and the inference precision of the confidence rule base system is improved. That is, the confidence rule base needs to be built not only by expert knowledge, but also by a time-consuming parameter learning process.
The extended confidence rule base method can solve the problem that the traditional confidence rule base is completely dependent on manual experience to establish time and labor, does not need a complicated initial confidence rule base generation mechanism and a time-consuming parameter iteration optimization training process, and can directly, simply and efficiently and automatically generate the extended confidence rule base from sample data. The extended confidence rule base method adopts a data driving mode to generate the confidence rules, each sample data correspondingly generates an extended confidence rule, and the number of the sample data determines the rule number of the extended confidence rule base. When the extended confidence rule base method is applied to complex problems, if the number of selected samples is small, incomplete rules are easily caused, and the inference precision of the established rule base is not high; if the number of the selected samples is too large, the samples are susceptible to noise data and rule redundancy is caused, and the inference precision is influenced when the rules are activated for inference. In addition, the quality of the sample data also affects the reasoning ability of the extended confidence rule base, and actually, the sample data often has incompleteness caused by incomplete data and inconsistency caused by mutual contradiction between data. Therefore, in the extended confidence rule base method, the selection of sample data is a difficult problem and is not suitable for large sample data modeling. In addition, it also requires the expert to give the weight of the premise attribute and the weight of the initial rule, and cannot completely avoid the dependence on the expert experience knowledge.
Aiming at the existing problems, on the basis of an expanded confidence rule base method, the process of avoiding manually selecting sample data to improve reasoning accuracy is researched, the dependence on expert experience knowledge is reduced, a confidence rule base model is established, a real automatic reasoning target is realized, and the method has very important significance for improving the generation efficiency and the reasoning accuracy of a confidence rule base and the knowledge automation and control intelligence of an industrial process.
Disclosure of Invention
The invention aims to solve the problem that the selection of sample data in the method for expanding the confidence rule base is difficult to cause that the sample data is not suitable for modeling large sample data; and the dependence of experience setting precondition attribute weight and initial rule weight on expert experience in the existing method for expanding the confidence rule base is reduced.
The technical scheme of the invention is as follows:
firstly, carrying out fuzzy clustering on a large amount of historical sample data by using a fuzzy C-means clustering algorithm to obtain a fuzzy clustering center V1, re-clustering the sample data of which the elements in the membership degree matrix are all smaller than 0.01 to obtain a fuzzy clustering center V2, and continuously re-clustering the sample data of which the elements in the membership degree matrix are all smaller than 0.01 until the elements in the membership degree matrix are not smaller than 0.01, thereby obtaining final fuzzy clustering center data (V1 ∪ V2 ∪ …).
And taking the obtained fuzzy clustering center data (V1 ∪ V2 ∪ …) as basic data for constructing an expanded confidence rule base, and determining the number of reference levels and utility values of each pre-condition attribute and conclusion attribute.
Converting the input-output pairs in the basic data into a confidence distribution expression form of the premise attributes and conclusion attributes of the extended confidence rule base shown in formula (1):
Figure RE-GDA0002422948340000021
wherein, TkIs the number of premise attributes in the kth rule, JiThe number of reference levels for the ith precondition attribute, N is the number of output result levels, θkRule weight of the kth rule;iis the attribute weight;
Figure BDA0002187776990000022
a utility value representing the jth reference level of the ith precondition attribute in the kth rule, where i is 1,2, L, Tk,j=1,2,L,Ji;JiThe number of reference levels representing the ith precondition attribute;
Figure BDA0002187776990000023
the confidence of the corresponding reference level is indicated, where i ═ 1,2, L, Tk,j=1,2,L,Ji;DnN is 1, 2L, N is the evaluation level of the output result, βn,kIndicates the evaluation result D in the k-th rulenThe confidence of (c).
For input-output pairsInputting a value xiIts precondition attribute is UiAccording to the formula (3), x can be expressediIs converted into UiAs shown in equation (2):
S(xi)={(Aijij);j=1,2,L,Ji} (2)
wherein the content of the first and second substances,
Figure BDA0002187776990000024
wherein, for the ith precondition attribute UiThe number of reference levels is JiAmong these different reference levels, the effective value of the jth reference level is set to Aij。αijFor a given value xiConversion to prerequisite Attribute UiBy examining the value xiThe degree of similarity to the corresponding reference level utility value is obtained.
Similarly, for the output value y in an input-output pairiIts conclusion attribute is ViAccording to the formula (5), y can be expressediConversion to ViAs shown in equation (4):
S(yi)={(Dijij);j=1,2,L,N} (4)
wherein the content of the first and second substances,
Figure BDA0002187776990000031
wherein N represents the number of evaluation levels of conclusion attributes; dijThe utility value of the jth evaluation level in the N evaluation levels representing the ith conclusion attribute βijRepresents a value yiTransition to conclusion Attribute ViThe confidence level at the jth evaluation level of (a) is also obtained by comparing the utility value of the corresponding evaluation level with the value yiThe degree of similarity of (a).
And then, calculating the weight of the precondition attribute based on the incidence relation of the variables, namely, respectively calculating the Pearson correlation coefficient of each precondition attribute and the conclusion attribute by using historical sample data, and normalizing the correlation coefficient to obtain the weight of the precondition attribute.
Calculating rule weight based on sample membership matrix, firstly calculating initial rule weight thetap. Assuming that the pth rule is generated by the pth clustering center, the ratio of the number of sample data belonging to the pth class to the total number of the whole sample data is the initial rule weight thetap. Then, according to the extended confidence rule base method, the rule weight is updated according to the inconsistency degree of the rule. The steps of updating the rule weights are as follows:
the first step is as follows: rule similarity is calculated.
First, the calculation formula (7) for similarity is obtained from the euclidean distance:
let X be { X ═ X1,x2,L,xn},Y={y1,y2,L,ynThe euclidean distance can be defined as follows:
Figure BDA0002187776990000032
then:
Sim(X,Y)=1-d(X,Y) (7)
rule similarities include the first piece of Similarity (SRA) and the second piece of Similarity (SRC). Assume that there are two rules with only one conclusion attribute:
Figure BDA0002187776990000033
Figure BDA0002187776990000034
wherein, UTRepresenting the Tth precondition attribute, wherein T represents the number of the precondition attributes;
Figure BDA0002187776990000035
representing the confidence distribution of the t-th precondition attribute in the ith rule; diRepresenting the confidence distribution of the conclusion attribute of the ith rule.
Then, the front piece similarity SRA and the back piece similarity SRC of the above two rules are defined as follows:
Figure BDA0002187776990000041
SRC(i,k)=Sim(Di,Dk) (11)
wherein, the similarity between the ith rule and the tth precondition attribute of the kth rule in the formula (10) can be calculated according to the formula (7)
Figure BDA0002187776990000042
And similarity Sim (D) of conclusion attributes of the ith rule and the kth rule in equation (11)i,Dk)。
The second step is that: and calculating the rule inconsistency.
First, using SRA and SRC, the degree of agreement between two rules is calculated as follows:
Cons(Ri,Rk)=exp{-(SRA(i,k)/SRC(i,k)-1.0)2/(1/SRA(i,k))2} (12)
wherein, Cons (R)i,Rk)∈[0,1]。
Then, the inconsistency of the ith rule for the whole rule base is calculated:
Figure BDA0002187776990000043
wherein, L*For the number of generated rules, then, the overall degree of inconsistency of the rule base is calculated:
Figure BDA0002187776990000044
the third step: the rule weights are updated.
The formula for calculating the rule weight is as follows:
Figure BDA0002187776990000045
wherein the content of the first and second substances,
Figure BDA0002187776990000046
the constant λ is a function of the update weight of the pth rule, depending on the particular application.
Therefore, a confidence rule base can be constructed and completed, and automatic generation of the confidence rules is realized.
The invention has the beneficial effects that: firstly, fuzzy clustering is carried out on sample data, a fuzzy clustering center of a sample and the sample which takes the clustering center as a circle center and has a certain set value of distance from the clustering center are used as basic data for generating an extended confidence rule base, and the problem that the data sample is difficult to select is solved; and then, the precondition attribute weight and the rule weight are calculated according to the variable incidence relation and the sample membership matrix, so that the degree of dependence on expert experience knowledge is reduced, and the defect of expanding the construction of a confidence rule base is overcome. By adopting the method and the device, the confidence rule is extracted from the actual production data, and compared with the manual setting of the initial confidence rule, the generation efficiency and the reasoning precision of the confidence rule base are improved.
Drawings
FIG. 1 is a flow chart of the fuzzy C-means clustering algorithm of the present invention;
fig. 2 is a flowchart of the confidence rule automatic generation method based on fuzzy clustering according to the present invention.
Detailed Description
Specific embodiments of the present invention will now be described with reference to the accompanying drawings, it being understood that the embodiments shown and described in the drawings are merely illustrative and are intended to illustrate the principles and methods of the present invention and not to limit the scope of the invention.
The invention relates to a confidence rule automatic generation method based on fuzzy clustering. Firstly, on the basis of carrying out fuzzy clustering on sample data for multiple times by using a fuzzy C-means clustering algorithm, taking an obtained fuzzy clustering center as basic data for generating an expanded confidence rule base; and calculating the weight of the precondition attribute and the rule weight according to the incidence relation of the variables and the membership matrix of the sample, thereby realizing the automatic generation of the confidence rule. The invention realizes the optimization control of the basic dosage of antimony roughing by using an automatic generation method of a fuzzy clustering confidence rule based on industrial production data. 530 groups of industrial historical data continuously collected in an antimony roughing procedure of a certain gold and antimony flotation plant in China are taken as cases for analysis, wherein 330 groups of industrial historical data are taken as training data, and the remaining 200 groups of industrial historical data are taken as test data. The data comprises ore feeding conditions, foam image characteristics, medicine adding amount and other process data, wherein the medicine adding amount comprises xanthate, black powder, copper sulfate, lead nitrate and second oil.
FIG. 1 shows a flow chart of the fuzzy C-means clustering algorithm of the present invention. Suppose the dataset to be clustered is X ═ X1,x2,L,xk,L,xnAnd (4) classifying the same into c types. The objective function of the fuzzy C-means clustering algorithm is shown in formula (1):
Figure BDA0002187776990000051
wherein
dik=||xk-vi|| (2)
And is
Figure BDA0002187776990000052
Wherein x isk∈RP,RPIs P-dimensional space, n is the number of data, c is more than or equal to 2<n,U=[μik]I is 1,2, L, c, k is 1,2, L, n is fuzzy classification matrix, muikRepresents a sample xkDegree of membership belonging to class i, and μikSatisfies the formula (3), V ═ V1,v2,L,vc) For the cluster center matrix, m ∈ [1, ∞) ] is the fuzzy weighting index, usually taking the values 1.1. ltoreq. m.ltoreq.2.5, dikThen the sample x is representedkTo the ith cluster center viEuclidean distance of.
The fuzzy C-means clustering algorithm minimizes the objective function to obtain the optimal clustering result. To obtain the extreme values of the objective function, the values are measured in μikAnd viThe derivation is performed, and when the constraint condition shown in equation (3) is satisfied, it is obtained:
Figure BDA0002187776990000053
Figure BDA0002187776990000054
the specific implementation process of the fuzzy C-means clustering algorithm can be summarized as follows:
Step1firstly, setting parameters of a fuzzy C-means clustering algorithm: the cluster category number c, the fuzzy weighting index m, the iteration stop threshold value and the maximum iteration number, and initializing a cluster center V(0)And let b equal to 0.
Step2Sequentially calculating a membership matrix U according to a formula (4) and a formula (5)(b)And a cluster center V(b+1)
Step3If V | |(b+1)-V(b)If | is less, stopping the operation; otherwise, let b be b +1, go to Step2
Fig. 2 is a flowchart illustrating an automatic confidence rule generation method based on fuzzy clustering according to the present invention, and the specific implementation steps thereof can be summarized as follows:
Step1fuzzy clustering is carried out on a large amount of historical sample data by using a fuzzy C-means clustering algorithm to obtain a fuzzy clustering center V1.
Step2Re-clustering sample data with elements less than 0.01 in the membership degree matrix to obtain a fuzzy clustering center V2.
Step3Repeat Step2And taking the obtained fuzzy clustering center data (V1 ∪ V2 ∪ …) as basic data for constructing an extended confidence rule base, and determining the number and utility values of the reference levels of each premise attribute and conclusion attribute.
Step4Converting input-output pairs in the underlying data to expanded as shown in equation (6)And the confidence distribution representation form of the premise attribute and the conclusion attribute of the confidence rule base.
Figure RE-GDA0002422948340000061
Wherein, TkIs the number of premise attributes in the kth rule, JiThe number of reference levels for the ith precondition attribute, N is the number of output result levels, θkRule weight of the kth rule;iis the attribute weight;
Figure BDA0002187776990000062
a utility value representing the jth reference level of the ith precondition attribute in the kth rule, where i is 1,2, L, Tk,j=1,2,L,Ji;JiThe number of reference levels representing the ith precondition attribute;
Figure BDA0002187776990000063
the confidence of the corresponding reference level is indicated, where i ═ 1,2, L, Tk,j=1,2,L,Ji;DnN is 1, 2L, N is the evaluation level of the output result, βn,kIndicates the evaluation result D in the k-th rulenThe confidence of (c).
For input value x in input-output pairiIts precondition attribute is UiAccording to the formula (8), x can be expressediIs converted into UiAs shown in equation (7):
S(xi)={(Aijij);j=1,2,L,Ji} (7)
wherein the content of the first and second substances,
Figure BDA0002187776990000064
wherein, for the ith precondition attribute UiThe number of reference levels is JiAmong these different reference levels, the effective value of the jth reference level is set to Aij。αijFor a given valuexiConversion to prerequisite Attribute UiBy examining the value xiThe degree of similarity to the corresponding reference level utility value is obtained.
Similarly, for the output value y in an input-output pairiIts conclusion attribute is ViAccording to the formula (10), y can be expressediConversion to ViThe confidence distribution form of (2) is shown as formula (9):
S(yi)={(Dijij);j=1,2,L,N} (9)
wherein the content of the first and second substances,
Figure BDA0002187776990000071
wherein N represents the number of evaluation levels of conclusion attributes; dijThe utility value of the jth evaluation level in the N evaluation levels representing the ith conclusion attribute βijRepresents a value yiTransition to conclusion Attribute ViThe confidence level at the jth evaluation level of (a) is also obtained by comparing the utility value of the corresponding evaluation level with the value yiThe degree of similarity of (a).
Step5Continuously executing Step4Each input-output pair will generate an extended confidence rule until all underlying data has generated the corresponding confidence rule.
Step6Finally, the rule weight and the premise attribute weight of each rule are calculated.
And (3) calculating the weight of the attribute of the precondition based on the incidence relation of the variables, namely calculating the Pearson correlation coefficient of each precondition attribute and each conclusion attribute respectively by using historical sample data, and normalizing the correlation coefficients to obtain the weight of the attribute of the precondition.
Calculating rule weight based on sample membership matrix, firstly calculating initial rule weight thetap. Assuming that the pth rule is generated by the pth clustering center, the ratio of the number of sample data belonging to the pth class to the total number of the whole sample data is the initial rule weight thetap. Then, according to the extended confidence ruleThe library method updates the rule weights according to the degree of inconsistency of the rules. The steps of updating the rule weights are as follows:
the first step is as follows: rule similarity is calculated.
First, the calculation formula (12) for similarity is obtained from the euclidean distance:
let X be { X ═ X1,x2,L,xn},Y={y1,y2,L,ynThe euclidean distance can be defined as follows:
Figure BDA0002187776990000072
then:
Sim(X,Y)=1-d(X,Y) (12)
rule similarities include the first piece of Similarity (SRA) and the second piece of Similarity (SRC). Assume that there are two rules with only one conclusion attribute:
Figure BDA0002187776990000073
Figure BDA0002187776990000074
wherein T represents the number of the premise attributes;
Figure BDA0002187776990000075
representing a confidence distribution of the t-th premise attribute in the ith rule; diRepresenting the confidence distribution of the conclusion attribute of the ith rule.
Then, the front piece similarity SRA and the back piece similarity SRC of the above two rules are defined as follows:
Figure BDA0002187776990000081
SRC(i,k)=Sim(Di,Dk) (16)
wherein, according toEquation (12) can calculate the similarity between the ith rule and the tth premise attribute of the kth rule in equation (15)
Figure BDA0002187776990000082
And similarity Sim (D) of conclusion attributes of the ith rule and the kth rule in equation (16)i,Dk)。
The second step is that: and calculating the rule inconsistency.
First, using SRA and SRC, the degree of agreement between two rules is calculated as follows:
Cons(Ri,Rk)=exp{-(SRA(i,k)/SRC(i,k)-1.0)2/(1/SRA(i,k))2} (17)
wherein, Cons (R)i,Rk)∈[0,1]。
Then, the inconsistency of the ith rule for the whole rule base is calculated:
Figure BDA0002187776990000083
wherein, L*For the number of generated rules, then, the overall degree of inconsistency of the rule base is calculated:
Figure BDA0002187776990000084
the third step: the rule weights are updated.
The formula for calculating the rule weight is as follows:
Figure BDA0002187776990000085
wherein the content of the first and second substances,
Figure BDA0002187776990000086
the constant λ is a function of the update weight of the pth rule, depending on the particular application.
According to the above 6 steps, a confidence rule base can be constructed.
The method provided by the invention is compared with performance of a local parameter Optimization post-confidence Rule library model (L OBRB, &lTtTtranslation = L "&gTtL &lTt/T &gTttemporal Optimization Belief Base) and a Global parameter Optimization post-confidence Rule library model (GOBRB, Global Optimization Belief Base) to obtain a result shown in table 2.
The invention designs an automatic confidence rule extraction method based on fuzzy clustering, which can effectively solve the problem that sample data is difficult to select in an extended confidence rule base method, and designs a calculation method based on rule weight and precondition attribute weight of data, thereby reducing the dependence on expert experience knowledge. The method is beneficial to improving the generation efficiency and the reasoning precision of the confidence rule base, and lays a good foundation for realizing knowledge automation and control intelligence of the industrial process.

Claims (1)

1. A confidence rule automatic generation method based on fuzzy clustering is characterized in that a fuzzy C-means clustering algorithm is used for carrying out fuzzy clustering on a large amount of historical sample data to obtain a fuzzy clustering center V1, the sample data with elements smaller than 0.01 in a membership degree matrix are clustered again to obtain a fuzzy clustering center V2, the sample data with elements smaller than 0.01 in the membership degree matrix are continuously clustered again until the elements in the membership degree matrix are not smaller than 0.01, and therefore final fuzzy clustering center data (V1 ∪ V2 ∪ …) is obtained;
the obtained fuzzy clustering center data (V1 ∪ V2 ∪ …) is used as basic data for constructing an extended confidence rule base, and input-output pairs in the basic data are converted into confidence distribution expression forms of the premise attributes and conclusion attributes of the extended confidence rule base, wherein the expression is shown as a formula (1):
Figure RE-FDA0002422948330000011
with a regular weight θkand attribute weighti
Wherein R iskDenotes the kth rule, TkIs the number of premise attributes in the kth rule, J1The number of reference levels for the 1 st prerequisite attribute,
Figure RE-FDA0002422948330000012
is the T thkThe number of levels of each precondition attribute, N is the number of output result levels, and thetakRule weight of the kth rule;iattribute weight of the ith attribute;
Figure RE-FDA0002422948330000013
a utility value representing the jth reference level of the ith precondition attribute in the kth rule, where i is 1,2, …, Tk,j=1,2,…,Ji;JiThe number of reference levels representing the ith precondition attribute;
Figure RE-FDA0002422948330000014
the confidence of the corresponding reference level is indicated, where i is 1,2, …, Tk,j=1,2,…,Ji;DnN is 1,2 …, N is the evaluation level of the output result, βn,kIndicates the evaluation result D in the k-th rulenThe confidence of (2);
calculating a precondition attribute weight based on the incidence relation of the variables, namely calculating the Pearson correlation coefficient of each precondition attribute and the conclusion attribute respectively by using historical sample data, and performing normalization processing to obtain the precondition attribute weight;
then, rule weights are calculated based on the sample membership matrix, and initial rule weights theta are calculated firstlyp(ii) a Assuming that the pth rule is generated by the pth clustering center, the ratio of the number of sample data belonging to the pth class to the total number of the whole sample data is the initial rule weight thetap(ii) a Then, according to the method of the extended confidence rule base, updating the rule weight according to the inconsistency of the rule; the steps of updating the rule weights are as follows:
the first step is as follows: calculating rule similarity;
first, a calculation formula (3) of similarity is obtained from the euclidean distance:
let X be { X ═ X1,x2,…,xn},Y={y1,y2,…,yn}, the Euclidean distance of which is defined as formula (2):
Figure RE-FDA0002422948330000015
then:
Sim(X,Y)=1-d(X,Y) (3)
the rule similarity comprises a front piece similarity SRA and a back piece similarity SRC; two rules with only one conclusion attribute are represented by equations (4) and (5):
Figure RE-FDA0002422948330000016
Figure RE-FDA0002422948330000017
wherein, UTRepresenting the Tth precondition attribute, wherein T represents the number of the precondition attributes;
Figure RE-FDA0002422948330000021
confidence score representing the t-th premise attribute in the ith ruleCloth; diA confidence distribution representing an attribute of the conclusion of the ith rule;
the pre-piece similarity SRA and the post-piece similarity SRC of the two rules of equations (4) and (5) are defined as equation (6) and (7), respectively:
Figure RE-FDA0002422948330000022
SRC(i,k)=Sim(Di,Dk) (7)
wherein, the similarity between the ith rule in the formula (6) and the tth precondition attribute of the kth rule is calculated according to the formula (3)
Figure RE-FDA0002422948330000023
And similarity Sim (D) of conclusion attributes of the ith rule and the kth rule in equation (7)i,Dk);
The second step is that: calculating the rule inconsistency;
first, using SRA and SRC, the degree of agreement between two rules is calculated as follows:
Cons(Ri,Rk)=exp{-[SRA(i,k)/SRC(i,k)-1.0]2/[1/SRA(i,k)]2} (8)
wherein, Cons (R)i,Rk)∈[0,1];
Then, the inconsistency of the ith rule for the whole rule base is calculated:
Figure RE-FDA0002422948330000024
wherein, L*For the number of generated rules, then, the overall degree of inconsistency of the rule base is calculated:
Figure RE-FDA0002422948330000025
the third step: updating the rule weight;
the formula for calculating the rule weight is as follows:
Figure RE-FDA0002422948330000026
wherein the content of the first and second substances,
Figure RE-FDA0002422948330000027
the updating weight of the p rule is obtained, and the value range of the constant lambda is (0, 1);
according to the steps, a confidence rule base is directly constructed from the data, and the automatic generation of the confidence rules is realized.
CN201910821869.5A 2019-09-02 2019-09-02 Confidence rule automatic generation method based on fuzzy clustering Pending CN111507365A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910821869.5A CN111507365A (en) 2019-09-02 2019-09-02 Confidence rule automatic generation method based on fuzzy clustering

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910821869.5A CN111507365A (en) 2019-09-02 2019-09-02 Confidence rule automatic generation method based on fuzzy clustering

Publications (1)

Publication Number Publication Date
CN111507365A true CN111507365A (en) 2020-08-07

Family

ID=71874022

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910821869.5A Pending CN111507365A (en) 2019-09-02 2019-09-02 Confidence rule automatic generation method based on fuzzy clustering

Country Status (1)

Country Link
CN (1) CN111507365A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112330093A (en) * 2020-09-30 2021-02-05 合肥工业大学 Intelligent early warning method and system oriented to cooperation of adaptive scheduling and unmanned production line
CN113162498A (en) * 2021-04-27 2021-07-23 谷芯(广州)技术有限公司 Permanent magnet synchronous motor vector control method and system based on fuzzy PI control
CN113379182A (en) * 2021-04-27 2021-09-10 云南电网有限责任公司昆明供电局 Middle and low voltage equipment health state assessment method based on multi-dimensional state parameters
CN114239738A (en) * 2021-12-21 2022-03-25 中国人民解放军国防科技大学 Medical data classification method for small sample and related equipment

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6148099A (en) * 1997-07-03 2000-11-14 Neopath, Inc. Method and apparatus for incremental concurrent learning in automatic semiconductor wafer and liquid crystal display defect classification
CN102740340A (en) * 2012-07-19 2012-10-17 中南大学 Data gathering method oriented to same configuration nodes in network cluster of wireless sensor
CN104820708A (en) * 2015-05-15 2015-08-05 成都睿峰科技有限公司 Cloud computing platform based big data clustering method and device
US20150286862A1 (en) * 2014-04-07 2015-10-08 Basware Corporation Method for Statistically Aided Decision Making
CN109655393A (en) * 2018-12-12 2019-04-19 昆明理工大学 A kind of air permeability of tipping paper detection method based on confidence rule base
CN110132603A (en) * 2019-05-16 2019-08-16 杭州电子科技大学 Boat diesel engine Fault Locating Method based on union confidence rule base and ant group algorithm

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6148099A (en) * 1997-07-03 2000-11-14 Neopath, Inc. Method and apparatus for incremental concurrent learning in automatic semiconductor wafer and liquid crystal display defect classification
CN102740340A (en) * 2012-07-19 2012-10-17 中南大学 Data gathering method oriented to same configuration nodes in network cluster of wireless sensor
US20150286862A1 (en) * 2014-04-07 2015-10-08 Basware Corporation Method for Statistically Aided Decision Making
CN104820708A (en) * 2015-05-15 2015-08-05 成都睿峰科技有限公司 Cloud computing platform based big data clustering method and device
CN109655393A (en) * 2018-12-12 2019-04-19 昆明理工大学 A kind of air permeability of tipping paper detection method based on confidence rule base
CN110132603A (en) * 2019-05-16 2019-08-16 杭州电子科技大学 Boat diesel engine Fault Locating Method based on union confidence rule base and ant group algorithm

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
LONG-HAOYANG 等: "A disjunctive belief rule-based expert system for bridge risk assessment with dynamic parameter optimization model", 《COMPUTERS & INDUSTRIAL ENGINEERING》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112330093A (en) * 2020-09-30 2021-02-05 合肥工业大学 Intelligent early warning method and system oriented to cooperation of adaptive scheduling and unmanned production line
CN112330093B (en) * 2020-09-30 2022-09-02 合肥工业大学 Intelligent early warning method and system for cooperation of adaptive scheduling and unmanned production line
CN113162498A (en) * 2021-04-27 2021-07-23 谷芯(广州)技术有限公司 Permanent magnet synchronous motor vector control method and system based on fuzzy PI control
CN113379182A (en) * 2021-04-27 2021-09-10 云南电网有限责任公司昆明供电局 Middle and low voltage equipment health state assessment method based on multi-dimensional state parameters
CN113162498B (en) * 2021-04-27 2021-12-21 谷芯(广州)技术有限公司 Permanent magnet synchronous motor vector control method and system based on fuzzy PI control
CN113379182B (en) * 2021-04-27 2022-09-16 云南电网有限责任公司昆明供电局 Middle and low voltage equipment health state assessment method based on multi-dimensional state parameters
CN114239738A (en) * 2021-12-21 2022-03-25 中国人民解放军国防科技大学 Medical data classification method for small sample and related equipment
CN114239738B (en) * 2021-12-21 2023-10-24 中国人民解放军国防科技大学 Medical data classification method and related equipment for small samples

Similar Documents

Publication Publication Date Title
CN111507365A (en) Confidence rule automatic generation method based on fuzzy clustering
Khayyam et al. A novel hybrid machine learning algorithm for limited and big data modeling with application in industry 4.0
Lin et al. Parameter tuning, feature selection and weight assignment of features for case-based reasoning by artificial immune system
CN109472088B (en) Shale gas-conditioned production well production pressure dynamic prediction method
CN114678080B (en) Converter end point phosphorus content prediction model, construction method and phosphorus content prediction method
CN110555459A (en) Score prediction method based on fuzzy clustering and support vector regression
Kumar et al. Advanced prediction of performance of a student in an university using machine learning techniques
Pei et al. Genetic algorithms for classification and feature extraction
Rabuñal et al. A new approach to the extraction of ANN rules and to their generalization capacity through GP
Tembusai et al. K-nearest neighbor with k-fold cross validation and analytic hierarchy process on data classification
CN114239400A (en) Multi-working-condition process self-adaptive soft measurement modeling method based on local double-weighted probability hidden variable regression model
Chandiok et al. Machine learning techniques for cognitive decision making
Hariri et al. Tipburn disorder detection in strawberry leaves using convolutional neural networks and particle swarm optimization
CN109214500B (en) Transformer fault identification method based on hybrid intelligent algorithm
Sun et al. A fuzzy brain emotional learning classifier design and application in medical diagnosis
CN117521063A (en) Malicious software detection method and device based on residual neural network and combined with transfer learning
Abdelkhalek et al. WE-AIRS: a new weighted evidential artificial immune recognition system
CN110766144B (en) Scalar coupling constant prediction system between atoms based on multi-layer decomposition fuzzy neural network
CN112836431A (en) Penicillin fermentation process fault prediction method based on PSO-LSTM
Jabbari et al. Obtaining accurate probabilistic causal inference by post-processing calibration
Chen Brain Tumor Prediction with LSTM Method
CN113743652B (en) Sugarcane squeezing process prediction method based on depth feature recognition
Listiana et al. Optimization of support vector machine using information gain and adaboost to improve accuracy of chronic kidney disease diagnosis
Wang et al. Data-driven Extraction Method of Belief Rule for Reagent Addition in Antimony Rougher Flotation
CN116226629B (en) Multi-model feature selection method and system based on feature contribution

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200807