CN115018006A - Dempster-Shafer framework-based classification method - Google Patents
Dempster-Shafer framework-based classification method Download PDFInfo
- Publication number
- CN115018006A CN115018006A CN202210776766.3A CN202210776766A CN115018006A CN 115018006 A CN115018006 A CN 115018006A CN 202210776766 A CN202210776766 A CN 202210776766A CN 115018006 A CN115018006 A CN 115018006A
- Authority
- CN
- China
- Prior art keywords
- follows
- framework
- bjs
- probability
- divergence
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 56
- 238000012549 training Methods 0.000 claims abstract description 27
- 238000012360 testing method Methods 0.000 claims abstract description 20
- 238000004364 calculation method Methods 0.000 claims abstract description 7
- 230000007246 mechanism Effects 0.000 claims abstract description 7
- 238000010276 construction Methods 0.000 claims abstract description 4
- 102100032249 Dystonin Human genes 0.000 claims description 37
- 230000006870 function Effects 0.000 claims description 27
- 230000008569 process Effects 0.000 claims description 21
- 239000002131 composite material Substances 0.000 claims description 9
- 238000005315 distribution function Methods 0.000 claims description 9
- 230000008901 benefit Effects 0.000 claims description 7
- 238000004422 calculation algorithm Methods 0.000 claims description 4
- 230000010354 integration Effects 0.000 claims description 4
- 238000011160 research Methods 0.000 claims description 4
- 238000013459 approach Methods 0.000 claims description 3
- 238000001514 detection method Methods 0.000 claims description 3
- 238000010606 normalization Methods 0.000 claims description 3
- 238000012545 processing Methods 0.000 claims description 3
- 230000007704 transition Effects 0.000 claims description 3
- 230000004927 fusion Effects 0.000 abstract description 6
- 230000008859 change Effects 0.000 abstract description 2
- 230000035945 sensitivity Effects 0.000 abstract description 2
- 238000010801 machine learning Methods 0.000 description 4
- 238000012706 support-vector machine Methods 0.000 description 3
- 238000007635 classification algorithm Methods 0.000 description 2
- 230000004075 alteration Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000008094 contradictory effect Effects 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 238000003066 decision tree Methods 0.000 description 1
- 239000005433 ionosphere Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000007500 overflow downdraw method Methods 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000002194 synthesizing effect Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Landscapes
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention discloses a classification method based on Dempster-Shafer framework, which comprises a discrimination framework, a quality function, a combination rule of evidence theory, posterior probability, fuzzy set theory, BJS difference between two BBAs, single case forms of PPT and PPT, an identification attribute framework, member calculation, single generation probability, training sample distribution, test sample distribution, BJS divergence discrimination, group discrimination probability, a weighting integral mechanism and BPA construction, the accuracy of the method is statistically superior to that of 11 methods, the decision result is more reliable and steady, the model is effective and reasonable, in addition, the method introducing the BJS divergence has high sensitivity to data change, thereby providing convenience for practical application, predicting that WFIG-DSF is particularly prominent in data fusion application, paying attention to data from different sensor sources, and considering the difference between the test sample and the training sample distribution, in order to make better decisions.
Description
Technical Field
The invention belongs to the technical field of weighted fuzzy individual generation and group discrimination classification correlation based on a Dempster-Shafer framework, and particularly relates to a classification method based on the Dempster-Shafer framework.
Background
In D-S theory, an identification framework is defined as a complete set of incompatible basic propositions, a subset of each framework called a proposition. The D-S theory is a combination rule of multi-source information, and has a function of synthesizing basic results of a plurality of sensors as an output.
Disclosure of Invention
The invention aims to provide a classification method based on a Dempster-Shafer framework.
In order to achieve the purpose, the invention provides the following technical scheme:
a classification method based on a Dempster-Shafer framework comprises a discrimination framework, a quality function, a combination rule of an evidence theory, a posterior probability, a fuzzy set theory, a BJS difference between two BBAs, single case forms of PPT and PPT, an identification attribute framework, member calculation, a single generation probability, training sample distribution, distribution of test samples, BJS divergence discrimination, a group discrimination probability, a weighted integral mechanism and BPA construction.
Preferably, the discrimination framework flow is as follows:
let Θ denote a mutually exclusive set, which may contain an infinite number of elements, e.g., Θ ═ θ 1 ,θ 2 ,…,θ j ,…,θ N In which θ j Is an element or event identifying the frame Θ, N is the number of elements, j is 1,2, …, N, the set of all subsets of Θ is called the quality function, and is defined as follows
Representing an empty set; { theta ] i ,θ j Indicates theta i Or theta j Event occurrence, for a system, we can use 2 no matter it is in any state Θ Represents it by an element of (1);
preferably, the quality function flow is as follows:
let m be from set 2 Θ Under theta [0, 1]]The focus element A represents any subset of the identification framework Θ, and m is defined as
The basic probability distribution function represents the initial distribution of trust established by the evidence, the basic probability distribution function of the event a is recorded as m (A) and is used for representing the trust degree of the evidence on the A, and m (A) is also commonly called a quality function and is constructed according to data obtained by detection or is given by people according to experience;
preferably, the flow of the combination rule of the evidence theory is as follows:
let E 1 And E 2 Is to identify two pieces of evidence under the theta framework, whose corresponding BPAs are m 1 And m 2 。A i And B i Is the focus element. The combination rule of Desmpter is defined as
The K value represents the degree of conflict between evidence sources, in most cases, for different evidences, two or more different BPAs are obtained due to different data sources, and at this time, in order to calculate the trust function and the likelihood function, two or more BPAs need to be synthesized into one BPA;
preferably, the posterior probability process is as follows:
let X i (i ═ 1, 2.., p) is a p-independent feature, and X denotes a p-dimensional feature vector. Y is formed by C ═ C 1 ,C 2 ,...,C N Is the same sort flag as the X state. Then posterior probability P (Y | X) i ),i=1…p,Y∈{C 1 ,C 2 ,...,C N It can be defined as follows:
in this study, we constructed models using both individual generative models and population generative models. This approach provides an operable way to balance individual differences with overall differences, and fuzzy set theory is used as a criterion to describe membership.
Preferably, the fuzzy set theory process is as follows:
is provided with C i (i ═ 1, 2.., n) is a class, and the basic probability of DST is defined as:
FST is an extension of the exact set. Compared with the exact set, the FST provides a standard for judging the attribution and non-attribution concepts, compared with the standard Bayes theory, the FST provides another flexible uncertainty modeling framework for processing uncertainty and complexity in practical application, because real-world samples are fuzzy, the factors can not be modeled by classical probability theory sometimes, so that the definition is not strict, and the FST has the using advantage,
the BJS divergence theory is the popularization of the KL divergence theory, mainly measures the similarity of two probabilities, solves the problem of asymmetric result in the KL theory, has good symmetry and boundary of the BJS divergence, and is more suitable for being applied in the evidence theory, although the gradient of a point is 0 when two distribution distances are far away, most of the problems encountered in the research are about the decision of high similarity, so that the use of the BJS divergence has certain advantages;
preferably, the BJS difference flow between the two BBAs is as follows:
suppose A i Is one of the elements of the recognition framework m, assuming that there are two BBAsm under the same recognition framework ω 1 And m 2 ,m 1 And m 2 The BJS deviation in between is defined as:
After the transition, BJS can be represented as
Wherein H (m) j )=-∑ i m j (A i )logm j (A i )(i=1,2,...,M;j=1,2)
The BJS divergence is similar to the JS divergence in form, but the BJS divergence replaces a probability distribution function with a quality function, which makes the BBA probability distribution when all assumptions of the confidence function are assigned to a single element, at which time the BJS divergence degenerates to the JS divergence;
preferably, the process of identifying the attribute framework is as follows:
let C be an identification framework comprising N mutually exclusive hypotheses Θ ═ C 1 ,C 2 ,...,C N }, distribution value of information sources 2 Θ Reference number C considering combination element i ,C j The identification frame for which i ≠ j is not greater than 2 is represented as follows:
Ω={{C 1 },...,{C N },{C 1 ,C 2 },...,{C i ,C j },…,{C N-1 ,C N }}.
since some properties are similar, the representations on the gaussian profiles overlap, and complex assumptions are therefore proposed to represent such cases.
Preferably, the member calculation process is as follows:
is provided withFor different classes of membership for each object x,for the variance of the class, ε is the sample mean, and the membership is calculated as follows:
dividing the generation probability calculated according to the membership degrees into the generation probability of single membership degree and the generation probability of composite membership degree, wherein the generation probability of single membership degree is represented by the membership degree of the category, and the generation probability of composite membership degree is calculated by the minimum value of t combination of two types of membership degree norms;
preferably, the single generation probability process is as follows:
is provided withAs a member of the combination, the sample to be tested is recorded asFor generating probabilitiesExpressed, as follows:
preferably, the training sample distribution process is as follows:
there are n sets of training samples, n being a multiple of m, epsilon P (C i ) And σ P (C i ) C representing the jth group of samples in turn i Sample mean and variance of attributes, training set distributionIs defined as
Wherein,
the distribution characteristics of the training set samples conform to Gaussian distribution, and assuming that the kth sample in a database needs to be tested, a test set is constructed by using m elements according to a time sequence modelThe text classification standard is classified according to each group m and is defined as
Preferably, the distribution flow of the test sample is as follows:
let the test set contain n elements, ε Q ({C i }),σ Q ({C i }) of samples in turn, which determines C for the j sample sets i Quality, and training set distributionIs defined as
The distribution characteristics of the training set samples conform to Gaussian distribution, and the method uses the BJS divergence to measure the difference between the training set and the test set;
preferably, the BJS divergence determination process is as follows:
the single attribute BJS divergence discrimination is expressed as:
the multi-attribute BJS divergence discrimination is defined as follows:
wherein,
preferably, the population discrimination probability process is as follows:
is provided withRepresenting group discrimination probability under single attribute, for corresponding group discrimination probability of composite attributeIs represented as follows:
preferably, the weighted integration mechanism flow is as follows:
let gamma denote the weighted integral result, and the heuristic algorithm is expressed by alpha and beta factors, defined as follows
Note that the learning factors (α, β) are different for different attribute classes;
preferably, the configuration BPA scheme is as follows:
let M k For the final BPA, this is also a weighted normalized expression, as follows:
wherein,
note that since Γ itself is not equal to 1, it is not an explicit BPA, and therefore, the normalization factor Γ was proposed to assist Γ in the last generation of BPAs.
Compared with the prior art, the invention provides a classification method based on a Dempster-Shafer framework, which has the following beneficial effects:
the WFIG-DSF provided by the invention utilizes the BJS divergence theory and the data characteristics to determine the attributes of the evidence and capture the uncertainty among classes, and the application of the electroencephalogram data not only controls the distribution characteristics of the whole data set, but also focuses on the characteristics of individual representative data, so that the method effectively measures the uncertainty among the evidences and reduces the harmful conflict among classifiers. The method can well capture high-conflict information, maintain important complementarity among classifiers and improve the fusion performance of the classifiers, in addition, the WFIG-DSF is evaluated based on 12 data sets in a UCI machine learning library, and compared with other existing classification methods, finally, a conclusion is drawn, the method has wide application value, the accuracy of the method is statistically superior to the performance of 11 methods through the accuracy test of UCI data, the decision result is more reliable and steady, the model is effective and reasonable, in addition, the method introducing the BJS divergence has high sensitivity to data change, convenience is provided for practical application, the WFIG-DSF is predicted to be particularly prominent in data fusion application, the method pays attention to data from different sensor sources, and the difference between the test sample and the distribution of the training sample is considered, in order to make better decisions.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The invention provides a technical scheme that:
a classification method based on a Dempster-Shafer framework comprises a discrimination framework, a quality function, a combination rule of an evidence theory, a posterior probability, a fuzzy set theory, a BJS difference between two BBAs, single case forms of PPT and PPT, an identification attribute framework, member calculation, a single generation probability, training sample distribution, distribution of test samples, BJS divergence discrimination, a group discrimination probability, a weighted integral mechanism and BPA construction.
The discrimination framework flow is as follows:
let Θ denote a mutually exclusive set, which may contain an infinite number of elements, e.g., Θ ═ θ 1 ,θ 2 ,…,θ j ,…,θ N In which θ j Is an element or event identifying the frame Θ, N is the number of elements, j is 1,2, …, N, the set of all subsets of Θ is called the quality function, and is defined as follows
Representing an empty set; { theta ] i ,θ j Indicates θ i Or theta j Event occurrence, for a system, we can use 2 no matter it is in any state Θ Represents it by an element of (1);
the quality function flow is as follows:
let m be from set 2 Θ At theta [0, 1]]The focus element A represents any subset of the identification framework Θ, and m is defined as
The basic probability distribution function represents the initial distribution of trust established by the evidence, the basic probability distribution function of the event a is recorded as m (A) and is used for representing the trust degree of the evidence on the A, and m (A) is also commonly called a quality function and is constructed according to data obtained by detection or is given by people according to experience;
the flow of the combination rule of evidence theory is as follows:
let E 1 And E 2 Is to identify two pieces of evidence under the theta framework, whose corresponding BPAs are m 1 And m 2 。A i And B i Is the focus element. The combination rule of Desmpter is defined as
The K value represents the degree of conflict between evidence sources, in most cases, for different evidences, two or more different BPA are obtained due to different data sources, and at the moment, in order to calculate the trust function and the likelihood function, the two or more BPA need to be synthesized into one BPA;
the posterior probability flow is as follows:
let X i (i ═ 1, 2.., p) is a p-independent feature, and X denotes a p-dimensional feature vector. Y is formed by C ═ C 1 ,C 2 ,...,C N Is the same sort flag as the X state. Then posterior probability P (Y | X) i ),i=1…p,Y∈{C 1 ,C 2 ,...,C N The following can be defined:
in this study, we constructed models using both individual generative models and population generative models. This approach provides an operable way to balance individual differences with overall differences, and fuzzy set theory is used as a criterion to describe membership.
The fuzzy set theory process is as follows:
is provided with C i (i ═ 1, 2.., n) is a class, and the basic probability of DST is defined as:
FST is an extension of the exact set. Compared with the exact set, the FST provides a standard for judging the attribution and non-attribution concepts, compared with the standard Bayes theory, the FST provides another flexible uncertainty modeling framework for processing uncertainty and complexity in practical application, because real-world samples are fuzzy, the factors can not be modeled by classical probability theory sometimes, so that the definition is not strict, and the FST has the using advantage,
the BJS divergence theory is the popularization of the KL divergence theory, mainly measures the similarity of two probabilities, solves the problem of asymmetric result in the KL theory, has good symmetry and boundary of the BJS divergence, and is more suitable for being applied in the evidence theory, although the gradient of a point is 0 when two distribution distances are far away, most of the problems encountered in the research are about the decision of high similarity, so that the use of the BJS divergence has certain advantages;
the BJS difference flow between the two BBAs is as follows:
suppose A i Is one of the elements of the recognition framework m, assuming that there are two BBAsm under the same recognition framework ω 1 And m 2 ,m 1 And m 2 The BJS deviation in between is defined as:
After the transition, BJS can be represented as
Wherein H (m) j )=-∑ i m j (A i )logm j (A i )(i=1,2,...,M;j=1,2)
The BJS divergence is similar to the JS divergence in form, but the BJS divergence replaces a probability distribution function with a quality function, which makes the BBA probability distribution when all assumptions of the confidence function are assigned to a single element, at which time the BJS divergence degenerates to the JS divergence;
the process of identifying the attribute framework is as follows:
let C be an identification framework comprising N mutually exclusive hypotheses Θ ═ C 1 ,C 2 ,...,C N }, distribution value of information sources 2 Θ Reference number C considering combination element i ,C j The identification frame for which i ≠ j is not greater than 2 is represented as follows:
Ω={{C 1 },...,{C N },{C 1 ,C 2 },...,{C i ,C j },…,{C N-1 ,C N }}.
since some properties are similar, the representations on the gaussian profiles overlap, and complex assumptions are therefore proposed to represent such cases.
The member calculation process is as follows:
is provided withFor different classes of membership for each object x,for the variance of the class, ε is the sample mean, and the membership is calculated as follows:
dividing the generation probability calculated according to the membership degrees into the generation probability of single membership degree and the generation probability of composite membership degree, wherein the generation probability of single membership degree is represented by the membership degree of the category, and the generation probability of composite membership degree is calculated by the minimum value of t combination of two types of membership degree norms;
the single generation probability flow is as follows:
is provided withAs a member of the combination, the sample to be tested is recorded asFor generating probabilitiesExpressed, as follows:
the training sample distribution process is as follows:
there are n sets of training samples, n being a multiple of m, epsilon P (C i ) And σ P (C i ) C representing the jth group of samples in turn i Sample mean and variance of attributes, training set distributionIs defined as
Wherein,
the distribution characteristics of the training set samples conform to Gaussian distribution, and assuming that the kth sample in a database needs to be tested, a test set is constructed by using m elements according to a time sequence modelThe text classification standard is classified according to each group m and is defined as
The distribution flow of the test samples is as follows:
let the test set contain n elements, ε Q ({C i }),σ Q ({C i }) represents the mean and variance of the samples in turn, which determines C for the j sample sets i Quality, and training set distributionIs defined as
The distribution characteristics of the training set samples conform to Gaussian distribution, and the method uses the BJS divergence to measure the difference between the training set and the test set;
the BJS divergence judging process comprises the following steps:
the single attribute BJS divergence discrimination is expressed as:
the multi-attribute BJS divergence discrimination is defined as follows:
wherein,
the population discrimination probability flow is as follows:
is provided withRepresenting group discrimination probability under single attribute, for corresponding group discrimination probability of composite attributeIs represented as follows:
the weighted integration mechanism flow is as follows:
let gamma denote the weighted integral result, and the heuristic algorithm is expressed by alpha and beta factors, defined as follows
Note that the learning factors (α, β) are different for different attribute classes;
the BPA build flow is as follows:
let M k For the final BPA, this is also a weighted normalized expression, as follows:
wherein,
note that since Γ itself is not equal to 1, it is not an explicit BPA, the normalization factor Γ was proposed to assist Γ in the last generation of BPAs;
combining BPA according to D-S rules
In combining the BPA, we combine the BPA for each individual information source generated according to equation [ eq14] using D-S rules, resulting in an overall BPA.
Probability transformation
After the total BPA is obtained, the total BPA is converted to individual decision probabilities according to equation [ eq11 ].
Terminal decision making
And obtaining a final decision result through PPT, and outputting the most possible sample category.
The UCI data set (https:// area-beta. ics. UCI. edu) is a classic machine learning test data set, and is suitable for pattern recognition and machine learning research. We used 12 kinds of data in UCI dataset, i.e. Isir, Heart, distance, Wine, Australian, Climate, Heapatitis, Waveform, Parkinsons, Forest, Ionosphere, Spambase and Sonar, to compare with eight most advanced classifiers, which are Naive Bayes (NB), Nearest Mean Classifier (NMC), k nearest neighbor (k-NN), decision tree (REPTree), Support Vector Machine (SVM), support vector machine and radial basis function (SVM-RBF), multilayer perceptron (MLP), Radial Basis Function Network (RBFN), and four classification algorithms based on D-S theory to find that the classification accuracy of the WFIG-DSF is better than other methods. The classification algorithms based on the D-S theory are respectively a k-nearest neighbor D-S theory (kNNDST), a classifier based on Normal Distribution (NDBC), evidence correction (Evicalib) and a weighted fuzzy D-S framework (WFDSF).
In classifier fusion, the classification results of different classifiers tend to be highly conflicting, sometimes even completely contradictory, which is likely to result in unreasonable fusion results, i.e., errors. This fusion method handles conflict situations very well, which is a major advantage of this method, when the selected data sets contain missing information, in handling such situations, since the proposed algorithm contains a time series and a BJS divergence signature, these missing values can be fitted with the corresponding variables in the D-S theory, specifically, if the data contains a missing value of a certain data, the missing value is considered as an uncertain problem, and the confidence of the missing value is 0, i.e., m (c) is 1. Furthermore, this step results in an increase in the population probability and a decrease in the individual probability. In addition to this missing value, the remaining attributes will be collected periodically, in fact, in order to improve the classification result of the different classifiers closer to the accuracy. By properly fitting and modeling partial inaccuracies, the degree of conflict between classifiers can be reduced. By combining the conventional attributes of other classifiers, the relative reliability of each classifier is evaluated by utilizing the incompatibility with other classifiers, and finally, the classifier with higher conflict degree with other classifiers submits smaller relative reliability value.
As the process of information diversification increases, the problem of classification of information attributes is brought up to the agenda. The method using machine learning and classifiers is a method widely used in data fusion, provides a weighted fuzzy individual generation and group discrimination classification rule based on a Dempster-Shafer framework, takes BPA as a plurality of classifiers to support different attribute sets, constructs a new target classification identification method, introduces fuzzy Bayes and BJS divergence theories to aggregate and classify information sources, determines the probability of a data point entering a cluster according to the distance between the characteristic of the attribute and the class centroid, and calculates the fuzzy membership of the cluster for classification.
Although embodiments of the present invention have been shown and described, it will be appreciated by those skilled in the art that changes, modifications, substitutions and alterations can be made in these embodiments without departing from the principles and spirit of the invention, the scope of which is defined in the appended claims and their equivalents.
Claims (7)
1. A classification method based on a Dempster-Shafer framework is characterized by comprising the following steps: the distinguishing and classifying method comprises the steps of distinguishing a frame, a quality function, a combination rule of an evidence theory, a posterior probability, a fuzzy set theory, a BJS difference between two BBAs, PPT, single case forms of the PPT, an identification attribute frame, member calculation, single generation probability, training sample distribution, test sample distribution, BJS divergence judgment, group distinguishing probability, a weighting integral mechanism and BPA construction.
2. The Dempster-Shafer framework-based classification method according to claim 1, characterized in that: the discrimination framework flow is as follows:
let Θ denote a mutually exclusive set, which may contain an infinite number of elements, e.g., Θ ═ θ 1 ,θ 2 ,…,θ j ,…,θ N In which θ j Is an element or event identifying the framework Θ, N is the number of elements, j is 1,2, …, N, the set of all subsets of Θ is called the quality function, and is defined as follows
Representing an empty set; { theta ] i ,θ j Indicates θ i Or theta j Event occurrence, for a system, we can use 2 no matter it is in any state Θ Represents it by an element of (1);
the quality function flow is as follows:
let m be from set 2 Θ Under theta [0, 1]]The focus element A represents any subset of the identification framework Θ, and m is defined as
The basic probability distribution function represents the initial distribution of trust established by the evidence, the basic probability distribution function of the event a is recorded as m (A) and is used for representing the trust degree of the evidence on the A, and m (A) is also commonly called a quality function and is constructed according to data obtained by detection or is given by people according to experience;
the flow of the combination rule of the evidence theory is as follows:
let E 1 And E 2 Is to identify two pieces of evidence under the theta framework, whose corresponding BPAs are m 1 And m 2 。A i And B i Is the focus element. The combination rule of Desmpter is defined as
The K value represents the degree of conflict between evidence sources, in most cases, for different evidences, two or more different BPA are obtained due to different data sources, and at the moment, in order to calculate the trust function and the likelihood function, the two or more BPA need to be synthesized into one BPA;
the posterior probability flow is as follows:
let X i (i ═ 1,2, …, p) is a p-independent feature, and X represents a p-dimensional feature vector. Y is formed by C ═ C 1 ,C 2 ,…,C N Is the same sort flag as the X state. Then the posterior probability P (Y |. X) i ),i=1…p,Y∈{C 1 ,C 2 ,…,C N The following can be defined:
in this study, we constructed models using both individual generative models and population generative models. This approach provides an operable way to balance individual differences with overall differences, and fuzzy set theory is used as a criterion to describe membership.
3. The Dempster-Shafer framework-based classification method according to claim 1, characterized in that: the fuzzy set theory process is as follows:
let C i (i ═ 1,2, …, n) is a class, and the basic probability of DST is defined as:
FST is an extension of the exact set. Compared with the exact set, the FST provides a standard for judging the attribution and non-attribution concepts, compared with the standard Bayes theory, the FST provides another flexible uncertainty modeling framework for processing uncertainty and complexity in practical application, because real-world samples are fuzzy, the factors can not be modeled by classical probability theory sometimes, so that the definition is not strict, and the FST has the using advantage,
the BJS divergence theory is the popularization of the KL divergence theory, mainly measures the similarity of two probabilities, solves the problem of asymmetric result in the KL theory, has good symmetry and boundary of the BJS divergence, and is more suitable for being applied in the evidence theory, although the gradient of a point is 0 when two distribution distances are far away, most of the problems encountered in the research are about the decision of high similarity, so that the use of the BJS divergence has certain advantages;
the BJS difference flow between the two BBAs is as follows:
suppose A i Is one of the elements of the recognition framework m, assuming that there are two BBAsm under the same recognition framework ω 1 And m 2 ,m 1 And m 2 The BJS deviation in between is defined as:
After the transition, BJS can be represented as
Wherein H (m) j )=-∑ i m j (A i )logm j (A i )(i=1,2,…,M;j=1,2)
The BJS divergence is similar to the JS divergence in form, but the BJS divergence replaces a probability distribution function with a quality function, which makes the BBA probability distribution when all assumptions of the confidence function are assigned to a single element, at which time the BJS divergence degenerates to the JS divergence;
the process of identifying the attribute framework is as follows:
let C be an identification framework comprising N mutually exclusive hypotheses Θ ═ C 1 ,C 2 ,…,C N }, distribution value of information sources 2 Θ Reference number C considering combination element i ,C j An identification frame with i ≠ j no greater than 2 is expressed as follows:
Ω={{C 1 },…,{C N },{C 1 ,C 2 },…,{C i ,C j },…,{C N-1 ,C N }}.
since some properties are similar, the representations on the gaussian profiles overlap, and complex assumptions are therefore proposed to represent such cases.
4. The Dempster-Shafer framework-based classification method according to claim 1, characterized in that: the member calculation process is as follows:
is provided withFor different classes of membership for each object x,for the variance of the class, ε is the sample mean, and the membership is calculated as follows:
dividing the generation probability calculated according to the membership degrees into the generation probability of single membership degree and the generation probability of composite membership degree, wherein the generation probability of single membership degree is represented by the membership degree of the category, and the generation probability of composite membership degree is calculated by the minimum value of t combination of two types of membership degree norms;
the single generation probability process is as follows:
is provided withAs a member of the combination, the sample to be tested is recorded asFor generating probabilitiesExpressed, as follows:
the training sample distribution process is as follows:
there are n sets of training samples, n being a multiple of m, epsilon P (C i ) And σ P (C i ) C representing the jth group of samples in turn i Sample mean and variance of attributes, training set distributionIs defined as
Wherein,
the distribution characteristics of the training set samples conform to Gaussian distribution, and assuming that the kth sample in a database needs to be tested, a test set is constructed by using m elements according to a time sequence modelThe text classification standard is classified according to m of each group and is defined as
5. The Dempster-Shafer framework-based classification method according to claim 1, characterized in that: the distribution flow of the test samples is as follows:
let the test set contain n elements, ε Q ({C i }),σ Q ({C i }) of samples in turn, which determines C for the j sample sets i Quality, and training set distributionIs defined as
The distribution characteristics of the training set samples conform to Gaussian distribution, and the method uses the BJS divergence to measure the difference between the training set and the test set;
the BJS divergence judging process is as follows:
the single attribute BJS divergence discrimination is expressed as:
the multi-attribute BJS divergence discrimination is defined as follows:
wherein,
6. the Dempster-Shafer framework-based classification method according to claim 1, characterized in that: the population discrimination probability process is as follows:
is provided withRepresenting group discrimination probability under single attribute, for corresponding group discrimination probability of composite attributeIs represented as follows:
7. the Dempster-Shafer framework-based classification method according to claim 1, characterized in that: the weighted integration mechanism flow is as follows:
let 'r' denote the weighted integration result, and the heuristic algorithm is expressed by alpha and beta factors, defined as follows
Note that the learning factors (α, β) are different for different attribute classes;
the BPA configuration flow is as follows:
let M k For the final BPA, this is also a weighted normalized expression, as follows:
wherein,
note that since Γ itself is not equal to 1, it is not an explicit BPA, the normalization factor Γ is proposed to assist Γ in the last generation BPA.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210776766.3A CN115018006A (en) | 2022-07-04 | 2022-07-04 | Dempster-Shafer framework-based classification method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210776766.3A CN115018006A (en) | 2022-07-04 | 2022-07-04 | Dempster-Shafer framework-based classification method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN115018006A true CN115018006A (en) | 2022-09-06 |
Family
ID=83078225
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210776766.3A Pending CN115018006A (en) | 2022-07-04 | 2022-07-04 | Dempster-Shafer framework-based classification method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115018006A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117909507A (en) * | 2024-03-19 | 2024-04-19 | 金盾检测技术股份有限公司 | AI-based data classification system |
CN118587549A (en) * | 2024-08-05 | 2024-09-03 | 杭州电子科技大学 | Ship classification method based on HRRP and SAR data credible decision fusion |
-
2022
- 2022-07-04 CN CN202210776766.3A patent/CN115018006A/en active Pending
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117909507A (en) * | 2024-03-19 | 2024-04-19 | 金盾检测技术股份有限公司 | AI-based data classification system |
CN117909507B (en) * | 2024-03-19 | 2024-05-17 | 金盾检测技术股份有限公司 | AI-based data classification system |
CN118587549A (en) * | 2024-08-05 | 2024-09-03 | 杭州电子科技大学 | Ship classification method based on HRRP and SAR data credible decision fusion |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109086658B (en) | Sensor data generation method and system based on generation countermeasure network | |
US8676740B2 (en) | Attribute estimation system, age estimation system, gender estimation system, age and gender estimation system and attribute estimation method | |
US7620217B2 (en) | Three-dimensional face recognition system and method | |
US7529403B2 (en) | Weighted ensemble boosting method for classifier combination and feature selection | |
CN115018006A (en) | Dempster-Shafer framework-based classification method | |
Sathiyanarayanan et al. | Identification of breast cancer using the decision tree algorithm | |
CN108960043A (en) | A kind of personage's family relationship construction method for electron album management | |
CN111061700A (en) | Hospitalizing migration scheme recommendation method and system based on similarity learning | |
Parvin et al. | A heuristic scalable classifier ensemble of binary classifier ensembles | |
CN109919236A (en) | A kind of BP neural network multi-tag classification method based on label correlation | |
Wang et al. | Unleash the black magic in age: a multi-task deep neural network approach for cross-age face verification | |
Chelly et al. | Hybridization schemes of the fuzzy dendritic cell immune binary classifier based on different fuzzy clustering techniques | |
Arowolo et al. | A hybrid dimensionality reduction model for classification of microarray dataset | |
CN111144462B (en) | Unknown individual identification method and device for radar signals | |
CN116363712A (en) | Palmprint palm vein recognition method based on modal informativity evaluation strategy | |
Kumar et al. | Imbalanced classification in diabetics using ensembled machine learning | |
CN113555110B (en) | Method and equipment for training multi-disease referral model | |
Trajdos et al. | A correction method of a binary classifier applied to multi-label pairwise models | |
Mehmood et al. | Classifier ensemble optimization for gender classification using genetic algorithm | |
Raja et al. | A Novel Feature Selection based Ensemble Decision Tree Classification Model for Predicting Severity Level of COPD | |
CN115082762A (en) | Target detection unsupervised domain adaptation system based on regional recommendation network center alignment | |
CN111860441B (en) | Video target identification method based on unbiased depth migration learning | |
CN115049870A (en) | Target detection method based on small sample | |
Manokhin | Machine learning for probabilistic prediction | |
Dang et al. | Inclusive Consistency based Quantitative Decision-making Framework for Incremental Automatic Target Recognition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |