CN115018006A - Dempster-Shafer framework-based classification method - Google Patents

Dempster-Shafer framework-based classification method Download PDF

Info

Publication number
CN115018006A
CN115018006A CN202210776766.3A CN202210776766A CN115018006A CN 115018006 A CN115018006 A CN 115018006A CN 202210776766 A CN202210776766 A CN 202210776766A CN 115018006 A CN115018006 A CN 115018006A
Authority
CN
China
Prior art keywords
follows
framework
bjs
probability
divergence
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210776766.3A
Other languages
Chinese (zh)
Inventor
肖富元
何华平
何立蜓
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to CN202210776766.3A priority Critical patent/CN115018006A/en
Publication of CN115018006A publication Critical patent/CN115018006A/en
Pending legal-status Critical Current

Links

Landscapes

  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention discloses a classification method based on Dempster-Shafer framework, which comprises a discrimination framework, a quality function, a combination rule of evidence theory, posterior probability, fuzzy set theory, BJS difference between two BBAs, single case forms of PPT and PPT, an identification attribute framework, member calculation, single generation probability, training sample distribution, test sample distribution, BJS divergence discrimination, group discrimination probability, a weighting integral mechanism and BPA construction, the accuracy of the method is statistically superior to that of 11 methods, the decision result is more reliable and steady, the model is effective and reasonable, in addition, the method introducing the BJS divergence has high sensitivity to data change, thereby providing convenience for practical application, predicting that WFIG-DSF is particularly prominent in data fusion application, paying attention to data from different sensor sources, and considering the difference between the test sample and the training sample distribution, in order to make better decisions.

Description

Dempster-Shafer framework-based classification method
Technical Field
The invention belongs to the technical field of weighted fuzzy individual generation and group discrimination classification correlation based on a Dempster-Shafer framework, and particularly relates to a classification method based on the Dempster-Shafer framework.
Background
In D-S theory, an identification framework is defined as a complete set of incompatible basic propositions, a subset of each framework called a proposition. The D-S theory is a combination rule of multi-source information, and has a function of synthesizing basic results of a plurality of sensors as an output.
Disclosure of Invention
The invention aims to provide a classification method based on a Dempster-Shafer framework.
In order to achieve the purpose, the invention provides the following technical scheme:
a classification method based on a Dempster-Shafer framework comprises a discrimination framework, a quality function, a combination rule of an evidence theory, a posterior probability, a fuzzy set theory, a BJS difference between two BBAs, single case forms of PPT and PPT, an identification attribute framework, member calculation, a single generation probability, training sample distribution, distribution of test samples, BJS divergence discrimination, a group discrimination probability, a weighted integral mechanism and BPA construction.
Preferably, the discrimination framework flow is as follows:
let Θ denote a mutually exclusive set, which may contain an infinite number of elements, e.g., Θ ═ θ 1 ,θ 2 ,…,θ j ,…,θ N In which θ j Is an element or event identifying the frame Θ, N is the number of elements, j is 1,2, …, N, the set of all subsets of Θ is called the quality function, and is defined as follows
Figure BDA0003727703950000011
Figure BDA0003727703950000012
Representing an empty set; { theta ] i ,θ j Indicates theta i Or theta j Event occurrence, for a system, we can use 2 no matter it is in any state Θ Represents it by an element of (1);
preferably, the quality function flow is as follows:
let m be from set 2 Θ Under theta [0, 1]]The focus element A represents any subset of the identification framework Θ, and m is defined as
Figure BDA0003727703950000021
The basic probability distribution function represents the initial distribution of trust established by the evidence, the basic probability distribution function of the event a is recorded as m (A) and is used for representing the trust degree of the evidence on the A, and m (A) is also commonly called a quality function and is constructed according to data obtained by detection or is given by people according to experience;
preferably, the flow of the combination rule of the evidence theory is as follows:
let E 1 And E 2 Is to identify two pieces of evidence under the theta framework, whose corresponding BPAs are m 1 And m 2 。A i And B i Is the focus element. The combination rule of Desmpter is defined as
Figure BDA0003727703950000022
Wherein
Figure BDA0003727703950000023
The K value represents the degree of conflict between evidence sources, in most cases, for different evidences, two or more different BPAs are obtained due to different data sources, and at this time, in order to calculate the trust function and the likelihood function, two or more BPAs need to be synthesized into one BPA;
preferably, the posterior probability process is as follows:
let X i (i ═ 1, 2.., p) is a p-independent feature, and X denotes a p-dimensional feature vector. Y is formed by C ═ C 1 ,C 2 ,...,C N Is the same sort flag as the X state. Then posterior probability P (Y | X) i ),i=1…p,Y∈{C 1 ,C 2 ,...,C N It can be defined as follows:
Figure BDA0003727703950000031
in this study, we constructed models using both individual generative models and population generative models. This approach provides an operable way to balance individual differences with overall differences, and fuzzy set theory is used as a criterion to describe membership.
Preferably, the fuzzy set theory process is as follows:
is provided with C i (i ═ 1, 2.., n) is a class, and the basic probability of DST is defined as:
Figure BDA0003727703950000032
FST is an extension of the exact set. Compared with the exact set, the FST provides a standard for judging the attribution and non-attribution concepts, compared with the standard Bayes theory, the FST provides another flexible uncertainty modeling framework for processing uncertainty and complexity in practical application, because real-world samples are fuzzy, the factors can not be modeled by classical probability theory sometimes, so that the definition is not strict, and the FST has the using advantage,
the BJS divergence theory is the popularization of the KL divergence theory, mainly measures the similarity of two probabilities, solves the problem of asymmetric result in the KL theory, has good symmetry and boundary of the BJS divergence, and is more suitable for being applied in the evidence theory, although the gradient of a point is 0 when two distribution distances are far away, most of the problems encountered in the research are about the decision of high similarity, so that the use of the BJS divergence has certain advantages;
preferably, the BJS difference flow between the two BBAs is as follows:
suppose A i Is one of the elements of the recognition framework m, assuming that there are two BBAsm under the same recognition framework ω 1 And m 2 ,m 1 And m 2 The BJS deviation in between is defined as:
Figure BDA0003727703950000041
wherein
Figure BDA0003727703950000042
And S (m) 1 ,m 2 )=∑ i m 1 (A i )=1(i=1,2,...,M;j=1,2),∑ i m j (A i )=1(i=1,2,...,M;j=1,2),
After the transition, BJS can be represented as
Figure BDA0003727703950000043
Wherein H (m) j )=-∑ i m j (A i )logm j (A i )(i=1,2,...,M;j=1,2)
The BJS divergence is similar to the JS divergence in form, but the BJS divergence replaces a probability distribution function with a quality function, which makes the BBA probability distribution when all assumptions of the confidence function are assigned to a single element, at which time the BJS divergence degenerates to the JS divergence;
preferably, the process of identifying the attribute framework is as follows:
let C be an identification framework comprising N mutually exclusive hypotheses Θ ═ C 1 ,C 2 ,...,C N }, distribution value of information sources 2 Θ Reference number C considering combination element i ,C j The identification frame for which i ≠ j is not greater than 2 is represented as follows:
Ω={{C 1 },...,{C N },{C 1 ,C 2 },...,{C i ,C j },…,{C N-1 ,C N }}.
since some properties are similar, the representations on the gaussian profiles overlap, and complex assumptions are therefore proposed to represent such cases.
Preferably, the member calculation process is as follows:
is provided with
Figure BDA0003727703950000051
For different classes of membership for each object x,
Figure BDA0003727703950000052
for the variance of the class, ε is the sample mean, and the membership is calculated as follows:
Figure BDA0003727703950000053
dividing the generation probability calculated according to the membership degrees into the generation probability of single membership degree and the generation probability of composite membership degree, wherein the generation probability of single membership degree is represented by the membership degree of the category, and the generation probability of composite membership degree is calculated by the minimum value of t combination of two types of membership degree norms;
preferably, the single generation probability process is as follows:
is provided with
Figure BDA0003727703950000054
As a member of the combination, the sample to be tested is recorded as
Figure BDA0003727703950000055
For generating probabilities
Figure BDA0003727703950000056
Expressed, as follows:
Figure BDA0003727703950000057
preferably, the training sample distribution process is as follows:
there are n sets of training samples, n being a multiple of m, epsilon P (C i ) And σ P (C i ) C representing the jth group of samples in turn i Sample mean and variance of attributes, training set distribution
Figure BDA0003727703950000058
Is defined as
Figure BDA0003727703950000059
Wherein,
Figure BDA00037277039500000510
Figure BDA00037277039500000511
the distribution characteristics of the training set samples conform to Gaussian distribution, and assuming that the kth sample in a database needs to be tested, a test set is constructed by using m elements according to a time sequence model
Figure BDA0003727703950000061
The text classification standard is classified according to each group m and is defined as
Figure BDA0003727703950000062
Preferably, the distribution flow of the test sample is as follows:
let the test set contain n elements, ε Q ({C i }),σ Q ({C i }) of samples in turn, which determines C for the j sample sets i Quality, and training set distribution
Figure BDA0003727703950000063
Is defined as
Figure BDA0003727703950000064
The distribution characteristics of the training set samples conform to Gaussian distribution, and the method uses the BJS divergence to measure the difference between the training set and the test set;
preferably, the BJS divergence determination process is as follows:
the single attribute BJS divergence discrimination is expressed as:
Figure BDA0003727703950000065
the multi-attribute BJS divergence discrimination is defined as follows:
Figure BDA0003727703950000066
wherein,
Figure BDA0003727703950000067
preferably, the population discrimination probability process is as follows:
is provided with
Figure BDA0003727703950000068
Representing group discrimination probability under single attribute, for corresponding group discrimination probability of composite attribute
Figure BDA0003727703950000069
Is represented as follows:
Figure BDA0003727703950000071
Figure BDA0003727703950000072
preferably, the weighted integration mechanism flow is as follows:
let gamma denote the weighted integral result, and the heuristic algorithm is expressed by alpha and beta factors, defined as follows
Figure BDA0003727703950000073
Note that the learning factors (α, β) are different for different attribute classes;
preferably, the configuration BPA scheme is as follows:
let M k For the final BPA, this is also a weighted normalized expression, as follows:
Figure BDA0003727703950000074
wherein,
Figure BDA0003727703950000075
note that since Γ itself is not equal to 1, it is not an explicit BPA, and therefore, the normalization factor Γ was proposed to assist Γ in the last generation of BPAs.
Compared with the prior art, the invention provides a classification method based on a Dempster-Shafer framework, which has the following beneficial effects:
the WFIG-DSF provided by the invention utilizes the BJS divergence theory and the data characteristics to determine the attributes of the evidence and capture the uncertainty among classes, and the application of the electroencephalogram data not only controls the distribution characteristics of the whole data set, but also focuses on the characteristics of individual representative data, so that the method effectively measures the uncertainty among the evidences and reduces the harmful conflict among classifiers. The method can well capture high-conflict information, maintain important complementarity among classifiers and improve the fusion performance of the classifiers, in addition, the WFIG-DSF is evaluated based on 12 data sets in a UCI machine learning library, and compared with other existing classification methods, finally, a conclusion is drawn, the method has wide application value, the accuracy of the method is statistically superior to the performance of 11 methods through the accuracy test of UCI data, the decision result is more reliable and steady, the model is effective and reasonable, in addition, the method introducing the BJS divergence has high sensitivity to data change, convenience is provided for practical application, the WFIG-DSF is predicted to be particularly prominent in data fusion application, the method pays attention to data from different sensor sources, and the difference between the test sample and the distribution of the training sample is considered, in order to make better decisions.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The invention provides a technical scheme that:
a classification method based on a Dempster-Shafer framework comprises a discrimination framework, a quality function, a combination rule of an evidence theory, a posterior probability, a fuzzy set theory, a BJS difference between two BBAs, single case forms of PPT and PPT, an identification attribute framework, member calculation, a single generation probability, training sample distribution, distribution of test samples, BJS divergence discrimination, a group discrimination probability, a weighted integral mechanism and BPA construction.
The discrimination framework flow is as follows:
let Θ denote a mutually exclusive set, which may contain an infinite number of elements, e.g., Θ ═ θ 1 ,θ 2 ,…,θ j ,…,θ N In which θ j Is an element or event identifying the frame Θ, N is the number of elements, j is 1,2, …, N, the set of all subsets of Θ is called the quality function, and is defined as follows
Figure BDA0003727703950000081
Figure BDA0003727703950000082
Representing an empty set; { theta ] i ,θ j Indicates θ i Or theta j Event occurrence, for a system, we can use 2 no matter it is in any state Θ Represents it by an element of (1);
the quality function flow is as follows:
let m be from set 2 Θ At theta [0, 1]]The focus element A represents any subset of the identification framework Θ, and m is defined as
Figure BDA0003727703950000091
The basic probability distribution function represents the initial distribution of trust established by the evidence, the basic probability distribution function of the event a is recorded as m (A) and is used for representing the trust degree of the evidence on the A, and m (A) is also commonly called a quality function and is constructed according to data obtained by detection or is given by people according to experience;
the flow of the combination rule of evidence theory is as follows:
let E 1 And E 2 Is to identify two pieces of evidence under the theta framework, whose corresponding BPAs are m 1 And m 2 。A i And B i Is the focus element. The combination rule of Desmpter is defined as
Figure BDA0003727703950000092
Wherein
Figure BDA0003727703950000093
The K value represents the degree of conflict between evidence sources, in most cases, for different evidences, two or more different BPA are obtained due to different data sources, and at the moment, in order to calculate the trust function and the likelihood function, the two or more BPA need to be synthesized into one BPA;
the posterior probability flow is as follows:
let X i (i ═ 1, 2.., p) is a p-independent feature, and X denotes a p-dimensional feature vector. Y is formed by C ═ C 1 ,C 2 ,...,C N Is the same sort flag as the X state. Then posterior probability P (Y | X) i ),i=1…p,Y∈{C 1 ,C 2 ,...,C N The following can be defined:
Figure BDA0003727703950000101
in this study, we constructed models using both individual generative models and population generative models. This approach provides an operable way to balance individual differences with overall differences, and fuzzy set theory is used as a criterion to describe membership.
The fuzzy set theory process is as follows:
is provided with C i (i ═ 1, 2.., n) is a class, and the basic probability of DST is defined as:
Figure BDA0003727703950000102
FST is an extension of the exact set. Compared with the exact set, the FST provides a standard for judging the attribution and non-attribution concepts, compared with the standard Bayes theory, the FST provides another flexible uncertainty modeling framework for processing uncertainty and complexity in practical application, because real-world samples are fuzzy, the factors can not be modeled by classical probability theory sometimes, so that the definition is not strict, and the FST has the using advantage,
the BJS divergence theory is the popularization of the KL divergence theory, mainly measures the similarity of two probabilities, solves the problem of asymmetric result in the KL theory, has good symmetry and boundary of the BJS divergence, and is more suitable for being applied in the evidence theory, although the gradient of a point is 0 when two distribution distances are far away, most of the problems encountered in the research are about the decision of high similarity, so that the use of the BJS divergence has certain advantages;
the BJS difference flow between the two BBAs is as follows:
suppose A i Is one of the elements of the recognition framework m, assuming that there are two BBAsm under the same recognition framework ω 1 And m 2 ,m 1 And m 2 The BJS deviation in between is defined as:
Figure BDA0003727703950000111
wherein
Figure BDA0003727703950000112
And S (m) 1 ,m 2 )=∑ i m 1 (A i )=1(i=1,2,...,M;j=1,2),∑ i m j (A i )=1(i=1,2,...,M;j=1,2),
After the transition, BJS can be represented as
Figure BDA0003727703950000113
Wherein H (m) j )=-∑ i m j (A i )logm j (A i )(i=1,2,...,M;j=1,2)
The BJS divergence is similar to the JS divergence in form, but the BJS divergence replaces a probability distribution function with a quality function, which makes the BBA probability distribution when all assumptions of the confidence function are assigned to a single element, at which time the BJS divergence degenerates to the JS divergence;
the process of identifying the attribute framework is as follows:
let C be an identification framework comprising N mutually exclusive hypotheses Θ ═ C 1 ,C 2 ,...,C N }, distribution value of information sources 2 Θ Reference number C considering combination element i ,C j The identification frame for which i ≠ j is not greater than 2 is represented as follows:
Ω={{C 1 },...,{C N },{C 1 ,C 2 },...,{C i ,C j },…,{C N-1 ,C N }}.
since some properties are similar, the representations on the gaussian profiles overlap, and complex assumptions are therefore proposed to represent such cases.
The member calculation process is as follows:
is provided with
Figure BDA0003727703950000121
For different classes of membership for each object x,
Figure BDA0003727703950000122
for the variance of the class, ε is the sample mean, and the membership is calculated as follows:
Figure BDA0003727703950000123
dividing the generation probability calculated according to the membership degrees into the generation probability of single membership degree and the generation probability of composite membership degree, wherein the generation probability of single membership degree is represented by the membership degree of the category, and the generation probability of composite membership degree is calculated by the minimum value of t combination of two types of membership degree norms;
the single generation probability flow is as follows:
is provided with
Figure BDA0003727703950000124
As a member of the combination, the sample to be tested is recorded as
Figure BDA0003727703950000125
For generating probabilities
Figure BDA0003727703950000126
Expressed, as follows:
Figure BDA0003727703950000127
the training sample distribution process is as follows:
there are n sets of training samples, n being a multiple of m, epsilon P (C i ) And σ P (C i ) C representing the jth group of samples in turn i Sample mean and variance of attributes, training set distribution
Figure BDA0003727703950000128
Is defined as
Figure BDA0003727703950000129
Wherein,
Figure BDA00037277039500001210
Figure BDA00037277039500001211
the distribution characteristics of the training set samples conform to Gaussian distribution, and assuming that the kth sample in a database needs to be tested, a test set is constructed by using m elements according to a time sequence model
Figure BDA0003727703950000131
The text classification standard is classified according to each group m and is defined as
Figure BDA0003727703950000132
The distribution flow of the test samples is as follows:
let the test set contain n elements, ε Q ({C i }),σ Q ({C i }) represents the mean and variance of the samples in turn, which determines C for the j sample sets i Quality, and training set distribution
Figure BDA0003727703950000133
Is defined as
Figure BDA0003727703950000134
The distribution characteristics of the training set samples conform to Gaussian distribution, and the method uses the BJS divergence to measure the difference between the training set and the test set;
the BJS divergence judging process comprises the following steps:
the single attribute BJS divergence discrimination is expressed as:
Figure BDA0003727703950000135
the multi-attribute BJS divergence discrimination is defined as follows:
Figure BDA0003727703950000136
wherein,
Figure BDA0003727703950000137
the population discrimination probability flow is as follows:
is provided with
Figure BDA0003727703950000138
Representing group discrimination probability under single attribute, for corresponding group discrimination probability of composite attribute
Figure BDA0003727703950000139
Is represented as follows:
Figure BDA0003727703950000141
the weighted integration mechanism flow is as follows:
let gamma denote the weighted integral result, and the heuristic algorithm is expressed by alpha and beta factors, defined as follows
Figure BDA0003727703950000142
Note that the learning factors (α, β) are different for different attribute classes;
the BPA build flow is as follows:
let M k For the final BPA, this is also a weighted normalized expression, as follows:
Figure BDA0003727703950000143
wherein,
Figure BDA0003727703950000144
note that since Γ itself is not equal to 1, it is not an explicit BPA, the normalization factor Γ was proposed to assist Γ in the last generation of BPAs;
combining BPA according to D-S rules
In combining the BPA, we combine the BPA for each individual information source generated according to equation [ eq14] using D-S rules, resulting in an overall BPA.
Probability transformation
After the total BPA is obtained, the total BPA is converted to individual decision probabilities according to equation [ eq11 ].
Terminal decision making
And obtaining a final decision result through PPT, and outputting the most possible sample category.
The UCI data set (https:// area-beta. ics. UCI. edu) is a classic machine learning test data set, and is suitable for pattern recognition and machine learning research. We used 12 kinds of data in UCI dataset, i.e. Isir, Heart, distance, Wine, Australian, Climate, Heapatitis, Waveform, Parkinsons, Forest, Ionosphere, Spambase and Sonar, to compare with eight most advanced classifiers, which are Naive Bayes (NB), Nearest Mean Classifier (NMC), k nearest neighbor (k-NN), decision tree (REPTree), Support Vector Machine (SVM), support vector machine and radial basis function (SVM-RBF), multilayer perceptron (MLP), Radial Basis Function Network (RBFN), and four classification algorithms based on D-S theory to find that the classification accuracy of the WFIG-DSF is better than other methods. The classification algorithms based on the D-S theory are respectively a k-nearest neighbor D-S theory (kNNDST), a classifier based on Normal Distribution (NDBC), evidence correction (Evicalib) and a weighted fuzzy D-S framework (WFDSF).
In classifier fusion, the classification results of different classifiers tend to be highly conflicting, sometimes even completely contradictory, which is likely to result in unreasonable fusion results, i.e., errors. This fusion method handles conflict situations very well, which is a major advantage of this method, when the selected data sets contain missing information, in handling such situations, since the proposed algorithm contains a time series and a BJS divergence signature, these missing values can be fitted with the corresponding variables in the D-S theory, specifically, if the data contains a missing value of a certain data, the missing value is considered as an uncertain problem, and the confidence of the missing value is 0, i.e., m (c) is 1. Furthermore, this step results in an increase in the population probability and a decrease in the individual probability. In addition to this missing value, the remaining attributes will be collected periodically, in fact, in order to improve the classification result of the different classifiers closer to the accuracy. By properly fitting and modeling partial inaccuracies, the degree of conflict between classifiers can be reduced. By combining the conventional attributes of other classifiers, the relative reliability of each classifier is evaluated by utilizing the incompatibility with other classifiers, and finally, the classifier with higher conflict degree with other classifiers submits smaller relative reliability value.
As the process of information diversification increases, the problem of classification of information attributes is brought up to the agenda. The method using machine learning and classifiers is a method widely used in data fusion, provides a weighted fuzzy individual generation and group discrimination classification rule based on a Dempster-Shafer framework, takes BPA as a plurality of classifiers to support different attribute sets, constructs a new target classification identification method, introduces fuzzy Bayes and BJS divergence theories to aggregate and classify information sources, determines the probability of a data point entering a cluster according to the distance between the characteristic of the attribute and the class centroid, and calculates the fuzzy membership of the cluster for classification.
Although embodiments of the present invention have been shown and described, it will be appreciated by those skilled in the art that changes, modifications, substitutions and alterations can be made in these embodiments without departing from the principles and spirit of the invention, the scope of which is defined in the appended claims and their equivalents.

Claims (7)

1. A classification method based on a Dempster-Shafer framework is characterized by comprising the following steps: the distinguishing and classifying method comprises the steps of distinguishing a frame, a quality function, a combination rule of an evidence theory, a posterior probability, a fuzzy set theory, a BJS difference between two BBAs, PPT, single case forms of the PPT, an identification attribute frame, member calculation, single generation probability, training sample distribution, test sample distribution, BJS divergence judgment, group distinguishing probability, a weighting integral mechanism and BPA construction.
2. The Dempster-Shafer framework-based classification method according to claim 1, characterized in that: the discrimination framework flow is as follows:
let Θ denote a mutually exclusive set, which may contain an infinite number of elements, e.g., Θ ═ θ 12 ,…,θ j ,…,θ N In which θ j Is an element or event identifying the framework Θ, N is the number of elements, j is 1,2, …, N, the set of all subsets of Θ is called the quality function, and is defined as follows
Figure FDA0003727703940000011
Figure FDA0003727703940000012
Representing an empty set; { theta ] ij Indicates θ i Or theta j Event occurrence, for a system, we can use 2 no matter it is in any state Θ Represents it by an element of (1);
the quality function flow is as follows:
let m be from set 2 Θ Under theta [0, 1]]The focus element A represents any subset of the identification framework Θ, and m is defined as
Figure FDA0003727703940000013
The basic probability distribution function represents the initial distribution of trust established by the evidence, the basic probability distribution function of the event a is recorded as m (A) and is used for representing the trust degree of the evidence on the A, and m (A) is also commonly called a quality function and is constructed according to data obtained by detection or is given by people according to experience;
the flow of the combination rule of the evidence theory is as follows:
let E 1 And E 2 Is to identify two pieces of evidence under the theta framework, whose corresponding BPAs are m 1 And m 2 。A i And B i Is the focus element. The combination rule of Desmpter is defined as
Figure FDA0003727703940000021
Wherein
Figure FDA0003727703940000022
The K value represents the degree of conflict between evidence sources, in most cases, for different evidences, two or more different BPA are obtained due to different data sources, and at the moment, in order to calculate the trust function and the likelihood function, the two or more BPA need to be synthesized into one BPA;
the posterior probability flow is as follows:
let X i (i ═ 1,2, …, p) is a p-independent feature, and X represents a p-dimensional feature vector. Y is formed by C ═ C 1 ,C 2 ,…,C N Is the same sort flag as the X state. Then the posterior probability P (Y |. X) i ),i=1…p,Y∈{C 1 ,C 2 ,…,C N The following can be defined:
Figure FDA0003727703940000023
in this study, we constructed models using both individual generative models and population generative models. This approach provides an operable way to balance individual differences with overall differences, and fuzzy set theory is used as a criterion to describe membership.
3. The Dempster-Shafer framework-based classification method according to claim 1, characterized in that: the fuzzy set theory process is as follows:
let C i (i ═ 1,2, …, n) is a class, and the basic probability of DST is defined as:
Figure FDA0003727703940000031
FST is an extension of the exact set. Compared with the exact set, the FST provides a standard for judging the attribution and non-attribution concepts, compared with the standard Bayes theory, the FST provides another flexible uncertainty modeling framework for processing uncertainty and complexity in practical application, because real-world samples are fuzzy, the factors can not be modeled by classical probability theory sometimes, so that the definition is not strict, and the FST has the using advantage,
the BJS divergence theory is the popularization of the KL divergence theory, mainly measures the similarity of two probabilities, solves the problem of asymmetric result in the KL theory, has good symmetry and boundary of the BJS divergence, and is more suitable for being applied in the evidence theory, although the gradient of a point is 0 when two distribution distances are far away, most of the problems encountered in the research are about the decision of high similarity, so that the use of the BJS divergence has certain advantages;
the BJS difference flow between the two BBAs is as follows:
suppose A i Is one of the elements of the recognition framework m, assuming that there are two BBAsm under the same recognition framework ω 1 And m 2 ,m 1 And m 2 The BJS deviation in between is defined as:
Figure FDA0003727703940000032
wherein
Figure FDA0003727703940000033
And S (m) 1 ,m 2 )=∑ i m 1 (A i )=1(i=1,2,…,M;j=1,2),∑ i m j (A i )=1(i=1,2,…,M;j=1,2),
After the transition, BJS can be represented as
Figure FDA0003727703940000041
Wherein H (m) j )=-∑ i m j (A i )logm j (A i )(i=1,2,…,M;j=1,2)
The BJS divergence is similar to the JS divergence in form, but the BJS divergence replaces a probability distribution function with a quality function, which makes the BBA probability distribution when all assumptions of the confidence function are assigned to a single element, at which time the BJS divergence degenerates to the JS divergence;
the process of identifying the attribute framework is as follows:
let C be an identification framework comprising N mutually exclusive hypotheses Θ ═ C 1 ,C 2 ,…,C N }, distribution value of information sources 2 Θ Reference number C considering combination element i ,C j An identification frame with i ≠ j no greater than 2 is expressed as follows:
Ω={{C 1 },…,{C N },{C 1 ,C 2 },…,{C i ,C j },…,{C N-1 ,C N }}.
since some properties are similar, the representations on the gaussian profiles overlap, and complex assumptions are therefore proposed to represent such cases.
4. The Dempster-Shafer framework-based classification method according to claim 1, characterized in that: the member calculation process is as follows:
is provided with
Figure FDA0003727703940000042
For different classes of membership for each object x,
Figure FDA0003727703940000043
for the variance of the class, ε is the sample mean, and the membership is calculated as follows:
Figure FDA0003727703940000051
dividing the generation probability calculated according to the membership degrees into the generation probability of single membership degree and the generation probability of composite membership degree, wherein the generation probability of single membership degree is represented by the membership degree of the category, and the generation probability of composite membership degree is calculated by the minimum value of t combination of two types of membership degree norms;
the single generation probability process is as follows:
is provided with
Figure FDA0003727703940000052
As a member of the combination, the sample to be tested is recorded as
Figure FDA0003727703940000053
For generating probabilities
Figure FDA0003727703940000054
Expressed, as follows:
Figure FDA0003727703940000055
the training sample distribution process is as follows:
there are n sets of training samples, n being a multiple of m, epsilon P (C i ) And σ P (C i ) C representing the jth group of samples in turn i Sample mean and variance of attributes, training set distribution
Figure FDA0003727703940000056
Is defined as
Figure FDA0003727703940000057
Wherein,
Figure FDA0003727703940000058
Figure FDA0003727703940000059
the distribution characteristics of the training set samples conform to Gaussian distribution, and assuming that the kth sample in a database needs to be tested, a test set is constructed by using m elements according to a time sequence model
Figure FDA00037277039400000510
The text classification standard is classified according to m of each group and is defined as
Figure FDA0003727703940000061
Figure FDA0003727703940000062
5. The Dempster-Shafer framework-based classification method according to claim 1, characterized in that: the distribution flow of the test samples is as follows:
let the test set contain n elements, ε Q ({C i }),σ Q ({C i }) of samples in turn, which determines C for the j sample sets i Quality, and training set distribution
Figure FDA0003727703940000063
Is defined as
Figure FDA0003727703940000064
The distribution characteristics of the training set samples conform to Gaussian distribution, and the method uses the BJS divergence to measure the difference between the training set and the test set;
the BJS divergence judging process is as follows:
the single attribute BJS divergence discrimination is expressed as:
Figure FDA0003727703940000065
the multi-attribute BJS divergence discrimination is defined as follows:
Figure FDA0003727703940000066
wherein,
Figure FDA0003727703940000067
6. the Dempster-Shafer framework-based classification method according to claim 1, characterized in that: the population discrimination probability process is as follows:
is provided with
Figure FDA0003727703940000076
Representing group discrimination probability under single attribute, for corresponding group discrimination probability of composite attribute
Figure FDA0003727703940000077
Is represented as follows:
Figure FDA0003727703940000071
Figure FDA0003727703940000072
7. the Dempster-Shafer framework-based classification method according to claim 1, characterized in that: the weighted integration mechanism flow is as follows:
let 'r' denote the weighted integration result, and the heuristic algorithm is expressed by alpha and beta factors, defined as follows
Figure FDA0003727703940000073
Note that the learning factors (α, β) are different for different attribute classes;
the BPA configuration flow is as follows:
let M k For the final BPA, this is also a weighted normalized expression, as follows:
Figure FDA0003727703940000074
wherein,
Figure FDA0003727703940000075
note that since Γ itself is not equal to 1, it is not an explicit BPA, the normalization factor Γ is proposed to assist Γ in the last generation BPA.
CN202210776766.3A 2022-07-04 2022-07-04 Dempster-Shafer framework-based classification method Pending CN115018006A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210776766.3A CN115018006A (en) 2022-07-04 2022-07-04 Dempster-Shafer framework-based classification method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210776766.3A CN115018006A (en) 2022-07-04 2022-07-04 Dempster-Shafer framework-based classification method

Publications (1)

Publication Number Publication Date
CN115018006A true CN115018006A (en) 2022-09-06

Family

ID=83078225

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210776766.3A Pending CN115018006A (en) 2022-07-04 2022-07-04 Dempster-Shafer framework-based classification method

Country Status (1)

Country Link
CN (1) CN115018006A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117909507A (en) * 2024-03-19 2024-04-19 金盾检测技术股份有限公司 AI-based data classification system
CN118587549A (en) * 2024-08-05 2024-09-03 杭州电子科技大学 Ship classification method based on HRRP and SAR data credible decision fusion

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117909507A (en) * 2024-03-19 2024-04-19 金盾检测技术股份有限公司 AI-based data classification system
CN117909507B (en) * 2024-03-19 2024-05-17 金盾检测技术股份有限公司 AI-based data classification system
CN118587549A (en) * 2024-08-05 2024-09-03 杭州电子科技大学 Ship classification method based on HRRP and SAR data credible decision fusion

Similar Documents

Publication Publication Date Title
CN109086658B (en) Sensor data generation method and system based on generation countermeasure network
US8676740B2 (en) Attribute estimation system, age estimation system, gender estimation system, age and gender estimation system and attribute estimation method
US7620217B2 (en) Three-dimensional face recognition system and method
US7529403B2 (en) Weighted ensemble boosting method for classifier combination and feature selection
CN115018006A (en) Dempster-Shafer framework-based classification method
Sathiyanarayanan et al. Identification of breast cancer using the decision tree algorithm
CN108960043A (en) A kind of personage's family relationship construction method for electron album management
CN111061700A (en) Hospitalizing migration scheme recommendation method and system based on similarity learning
Parvin et al. A heuristic scalable classifier ensemble of binary classifier ensembles
CN109919236A (en) A kind of BP neural network multi-tag classification method based on label correlation
Wang et al. Unleash the black magic in age: a multi-task deep neural network approach for cross-age face verification
Chelly et al. Hybridization schemes of the fuzzy dendritic cell immune binary classifier based on different fuzzy clustering techniques
Arowolo et al. A hybrid dimensionality reduction model for classification of microarray dataset
CN111144462B (en) Unknown individual identification method and device for radar signals
CN116363712A (en) Palmprint palm vein recognition method based on modal informativity evaluation strategy
Kumar et al. Imbalanced classification in diabetics using ensembled machine learning
CN113555110B (en) Method and equipment for training multi-disease referral model
Trajdos et al. A correction method of a binary classifier applied to multi-label pairwise models
Mehmood et al. Classifier ensemble optimization for gender classification using genetic algorithm
Raja et al. A Novel Feature Selection based Ensemble Decision Tree Classification Model for Predicting Severity Level of COPD
CN115082762A (en) Target detection unsupervised domain adaptation system based on regional recommendation network center alignment
CN111860441B (en) Video target identification method based on unbiased depth migration learning
CN115049870A (en) Target detection method based on small sample
Manokhin Machine learning for probabilistic prediction
Dang et al. Inclusive Consistency based Quantitative Decision-making Framework for Incremental Automatic Target Recognition

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination