CN113222045B - Semi-supervised fault classification method based on weighted feature alignment self-encoder - Google Patents
Semi-supervised fault classification method based on weighted feature alignment self-encoder Download PDFInfo
- Publication number
- CN113222045B CN113222045B CN202110575307.4A CN202110575307A CN113222045B CN 113222045 B CN113222045 B CN 113222045B CN 202110575307 A CN202110575307 A CN 202110575307A CN 113222045 B CN113222045 B CN 113222045B
- Authority
- CN
- China
- Prior art keywords
- encoder
- unlabeled
- labeled
- sample
- self
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 45
- 238000012549 training Methods 0.000 claims abstract description 47
- 238000013145 classification model Methods 0.000 claims abstract description 25
- 238000009826 distribution Methods 0.000 claims description 28
- 238000004519 manufacturing process Methods 0.000 claims description 18
- 239000013598 vector Substances 0.000 claims description 12
- 238000013528 artificial neural network Methods 0.000 claims description 11
- 238000004422 calculation algorithm Methods 0.000 claims description 10
- 238000000605 extraction Methods 0.000 claims description 10
- 230000008569 process Effects 0.000 claims description 10
- 238000004364 calculation method Methods 0.000 claims description 9
- 230000002159 abnormal effect Effects 0.000 claims description 7
- 239000010754 BS 2869 Class F Substances 0.000 claims description 3
- 239000013585 weight reducing agent Substances 0.000 claims description 3
- 238000005065 mining Methods 0.000 abstract description 3
- 230000006870 function Effects 0.000 description 13
- 230000008859 change Effects 0.000 description 5
- 238000003745 diagnosis Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 2
- 239000000498 cooling water Substances 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 238000003889 chemical engineering Methods 0.000 description 1
- 238000001311 chemical methods and process Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 239000003292 glue Substances 0.000 description 1
- 238000009776 industrial production Methods 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 239000007788 liquid Substances 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000003908 quality control method Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000008439 repair process Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000000926 separation method Methods 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/23—Clustering techniques
- G06F18/232—Non-hierarchical techniques
- G06F18/2321—Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Computing Systems (AREA)
- Molecular Biology (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- General Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Health & Medical Sciences (AREA)
- Probability & Statistics with Applications (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a semi-supervised fault classification method based on a weighted feature alignment self-encoder. Then, the weight of the unlabeled sample is calculated according to the probability density function of the error reconstructed by the training data. Further, a semi-supervised classification model based on the weighted feature alignment self-encoder is constructed by utilizing the labeled sample set, the unlabeled sample set and the corresponding weights. The weighted feature alignment self-encoder classification model designs a cross entropy training loss function based on weighted Sinkhorn distance, and the function enables the model to use labeled data and unlabeled data at the fine tuning stage, so that not only can deep mining of data information be realized, but also the generalization capability of a network model can be improved. Meanwhile, due to the introduction of a weighting strategy, the robustness of the model is obviously improved.
Description
Technical Field
The invention belongs to the field of industrial process control, and particularly relates to a semi-supervised fault classification method based on a weighted feature alignment self-encoder.
Background
Modern industrial processes are moving towards large scale, complex processes. How to ensure the safety of the production process is one of key problems which are focused on and need to be solved in the field of industrial process control. The fault diagnosis is a key technology for guaranteeing the safe operation of the industrial process, and has important significance for improving the product quality and the production efficiency. The fault classification belongs to a link in fault diagnosis, and automatic identification and judgment of fault types are realized by learning from historical fault information, so that production personnel are helped to quickly locate and repair the faults, and further loss caused by the faults is avoided. With the continuous development and progress of modern measurement means, a great deal of data is accumulated in the industrial production process. The data describes the actual conditions of each production stage of the manufacturing, provides valuable data resources for reading, analyzing and optimizing the manufacturing process, and is an intelligent source for realizing intelligent manufacturing. Therefore, how to reasonably utilize the data information accumulated in the manufacturing process to establish a data-driven intelligent analysis model to better serve the intelligent decision and quality control of the manufacturing process is a hot point of great concern in the industry. The data-driven fault classification method utilizes intelligent analysis technologies such as machine learning and deep learning to deeply mine, model and analyze industrial data and provide a data-driven fault diagnosis mode for users and industries. Most of the existing data-driven fault classification methods belong to supervised learning methods, and when sufficient labeled data can be obtained, the model can obtain excellent performance. However, it is difficult to obtain large, sufficient tagged data in certain industrial scenarios. Thus, there is often a large amount of unlabeled data and a small amount of labeled data. In order to effectively utilize the unlabeled data to improve the classification performance of the model, a fault classification method based on semi-supervised learning is gradually receiving attention. However, most existing semi-supervised fault classification methods mostly rely on certain data assumptions, such as semi-supervised learning methods based on statistical learning, semi-supervised learning methods based on graphs, and other methods for labeling unlabeled data based on cooperative training, self-training, etc., which all rely on one assumption, namely: the labeled and unlabeled swatches belong to the same distribution. However, this assumption has its limitation, data collected by an industrial process often include a large amount of noise and abnormal points, and may drift working conditions, labeled data is often manually screened and labeled by experts in the process field, while unlabeled samples are not screened, so that there is a high possibility that abnormal data different from the labeled data may occur in the unlabeled data. When the distribution of the non-labeled data is inconsistent with that of the labeled data, the performance of the semi-supervised algorithm is reduced, even lower than that of the supervised algorithm which only uses the labeled data for training. Therefore, it is desirable to provide a robust semi-supervised learning method, so that the model can still accurately implement fault classification when the labeled data and the unlabeled data have inconsistent distribution.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provides a semi-supervised fault classification method based on a weighted feature alignment self-encoder, which comprises the following steps:
a semi-supervised fault classification method based on weighted feature alignment self-encoder includes the following steps:
the method comprises the following steps: collecting normal working condition data and various fault data of an industrial process to obtain a training data set for modeling: sample set with labelsAnd unlabeled sample setWherein x represents an input sample, y represents a sample label, m represents the number of labeled samples, and n represents the number of unlabeled samples;
step two: constructing a stacking self-encoder model for reconstruction, and training the stacking self-encoder model by using a labeled sample set;
step three: estimating the probability density distribution of the reconstruction error of the training data, calculating the weight of the label-free sample, and further constructing a weighted feature alignment self-encoder classification model;
step four: and acquiring field working data, inputting the weighting characteristics, aligning the self-encoder classification model, and outputting a corresponding fault category.
Further, the second step is specifically divided into the following sub-steps:
(2.1) constructing a stacked self-encoder model for reconstruction, comprising a multi-layer encoder and a decoder, wherein the output of the model is the reconstruction of the input, and the calculation formula is as follows:
wherein x represents the input, zkRepresenting the extracted k-th layer features, k representing the k-th layer of the stacked self-encoder,andrepresenting the weight vector and the disparity vector of the encoder and decoder respectively,reconstruction of the input by the representative model;
(2.2) training the stacked self-encoder model by adopting the labeled samples constructed in the step one and adopting a random gradient descent algorithm, wherein a model training loss function is defined as a reconstruction error of an input, and the reconstruction error is represented by the following formula:
wherein,representing the ith labeled input sample,representing the reconstruction of the stacked auto-encoder;
(2.3) calculating the reconstruction error of the labeled sample by using the trained stacked self-encoder modelWherein the reconstruction error of a single sample is calculated with reference to the following formula:
further, the third step is specifically divided into the following sub-steps:
(3.1) calculating the reconstruction error E of the labeled exemplarslCompliance chi2Distribution ofDistribution parameters g and h of
g·h=mean(El) (5)
2g2·h=variance(El) (6)
(3.2) calculating reconstruction error of unlabeled exemplarThe reconstruction error calculation formula of a single sample is the same as the formula (4);
(3.3) calculating the reconstruction error E of the unlabeled exemplarsuIn distribution ElProbability of occurrence ofTo PuNormalizing to obtain the weight of the unlabeled sample
And (3.4) constructing a weighted feature alignment self-encoder classification model, and training the weighted feature alignment self-encoder classification model by adopting a labeled sample set, an unlabeled sample set and corresponding weights. The training process comprises the following steps: unsupervised pre-training and supervised fine tuning. In the unsupervised pre-training phase, labeled samples and unlabeled samples are used together to train a stacked self-encoder. The unsupervised pre-training method is the same as the steps (2.1) - (2.3). The supervised fine tuning is formed by adding a fully-connected neural network layer on a stacked self-encoder obtained by unsupervised pre-training and using the fully-connected neural network layer as output of categories, so as to obtain deep extraction features and category labels of the labeled samples and deep extraction features and predicted category label output of the unlabeled samples, and a specific calculation formula is as follows:
wherein,represents the deep-extracted features of the ith labeled sample,class labels representing the predicted ith labeled exemplar, { wc,bcRepresenting weight vectors and deviation vectors of the fully connected neural network layer;represents a deep extraction feature of the unlabeled exemplar,a class label output representing a prediction;
(3.7) assuming the number of classes as F, obtaining deep extraction features of labeled exemplars and unlabeled exemplars corresponding to each class F e FAndand weight of unlabeled exemplars
(3.8) calculating a training loss function of the weighted feature alignment self-encoder classification model using the following formula:
wherein, crossentropy represents a cross entropy loss function,the representative weighted Sinkhorn distance function is used for measuring the distance between the characteristic distribution of the labeled data and the characteristic distribution of the unlabeled data belonging to the same category, and meanwhile, the weight reduction of the abnormal unlabeled sample with larger reconstruction error is realized; alpha is the weight of the Sinkhorn distance,l being a network parameter2Regularization penalty term, β is its weight, pijRepresenting features of labeled exemplars i corresponding to category fFeatures to unlabeled sample jTransition probability of dijRepresenting features of labeled exemplars i corresponding to class fFeatures to unlabeled sample jThe distance of (a) to (b),represents the weight of the unlabeled exemplar j corresponding to the class f, and mf and nf represent the number of labeled and unlabeled exemplars corresponding to the class f, respectively.
The invention has the following beneficial effects:
the invention provides a robust semi-supervised fault classification method based on a weighted feature alignment self-encoder, aiming at the problem of performance degradation of a traditional semi-supervised classification model when labeled data and unlabelled data are not distributed uniformly. The method designs a model training loss function based on a weighting and feature alignment strategy. The introduction of the weighting strategy improves the robustness of the semi-supervised classification model and reduces the problem of performance reduction of the classification model caused by inconsistent sample distribution. And the introduction of the characteristic alignment strategy enables the model to use the labeled data and the unlabeled data at the same time in the fine tuning stage, so that the deep mining of data information can be realized, and the generalization capability and classification performance of the network model can be improved.
Drawings
FIG. 1 is a schematic diagram of a stacked self-encoder;
FIG. 2 is a TE process flow diagram;
FIG. 3 is a schematic diagram of data log reconstruction errors;
FIG. 4 is a graph illustrating classification accuracy of different algorithms.
Detailed Description
The present invention will be described in detail below with reference to the accompanying drawings and preferred embodiments, and the objects and effects of the present invention will become more apparent, it being understood that the specific embodiments described herein are merely illustrative of the present invention and are not intended to limit the present invention.
The semi-supervised fault classification method based on the weighted feature alignment self-encoder comprises the steps of firstly using labeled data to carry out reconstruction pre-training on a stacked self-encoder, and estimating probability density distribution of reconstruction errors. Then, the weight of the label-free sample is calculated according to the probability density function of the error reconstructed by the training data. Further, a semi-supervised classification model based on the weighted feature alignment self-encoder is constructed by utilizing the labeled sample set, the unlabeled sample set and the corresponding weights. The weighted feature alignment self-encoder classification model designs a cross entropy training loss function based on weighted Sinkhorn distance, and the function enables the model to use labeled data and unlabeled data at the fine tuning stage, so that not only can deep mining of data information be realized, but also the generalization capability of a network model can be improved. Meanwhile, due to the introduction of a weighting strategy, the robustness of the model is obviously improved.
The method comprises the following specific steps:
the method comprises the following steps: collecting normal working condition data and various fault data of an industrial process to obtain a training data set for modeling: labeled sample setAnd unlabeled exemplar setWherein x represents an input sample, y represents a sample label, m represents the number of labeled samples, and n represents the number of unlabeled samples;
step two: constructing a stacking self-encoder model for reconstruction, and training the stacking self-encoder model by utilizing a labeled sample set; the method is specifically divided into the following substeps:
(2.1) constructing a stacked self-encoder model for reconstruction, comprising a multi-layer encoder and a decoder, wherein the output of the model is the reconstruction of the input, and the calculation formula is as follows:
wherein x represents the input, zkRepresenting the extracted k-th layer features, k representing the k-th layer of the stacked self-encoder,andrepresenting the weight vector and the disparity vector of the encoder and decoder respectively,reconstruction of the representative model from the input;
(2.2) training the stacked self-encoder model by adopting the labeled sample set constructed in the step one and adopting a random gradient descent algorithm, wherein a model training loss function is defined as a reconstruction error of an input, and the reconstruction error is represented by the following formula:
wherein,represents the ith labeled input sample,representing the reconstruction of the stacked auto-encoder;
(2.3) calculating the reconstruction error of the labeled sample by using the trained stacked self-encoder modelWherein the reconstruction error of a single sample is calculated with reference to the following formula:
step three: estimating the probability density distribution of the reconstruction error of the training data, calculating the weight of the label-free sample, and further constructing a weighted feature alignment self-encoder classification model;
the third step is specifically divided into the following substeps:
(3.1) calculating the reconstruction error E of the labeled exemplarlCompliance chi2Distribution ofDistribution parameters g and h of
g·h=mean(El) (5)
2g2·h=variance(El) (6)
(3.2) calculating reconstruction error of unlabeled exemplarThe calculation formula of the reconstruction error of a single sample is the same as the formula (4);
(3.3) calculating the reconstruction error E of the unlabeled exemplarsuIn distribution ElProbability of occurrence ofTo PuNormalizing to obtain the weight of the unlabeled sample
And (3.4) constructing a weighted feature alignment self-encoder classification model, and training the weighted feature alignment self-encoder classification model by adopting a labeled sample set, an unlabeled sample set and corresponding weights. The training process can be divided into: unsupervised pre-training and supervised fine tuning.
In the unsupervised pre-training phase, labeled samples and unlabeled samples are used together to train a stacked self-encoder. The unsupervised pre-training method is the same as the steps (2.1) - (2.3), namely, a stacking self-encoder model for reconstruction is constructed firstly, and then the stacking self-encoder is trained by using the labeled sample and the unlabeled sample;
the supervised fine tuning is formed by adding a fully-connected neural network layer on a stacked self-encoder obtained by unsupervised pre-training and using the fully-connected neural network layer as output of categories, so as to obtain deep extraction features and category labels of the labeled samples and deep extraction features and predicted category label output of the unlabeled samples, and a specific calculation formula is as follows:
wherein,represents the deep-extracted features of the ith labeled sample,class labels representing the predicted ith labeled exemplar, { wc,bcRepresenting weight vectors and deviation vectors of the fully connected neural network layer;represents a deep-extracted feature of the unlabeled exemplar,a class label output representing a prediction;
(3.7) assuming that the number of classes is F, deep-layer extracted features of labeled samples and unlabeled samples corresponding to each class F e F are obtained according to the following formulaAndand weight of unlabeled exemplars
(3.8) calculating a training loss function of the weighted feature alignment self-encoder classification model by adopting the following formula:
wherein, crossentropy represents a cross entropy loss function,representing a weighted Sinkhorn distance function, alpha is the weight of the Sinkhorn distance,l being a network parameter2Regularization penalty term, β is its weight, pijRepresenting features of labeled exemplars i corresponding to class fFeatures to unlabeled sample jTransition probability of dijRepresenting features of labeled exemplars i corresponding to class fTo noneCharacteristics of Label sample jThe distance of (a) to (b),represents the weight of the unlabeled exemplar j corresponding to the class f, and mf and nf represent the number of labeled and unlabeled exemplars corresponding to the class f, respectively. The main purpose of the newly designed training loss function based on the weighted Sinkhorn distance is two. One is to align the labeled data and unlabeled data belonging to the same class in the fine tuning stage by stacking the features extracted from the encoder so that their distributions are close. And the other is that the weight reduction of the abnormal unlabeled sample with larger reconstruction error is realized through the weighted Sinkhorn characteristic distance with the unlabeled sample weight.
Step four: and acquiring field working data, inputting the weighted features to align the self-encoder classification model, and outputting corresponding fault categories.
The validity of the method of the invention is verified below with a specific industrial process example. All data are collected on a Tennessee-Eastman (TE) chemical engineering experiment simulation platform in the United states, and the platform is widely applied to the field of fault diagnosis and fault classification as a typical chemical process research object. The TE process is schematically shown in FIG. 2, and its main equipment includes a continuous stirred tank reactor, a gas-liquid separation column, a centrifugal compressor, a partial condenser and a reboiler. The modeled process data contained 16 process variables and 10 fault categories, and the detailed process variable and fault information descriptions are shown in tables 1 and 2, respectively.
TABLE 1
TABLE 2
Fault numbering | Description of the invention | Type of failure |
1 | A/C describes the feed flow ratio variation (stream 4) | Step change |
5 | Condenser cooling water inlet temperature change | Step change |
7 | Material C pressure loss (stream 4) | |
10 | Temperature Change of Material C (stream 4) | Random variable |
14 | Cooling water valve of reactor | Viscous glue |
The collected data contains a total of 3600 samples from 6 classes, 600 samples for each class. The collected data was divided into training data (containing 300 labeled data and 3000 unlabeled data) and test data (containing 300 labeled data). In order to simulate the situation that the distribution of the non-tag data is inconsistent with that of the tag data, Gaussian noise is added into the original non-tag data according to a certain proportion.
Fig. 3 shows log reconstruction errors of labeled data, normal unlabeled data, and abnormal unlabeled data that are not in accordance with the distribution of the labeled data under the stacked self-encoder reconstruction model. As is apparent from fig. 3, the reconstruction errors of the labeled data and the normal unlabeled data are relatively close, while the reconstruction error of the abnormal unlabeled data is significantly larger than the reconstruction errors of the labeled data and the normal unlabeled data. This is the basis for detecting abnormally distributed unlabeled data from the encoder based on weighted feature alignment.
Fig. 4 shows the classification accuracy of the three algorithms under different labeled and unlabeled data distribution inconsistent ratios. The MLP method is a supervised neural network classification model, the Tri-tracking method is a neural network classification model obtained based on cooperative training, and the Weighted FA-SAE method is a Weighted feature alignment-based self-encoder classification model provided by the invention. Tri-tracking and Weighted FA-SAE belong to a semi-supervised deep learning network model. As can be seen from the figure, the classification performance of most semi-supervised learning algorithms is superior to that of supervised algorithms; in addition, with the gradual expansion of the distribution inconsistency ratio of the labeled data and the unlabeled data, the performance of the semi-supervised algorithm is reduced, wherein when the distribution inconsistency reaches 90%, the classification precision of the Tri-tracking method is even lower than that of the supervised MLP method. In contrast, the Weighted FA-SAE method provided by the invention has better classification performance than MLP and Tri-tracking methods under different degrees of distribution inconsistency rate.
It will be understood by those skilled in the art that the foregoing is only a preferred embodiment of the invention and is not intended to limit the invention to the particular forms disclosed, and that modifications may be made, or equivalents may be substituted for elements thereof, while remaining within the scope of the claims that follow. All modifications, equivalents and the like which come within the spirit and principle of the invention are intended to be included within the scope of the invention.
Claims (2)
1. A semi-supervised fault classification method based on a weighted feature alignment self-encoder is characterized by comprising the following steps:
the method comprises the following steps: collecting normal working condition data and various fault data of an industrial process to obtain a training data set for modeling: sample set with labelsAnd unlabeled sample setWherein x represents an input sample, y represents a sample label, m represents the number of labeled samples, and n represents the number of unlabeled samples;
step two: constructing a stacking self-encoder model for reconstruction, and training the stacking self-encoder model by utilizing a labeled sample set;
step three: estimating the probability density distribution of the reconstruction error of the training data, calculating the weight of the label-free sample, and further constructing a weighted feature alignment self-encoder classification model;
the third step is specifically divided into the following substeps:
(3.1) calculating the reconstruction error E of the labeled exemplarslCompliance chi2Distribution ofDistribution parameters g and h of
g·h=mean(El)
2g2·h=variance(El)
(3.2) calculating reconstruction error of unlabeled exemplarThe reconstruction error calculation formula for a single sample is as follows:
(3.3) calculating the reconstruction error E of the unlabeled exemplarsuIn distribution ElProbability of occurrence ofTo PuNormalizing to obtain the weight of the unlabeled sample
(3.4) constructing a weighted feature alignment self-encoder classification model, and training the weighted feature alignment self-encoder classification model by adopting a labeled sample set, an unlabeled sample set and corresponding weights; the training process comprises the following steps: unsupervised pre-training and supervised fine tuning; in an unsupervised pre-training stage, a stack self-encoder is trained by adopting a labeled sample and an unlabeled sample together; the supervised fine tuning is formed by adding a fully-connected neural network layer on a stacked self-encoder obtained by unsupervised pre-training and using the fully-connected neural network layer as output of categories, so as to obtain deep extraction features and category labels of the labeled samples and deep extraction features and predicted category label output of the unlabeled samples, and a specific calculation formula is as follows:
wherein,represents the deep-extracted features of the ith labeled sample,class label representing predicted ith labeled sample, { wc,bcRepresenting weight vectors and deviation vectors of the fully connected neural network layer;represents a deep extraction feature of the unlabeled exemplar,a class label output representing a prediction;
(3.5) the number of classes is F, and deep extraction features of labeled samples and unlabeled samples corresponding to each class F epsilon F are obtainedAndand weight of unlabeled exemplars
(3.6) calculating a training loss function of the weighted feature alignment self-encoder classification model using the following formula:
wherein, crossentropy represents a cross entropy loss function,the representative weighted Sinkhorn distance function is used for measuring the distance between the characteristic distribution of the labeled data and the characteristic distribution of the unlabeled data belonging to the same category, and meanwhile, the weight reduction of the abnormal unlabeled sample with larger reconstruction error is realized; alpha is the weight of the Sinkhorn distance,l being a network parameter2Regularization penalty term, β is its weight, pijRepresenting features of labeled exemplars i corresponding to category fFeatures to unlabeled sample jTransition probability of dijRepresenting features of labeled exemplars i corresponding to class fFeatures to unlabeled sample jThe distance of (a) to (b),represents the weight of the unlabeled exemplar j corresponding to the class f, and mf and nf represent the number of labeled and unlabeled exemplars corresponding to the class f, respectively;
step four: and acquiring field working data, inputting the weighted features to align the self-encoder classification model, and outputting corresponding fault categories.
2. The semi-supervised fault classification method based on weighted feature alignment self-encoder according to claim 1, wherein the second step is specifically divided into the following sub-steps:
(2.1) constructing a stacked self-encoder model for reconstruction, comprising a multi-layer encoder and a decoder, wherein the output of the model is the reconstruction of the input, and the calculation formula is as follows:
wherein x represents the input, zkRepresenting the extracted k-th layer features, k representing the k-th layer of the stacked self-encoder,andrepresenting the weight vector and the disparity vector of the encoder and decoder respectively,reconstruction of the input by the representative model;
(2.2) training the stacked self-encoder model by adopting the labeled samples constructed in the step one and adopting a random gradient descent algorithm, wherein a model training loss function is defined as a reconstruction error of an input, and the reconstruction error is represented by the following formula:
wherein,representing the ith labeled input sample,representing the reconstruction of the stacked auto-encoder;
(2.3) calculating the reconstruction error of the labeled sample by using the trained stacked self-encoder modelWherein the reconstruction error of a single sample is calculated with reference to the following formula:
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110575307.4A CN113222045B (en) | 2021-05-26 | 2021-05-26 | Semi-supervised fault classification method based on weighted feature alignment self-encoder |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110575307.4A CN113222045B (en) | 2021-05-26 | 2021-05-26 | Semi-supervised fault classification method based on weighted feature alignment self-encoder |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113222045A CN113222045A (en) | 2021-08-06 |
CN113222045B true CN113222045B (en) | 2022-06-24 |
Family
ID=77098569
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110575307.4A Active CN113222045B (en) | 2021-05-26 | 2021-05-26 | Semi-supervised fault classification method based on weighted feature alignment self-encoder |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113222045B (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113705729B (en) * | 2021-09-27 | 2024-06-25 | 中原动力智能机器人有限公司 | Garbage classification model modeling method, garbage classification device and medium |
CN115184054B (en) * | 2022-05-30 | 2022-12-27 | 深圳技术大学 | Mechanical equipment semi-supervised fault detection and analysis method, device, terminal and medium |
CN114819108B (en) * | 2022-06-22 | 2022-10-04 | 中国电力科学研究院有限公司 | Fault identification method and device for comprehensive energy system |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111026058A (en) * | 2019-12-16 | 2020-04-17 | 浙江大学 | Semi-supervised deep learning fault diagnosis method based on Watherstein distance and self-encoder |
CN112183581A (en) * | 2020-09-07 | 2021-01-05 | 华南理工大学 | Semi-supervised mechanical fault diagnosis method based on self-adaptive migration neural network |
-
2021
- 2021-05-26 CN CN202110575307.4A patent/CN113222045B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111026058A (en) * | 2019-12-16 | 2020-04-17 | 浙江大学 | Semi-supervised deep learning fault diagnosis method based on Watherstein distance and self-encoder |
CN112183581A (en) * | 2020-09-07 | 2021-01-05 | 华南理工大学 | Semi-supervised mechanical fault diagnosis method based on self-adaptive migration neural network |
Non-Patent Citations (2)
Title |
---|
Semi-Supervised Bearing Fault Diagnosis and Classification Using Variational Autoencoder-Based Deep Generative Models;Shen Zhang et al.;《IEEE SENSORS JOURNAL》;20210301;第6476-6486页 * |
基于循环神经网络的半监督动态软测量建模方法;邵伟明等;《电子测量与仪器学报》;20191115(第11期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN113222045A (en) | 2021-08-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113222045B (en) | Semi-supervised fault classification method based on weighted feature alignment self-encoder | |
CN108875772B (en) | Fault classification model and method based on stacked sparse Gaussian Bernoulli limited Boltzmann machine and reinforcement learning | |
CN103914064B (en) | Based on the commercial run method for diagnosing faults that multi-categorizer and D-S evidence merge | |
CN109146246B (en) | Fault detection method based on automatic encoder and Bayesian network | |
CN106649789B (en) | It is a kind of based on the industrial process Fault Classification for integrating semi-supervised Fei Sheer and differentiating | |
CN113642754B (en) | Complex industrial process fault prediction method based on RF noise reduction self-coding information reconstruction and time convolution network | |
CN106843195B (en) | The Fault Classification differentiated based on adaptive set at semi-supervised Fei Sheer | |
CN108875771A (en) | A kind of failure modes model and method being limited Boltzmann machine and Recognition with Recurrent Neural Network based on sparse Gauss Bernoulli Jacob | |
CN113222046B (en) | Feature alignment self-encoder fault classification method based on filtering strategy | |
CN101169623A (en) | Non-linear procedure fault identification method based on kernel principal component analysis contribution plot | |
CN112904810B (en) | Process industry nonlinear process monitoring method based on effective feature selection | |
CN108375965A (en) | A kind of nongausian process monitoring method rejected based on changeable gauge block crossing dependency | |
CN111026058A (en) | Semi-supervised deep learning fault diagnosis method based on Watherstein distance and self-encoder | |
CN108345284A (en) | A kind of quality dependent failure detection method becoming gauge block based on two | |
CN110765587A (en) | Complex petrochemical process fault diagnosis method based on dynamic regularization judgment local retention projection | |
CN110175640A (en) | A kind of Fault Diagnosis Method of Electro-hydraulic based on machine learning | |
WO2023273249A1 (en) | Tsvm-model-based abnormality detection method for automatic verification system of smart electricity meter | |
CN108830006B (en) | Linear-nonlinear industrial process fault detection method based on linear evaluation factor | |
CN114757269A (en) | Complex process refined fault detection method based on local subspace-neighborhood preserving embedding | |
CN116204825A (en) | Production line equipment fault detection method based on data driving | |
CN114492614A (en) | Method and device for classifying faults in hot rolling process of strip steel based on ensemble learning | |
CN111914886B (en) | Nonlinear chemical process monitoring method based on online brief kernel learning | |
CN115564021A (en) | Fault root cause sequencing method in polyester fiber polymerization process | |
CN115618708A (en) | Equipment health state prediction method based on incremental inform algorithm | |
CN114120043A (en) | Method for detecting abnormal pumping well based on production dynamic data and indicator diagram |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |