CN112613226A - Feature enhancement method for residual life prediction - Google Patents

Feature enhancement method for residual life prediction Download PDF

Info

Publication number
CN112613226A
CN112613226A CN202011432003.4A CN202011432003A CN112613226A CN 112613226 A CN112613226 A CN 112613226A CN 202011432003 A CN202011432003 A CN 202011432003A CN 112613226 A CN112613226 A CN 112613226A
Authority
CN
China
Prior art keywords
data
test
sample
training
rul
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011432003.4A
Other languages
Chinese (zh)
Other versions
CN112613226B (en
Inventor
孙希明
王哲夫
王天诚
郭迪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Dalian University of Technology
Original Assignee
Dalian University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Dalian University of Technology filed Critical Dalian University of Technology
Priority to CN202011432003.4A priority Critical patent/CN112613226B/en
Publication of CN112613226A publication Critical patent/CN112613226A/en
Application granted granted Critical
Publication of CN112613226B publication Critical patent/CN112613226B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F30/00Computer-aided design [CAD]
    • G06F30/20Design optimisation, verification or simulation
    • G06F30/27Design optimisation, verification or simulation using machine learning, e.g. artificial intelligence, neural networks, support vector machines [SVM] or training a model
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F30/00Computer-aided design [CAD]
    • G06F30/10Geometric CAD
    • G06F30/15Vehicle, aircraft or watercraft design
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2119/00Details relating to the type or aim of the analysis or the optimisation
    • G06F2119/04Ageing analysis or optimisation against ageing

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • Geometry (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Computer Hardware Design (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Medical Informatics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Automation & Control Theory (AREA)
  • Aviation & Aerospace Engineering (AREA)
  • Computational Mathematics (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Pure & Applied Mathematics (AREA)
  • Testing Of Devices, Machine Parts, Or Other Structures Thereof (AREA)

Abstract

A characteristic enhancement method for residual life prediction belongs to the field of aircraft engine fault prediction and health management. First, the engine sensor data is normalized and the remaining life values of the training set data and the test set data are calculated. Secondly, the training data set and the testing data set are subjected to dimensionality reduction, and a training data set sample and a testing data set sample are extracted in a sliding window mode. And thirdly, performing data characteristic enhancement on the training set sample and the test and sample, and determining the sample after the data characteristic enhancement. And finally, setting a long-term and short-term memory structure of the deep learning neural network, adding a sample for training, and predicting test data by using the neural network model. The deep neural network model is established based on a data-driven form, is irrelevant to the actual engine model, can be migrated to engines of different models for use by training different data sets, and has certain universality.

Description

Feature enhancement method for residual life prediction
Technical Field
The invention provides a novel Tensorflow deep learning framework-based data enhancement method for predicting the residual life of an aircraft engine system or a component, and belongs to the field of aircraft engine fault Prediction and Health Management (PHM).
Background
The subject of the invention is the turbofan engine degradation simulation dataset disclosed by NASA. The data set was used for engine degradation simulation using C-MAPSS, four different sets were simulated under different combinations of conditions and failure modes, and multiple sensor channel data were recorded to characterize the failure evolution.
The data set is divided into four sub-data sets, each of which is divided into a training data set, a test data set, and a true system or component remaining life value (RUL). For each subdata set, the training data set and the test data set are composed of a plurality of parameters, the parameters include engine number, time series, 3 engine operation setting parameters and 21 sensor parameter data, and the parameters are time series parameters.
The engine operates normally at the beginning of each time series and then a fault occurs at some point in the time series. In the training set, the fault continues to increase until the system fails. In the test set, the time series ends at a time prior to the system failure. What we need to do is to predict the number of operating cycles remaining before the test set fails, i.e., the number of operating cycles that the engine will continue to operate after the last operating cycle, based on the complete training data set given. The RUL file for each subdata set also provides the actual Remaining Useful Life (RUL) value for the engine to which the test data corresponds.
For a time series turbofan engine degradation simulation data set, most of the current methods adopt a Recurrent Neural Network (RNN) to train and predict the data set, and a long-short term memory network (LSTM) is also an RNN network model which proves to be very successful and is widely applied to the aspects of speech recognition, machine translation and the like. LSTM solves the problem of gradient disappearance or gradient explosion that typical RNNs can exhibit, using input, forgetting and output gates to control the flow of information, and the inherent sequential nature of sensor data for this turbofan engine also means that it is suitable for use with LSTM for RUL prediction.
Disclosure of Invention
The invention provides a feature extraction method based on data increase, aiming at the problem that the features and information of an original data set cannot be fully learned and mined by adding data directly using standard normalization into an LSTM network for training.
The technical scheme of the invention is as follows:
a novel feature enhancement method for remaining life prediction comprises the following steps:
s1, normalization processing is carried out on engine sensor data, and the method comprises the following steps:
s1.1 reading public NASA turbofan engine training set data train _ FD001, test set data test _ FD001, and remaining life data set RUL _ FD001 (the public data set includes four divided sub-data sets, each of which is divided into a training data set, a test data set, and a remaining life data set, here the first sub-data set is taken as an example);
s1.2, the sensor parameter data of the training set data train _ FD001 and the test set data test _ FD001 are time series data, the sensor data are normalized by adopting a maximum and minimum normalization method, and normalized training set data train _01_ ru and test set data test _01_ ru are obtained, wherein the normalization formula is as follows:
Figure BDA0002826813810000021
wherein x is sensor parameter data before normalization, and x' isNormalized sensor parameter data, xminIs the minimum value of the sensor parameter, xmaxIs the maximum value of the sensor parameter;
s2, calculating the residual life RUL values of the training set data and the test set data according to the training set data train _01_ ru, the test set data test _01_ ru and the residual life data set RUL _ FD001 obtained in the step S1, and the method comprises the following steps:
s2.1 finding the maximum engine cycle number cycle corresponding to each engine number id of the training data setmax_train(the first and second columns of parameters in the training dataset data are engine number and engine cycle number, respectively), cycle is usedmax_trainMinus the number of engine cycles cyclet_trainObtaining the corresponding residual service life RUL under the engine cycle numbert_trainA value;
s2.2 finding the maximum engine cycle number cycle corresponding to each engine number id of the test data settest(the first and second columns of parameters in the test data set are the engine number and the engine cycle number respectively), and the engine cycle number is added with the RUL value in the remaining life data set (the data set only contains the remaining life RUL value corresponding to each engine number), so as to obtain the maximum cycle number cycle of the enginemax_testUsing cyclemax_testMinus the number of engine cycles cyclet_testObtaining the corresponding residual service life RUL under the engine cycle numbert_testA value;
s2.3 setting the maximum residual Life RULmaxWith a value of 125, recalculating the residual lifetime RUL of the training set datat_trainResidual lifetime RUL of value and test set datat_testThe value, the calculation formula is as follows:
Figure BDA0002826813810000031
in the formula, RULtResidual Life RUL for training set datat_trainRemaining lifetime RUL of value or test set datat_testA value;
s3, performing dimension reduction processing on the normalized training data set and the test data set obtained in the step S2, namely removing the sensor parameter data with parameter values unchanged from the data set in the sensor parameter data, and taking the rest sensor parameter data as original sample data;
s4, obtaining the residual life RUL of the training set data according to the training data set train _01_ ru and the test data set test _01_ ru which are subjected to dimension reduction processing in S3 and the training set data obtained in S2t_trainResidual lifetime RUL of value and test set datat_testValues, extracting training data set samples and testing data set samples in a sliding window manner, comprising the steps of:
s4.1 finding the maximum engine cycle number cycle corresponding to each engine number id of the training data setmax_trainAnd taking out the minimum cycle thereofmax_train_minSetting the sliding window size winSize close to and not exceeding the minimum cyclemax_train_min
S4.2, taking the size of the sliding window winSize as a row, taking the dimension reduction sensor parameter data FeaSIze as a column, extracting samples of a training data set train _01_ ru and a test data set test _01_ ru, taking the samples as a 2-dimensional matrix with the winSize as a row and the FeaSIze as a column, and taking the output of a corresponding network as the residual life of the engine (sampling the RUL corresponding to the last time sequence of the samplet) At this time, the input sample of the training data set is train x _ New, the input sample of the testing data set is testX _ New, the output sample of the training data set is train y, and the output sample of the testing data set is testY;
s5, performing data feature enhancement on the training set sample train X _ New and the test and sample testX _ New acquired in the step S4 in a manual feature extraction mode, adding four feature values of ridge regression weight, sample mean value, sample maximum value and sample minimum value to each sample, expanding sample dimensionality, and determining a sample after data feature enhancement, wherein the method comprises the following steps:
s5.1, respectively calculating a ridge regression weight value coef, a mean value mean, a maximum value max and a minimum value min of each sensor parameter for the training set sample train X _ New and the test and sample testX _ New obtained in the step S4;
s5.2 for the four characteristic values obtained, winSize is added to enhanceObtaining a final data feature enhanced sample from the samples, wherein the sample is a 2-dimensional matrix with winSize +4 as rows and FeaSeze as columns, the final training set sample is defined as train X _ reNew, the final testing set sample is defined as testX _ New, and the corresponding output engine residual life RUL istThe values are unchanged, i.e. trainY and testY are unchanged;
s6, setting a deep learning neural network long-short term memory (LSTM) structure, adding a sample for training, and predicting test data by using a trained neural network model, wherein the method comprises the following steps:
s6.1, establishing a Sequential model structure which is a linear stack of a network layer;
s6.2, adding an LSTM layer and an NN layer, wherein the LSTM layer structure is generally set to be a 3-4-layer neural network structure, the NN layer is generally set to be a 1-2-layer neural network structure, a Dropout function is used for preventing overfitting of neural network model learning, and a Reynolds function (RELU) is adopted as an activation function;
s6.3, selecting an optimizer of the network, a target loss function loss, a batch size and a training time epochs, and adding an input sample train X _ reNew and an output sample train Y into the network for training to obtain a prediction model;
s6.4, predicting the testset testX _ reenw by the trained prediction model, carrying out model prediction comparison by combining testY to obtain prediction results rmse and score,
wherein rmse is the root mean square error and the calculation mode is
Figure BDA0002826813810000051
In the above formula, n is the number of samples, hiPredicting RUL for remaining lifepreAnd actual remaining life value RULtrue(i.e., the RUL _ FD001 data RUL value).
score is a scoring function calculated by
Figure BDA0002826813810000052
Likewise, n is the number of samples, hiPredicting RUL for remaining lifepreAnd actual remaining life value RULtrueThe error between.
The invention has the beneficial effects that: compared with a deep LSTM network which directly performs normalized training on data, the root mean square error rmse and score of a target function of the deep LSTM network are reduced to a certain degree, the prediction accuracy of the network model is improved to a certain degree, and in addition, the results of multiple tests also show that the prediction stability of the network model is improved to a certain degree. The method establishes the deep neural network model based on a data-driven form, and is irrelevant to the actual engine model, so that the model can be conveniently transferred to engines of different models for use by training different data sets, and the method has certain universality.
Drawings
FIG. 1 is a flow diagram of a CMPASS dataset system or component remaining life prediction based on deep learning;
FIG. 2 is a flow chart of data processing in three different ways (wherein the third method is the method provided by the present invention);
FIG. 3 is a diagram of a neural network model architecture in accordance with the present disclosure.
Detailed Description
The new feature enhancement method for remaining life prediction according to the present invention will be further described with reference to the accompanying drawings.
The present invention relies on the turbine fan engine degradation simulation dataset disclosed in the background by NASA, and is based on a new feature enhancement method flow for residual life prediction as shown in FIG. 1.
Fig. 2 is a flow chart of three methods for processing data and training a network respectively, the first two methods in fig. 2 are conventional methods, the first method is to determine samples in a sliding window manner after only performing normalization processing on original data to perform network training for predicting RUL, the second method is similar to the first method except that feature values are obtained from the original data to perform subsequent network training and prediction operations instead of the original data, the method is the third method, and the main steps include:
s1, determining a data set as a degradation simulation data set of a turbofan engine disclosed by NASA, wherein the data set comprises four sub data sets, each sub data set comprises a training data set, a testing data set and an RUL data set, and the purpose is to train a deep learning neural network according to the training data set for testing the data set and compare the difference between a prediction result and actual RUL data;
s2, carrying out normalization processing on the data of the engine sensor, and comprising the following steps:
s2.1 reading public NASA turbofan engine training set data train _ FD001, test set data test _ FD001, and remaining life data set RUL _ FD001 (the public data set includes four divided sub-data sets, each of which is divided into a training data set, a test data set, and a remaining life data set, here the first sub-data set is taken as an example);
s2.2, the sensor parameter data of the training set data train _ FD001 and the test set data test _ FD001 which are read are time series data, normalization processing is carried out on the sensor data by adopting a maximum and minimum normalization method, normalized training set data train _01_ ru and normalized test set data test _01_ ru are obtained, and the normalization formula is as follows:
Figure BDA0002826813810000071
wherein x is sensor parameter data before normalization, x' is sensor parameter data after normalization, and xminIs the minimum value of the sensor parameter, xmaxIs the maximum value of the sensor parameter;
s3, calculating the residual life RUL values of the training set data and the test set data according to the training set data train _01_ ru, the test set data test _01_ ru and the residual life data set RUL _ FD001 obtained in the step S2, and the method comprises the following steps:
s3.1 finding the maximum engine cycle number cycle corresponding to each engine number id of the training data setmax_train(the first and second columns of parameters in the training dataset data are engine number and engine cycle number, respectively), cycle is usedmax_trainMinus the number of engine cycles cyclet_trainObtaining the corresponding residual service life RUL under the engine cycle numbert_trainA value;
s3.2 finding the maximum engine cycle number cycle corresponding to each engine number id of the test data settest(the first and second columns of parameters in the test data set are the engine number and the engine cycle number respectively), and the engine cycle number is added with the RUL value in the remaining life data set (the data set only contains the remaining life RUL value corresponding to each engine number), so as to obtain the maximum cycle number cycle of the enginemax_testUsing cyclemax_testMinus the number of engine cycles cyclet_testObtaining the corresponding residual service life RUL under the engine cycle numbert_testA value;
s3.3 setting the maximum residual Life RULmaxWith a value of 125, recalculating the residual lifetime RUL of the training set datat_trainResidual lifetime RUL of value and test set datat_testThe value, the calculation formula is as follows:
Figure BDA0002826813810000081
in the formula, RULtResidual Life RUL for training set datat_trainRemaining lifetime RUL of value or test set datat_testA value;
s4, performing dimensionality reduction on the normalized training data set and the test data set obtained in the S3, namely removing sensor parameter data with parameter values unchanged from the data set (the data set can be observed to find that 7 sensor parameter values in the sensor parameter data keep constant values in a given time sequence, and if the sensor parameters are S1, S2 and so on, S1, S5, S6, S10, S16, S18 and S19 parameter values are constant respectively, and removing the data to achieve a dimensionality reduction effect), and taking the rest sensor parameter data as original sample data;
s5, obtaining the residual life RUL of the training set data according to the training data set train _01_ ru and the test data set test _01_ ru which are subjected to dimension reduction processing in S4 and the training set data obtained in S3t_trainResidual lifetime RUL of value and test set datat_testValues, extracting training data set samples and testing data set samples in a sliding window manner, comprising the steps of:
s5.1 finding the maximum engine cycle number cycle corresponding to each engine number id of the training data setmax_trainAnd taking out the minimum cycle thereofmax_train_min(the values corresponding to the four subdata sets are found to be 31,21,38 and 19 respectively in the experiment), and the size of the sliding window winSize is set to be close to and not exceed the minimum value cyclemax_train_min(setting the sliding window sizes winSize to 30,20,38,19, respectively);
s5.2, taking the size of the sliding window winSize as a row, taking the dimension reduction sensor parameter data FeaSIze as a column, extracting samples of a training data set train _01_ ru and a test data set test _01_ ru, taking the samples as a 2-dimensional matrix with the winSize as a row and the FeaSIze as a column, and taking the output of a corresponding network as the residual life of the engine (sampling the RUL corresponding to the last time sequence of the samplet) At this time, the input sample of the training data set is train x _ New, the input sample of the testing data set is testX _ New, the output sample of the training data set is train y, and the output sample of the testing data set is testY;
s6, performing data feature enhancement on the training set sample train X _ New and the test and sample testX _ New acquired in the step S5 in a manual feature extraction mode, adding four feature values of ridge regression weight, sample mean value, sample maximum value and sample minimum value to each sample, expanding sample dimensionality, and determining a sample after data feature enhancement, wherein the method comprises the following steps:
s6.1, respectively calculating a ridge regression weight value coef, a mean value mean, a maximum value max and a minimum value min of each sensor parameter for the training set sample train X _ New and the test and sample testX _ New obtained in S5, wherein the ridge regression weight value coef can be obtained by a RidgeCV () function in a linear _ model of a sketch kit, and the mean value, the maximum value and the minimum value of the sensor parameters can be obtained by a corresponding function mean (), max () and min () of the numpy kit;
s6.2, adding the obtained four characteristic values into a sample in a winSize enhancing mode to obtain a final data characteristic enhanced sample, wherein the sample is a 2-dimensional matrix with winSize +4 as rows and FeaSize as columns, the final training set sample is defined as train X _ reNew, the final test set sample is testX _ New, and the corresponding output engine residual life RUL istThe values are unchanged, i.e. trainY and testY are unchanged;
s7, setting a deep learning neural network long-short term memory (LSTM) structure, adding a sample for training, and predicting test data by using a trained neural network model, wherein the method comprises the following steps:
s7.1, establishing a Sequential model structure which is a linear stack of a network layer;
s7.2, adding an LSTM layer and an NN layer, wherein the LSTM layer structure is generally set to be a 3-4-layer neural network structure, the NN layer is generally set to be a 1-2-layer neural network structure, a Dropout function is used for preventing overfitting of neural network model learning, a Reynolds function (RELU) is adopted as an activation function, the network structure is shown in figure 3, namely, the 4-layer LSTM network structure is adopted, the number of memory units is respectively 64,64,8 and 8, a two-layer NN structure is adopted, and the number of neuron nodes is respectively 16 and 1;
the expression of the RELU function is as follows
f(x)=max(0,x)
Where x is the functional input, for a given input to a neuron, the above equation becomes
f(x)=max(0,x)=max(0,WTx+b)
In the formula WTIs the weight of the neuron, and b is the bias of the neuron;
s7.3, selecting an optimizer of the network as rmsprop, a target loss function loss as mean square error mse, a batch size of 512 and training times epochs of 100, and adding an input sample train X _ reNew and an output sample train Y into the network for training to obtain a prediction model;
s7.4, predicting the testset testX _ reenw by the trained prediction model, carrying out model prediction comparison by combining testY to obtain prediction results rmse and score,
wherein rmse is the root mean square error and the calculation mode is
Figure BDA0002826813810000101
In the above formula, n is the number of samples, hiPredicting RUL for remaining lifepreAnd actual remaining life value RULtrue(i.e., the RUL _ FD001 data RUL value).
score is a scoring function calculated by
Figure BDA0002826813810000102
Likewise, n is the number of samples, hiPredicting RUL for remaining lifepreAnd actual remaining life value RULtrueThe error between.
Tables 1,2 and 3 are the results of the simulation of the sub-data set 1 of the turbofan engine degradation simulation data set disclosed by NASA according to the three different data processing methods in fig. 2. The prediction results of the LSTM network model obtained by the common direct normalization processing method have the RMSE mean value mean of 13.965, std of 0.520, score of 341.269 and std of 107.875, the second model prediction result of LSTM network training by using characteristic values instead of original data has the RMSE mean value mean of 13.904, std of 0.527, score of 297.373 and std of 36.840, the results obtained by the data enhancement method are that the RMSE mean value mean is 12.479, std of 0.348, score of 279.212 and std of 21.889, and it can be seen that the root mean square errors obtained by the first method and the second method are not greatly different, the scoring function obtained by the second method is effectively reduced compared with the first method, the standard deviation is also reduced, and the stability is better. The data enhancement method in the present invention can obtain better prediction performance, further decrease the root mean square error and the scoring function, and show better stability compared with the former two methods.
TABLE 1 prediction result table of LSTM trained based on common normalization method of subdata set 1
RMSE Score
1 13.192 327.142
2 13.263 234.108
3 13.814 261.377
4 14.474 604.091
5 13.955 428.477
6 14.52 312.76
7 14.477 307.913
8 13.82 251.531
9 14.665 274.354
10 13.469 344.024
mean 13.965 341.269
std 0.520 107.875
TABLE 2 prediction results Table for LSTM trained based on the feature data of subdata set 1
RMSE Score
1 13.634 293.456
2 13.519 263.007
3 13.788 296.983
4 14.646 274.509
5 13.497 271.426
6 13.566 287.183
7 14.483 379.878
8 14.694 354.434
9 14.143 284.636
10 13.073 268.22
mean 13.904 297.373
std 0.527 36.840
TABLE 3 prediction results Table for the present method based on subdata set 1
Figure BDA0002826813810000111
Figure BDA0002826813810000121
Table 4, table 5 and table 6 the results obtained from the simulation of the subset of turbofan engine degradation simulation data sets 2,3,4 disclosed by NASA according to the method herein.
TABLE 4 prediction results Table for the present methods based on subdata set 2
RMSE Score
1 17.443 1409.015
2 16.709 1308.895
3 17.076 1342.558
4 17.384 1511.902
5 17.656 1809.671
6 17.081 1292.520
7 17.041 1383.842
8 17.239 1323.760
9 17.143 1260.521
10 17.314 1587.507
mean 17.209 1423.019
std 0.248 160.927
TABLE 5 prediction results Table for the present methods based on subdata set 3
RMSE Score
1 12.175 253.251
2 11.935 255.896
3 12.332 259.733
4 12.492 271.425
5 12.687 343.225
6 12.332 255.477
7 12.254 267.069
8 12.213 231.005
9 11.837 234.359
10 12.414 259.910
mean 12.267 263.135
std 0.237 29.308
Table 6 prediction results table for the present method based on subdata set 4
Figure BDA0002826813810000122
Figure BDA0002826813810000131
In conclusion, the novel feature enhancement method for predicting the residual life is feasible, the prediction performance can be obviously improved, the prediction capability is stable, the root mean square error and the scoring function of the feature enhancement method are effectively reduced, and the results are reasonable and are not accidental results generated by network fluctuation.
The above-mentioned embodiments only express the embodiments of the present invention, but not should be understood as the limitation of the scope of the invention patent, it should be noted that, for those skilled in the art, many variations and modifications can be made without departing from the concept of the present invention, and these all fall into the protection scope of the present invention.

Claims (4)

1. A feature enhancement method for remaining life prediction, comprising the steps of:
s1, normalization processing is carried out on engine sensor data, and the method comprises the following steps:
s1.1, reading training set data, test set data and residual life data set data of the turbofan engine;
s1.2, the sensor parameter data of the training set data train _ FD001 and the test set data test _ FD001 which are read are time series data, and the sensor data are normalized by adopting a maximum and minimum normalization method to obtain normalized training set data and normalized test set data;
s2, calculating the residual life RUL values of the training set data and the test set data according to the residual life data set obtained in the step S1 and the training set data and the test set data after normalization processing, and comprising the following steps of:
s2.1 according to the maximum engine cycle number cycle corresponding to each engine number of the training data setmax_trainUsing cyclemax_trainMinus the number of engine cycles cyclet_trainObtaining the corresponding residual service life RUL under the engine cycle numbert_trainA value;
s2.2 according to the test data set, the maximum engine cycle number cycle corresponding to each engine numbertestAdding the RUL value in the remaining life data set to the number of engine cycles to obtain the maximum number of engine cycles cyclemax_testUsing cyclemax_testMinus the number of engine cycles cyclet_testObtaining the corresponding residual service life RUL under the engine cycle numbert_testA value;
s2.3 setting the maximum residual Life RULmaxValue, recalculation of training set data residual Life RULt_trainResidual lifetime RUL of value and test set datat_testThe value, the calculation formula is as follows:
Figure FDA0002826813800000011
in the formula, RULtResidual Life RUL for training set datat_trainRemaining lifetime RUL of value or test set datat_testA value;
s3, performing dimension reduction on the normalized training data set and the test data set obtained in the step S2, removing the sensor parameter data with parameter values unchanged from the data set in the sensor parameter data, and taking the rest sensor parameter data as original sample data;
s4, according to the training data set and the test data set subjected to the dimensionality reduction processing in the step S3 and the residual life RUL of the training set data obtained in the step S2t_trainResidual lifetime RUL of value and test set datat_testValues, extracting training data set samples and testing data set samples in a sliding window manner, comprising the steps of:
s4.1 finding the maximum engine cycle number cycle corresponding to each engine number of the training data setmax_trainAnd taking out the minimum cycle thereofmax_train_minSetting the sliding window size winSize close to and not exceeding the minimum cyclemax_train_min
S4.2, taking the size of a sliding window as a row and the parameter data of the dimension reduction sensor as a column, extracting samples of a training data set and a testing data set, wherein the output of a corresponding network is the residual life of the engine, the input sample of the training data set at the moment is train X _ New, the input sample of the testing data set is testX _ New, the output sample of the training data set is train Y, and the output sample of the testing data set is testY;
s5, performing data feature enhancement on the training set sample train X _ New and the test and sample testX _ New acquired in the step S4 by using a manual feature extraction mode, adding four feature values of ridge regression weight, sample mean value, sample maximum value and sample minimum value to each sample, expanding sample dimensionality, and determining a sample after data feature enhancement, wherein the method comprises the following steps:
s5.1, respectively calculating a ridge regression weight value coef, a mean value mean, a maximum value max and a minimum value min of each sensor parameter for the training set sample train X _ New and the test and sample testX _ New obtained in the step S4;
s5.2, adding the obtained four characteristic values into a sample in a winSize enhancement mode to obtain a final data characteristic enhanced sample; defining the final training set sample as train X _ reNew, and the final testing set sample as testX _ New, and corresponding to the residual service life RUL of the output enginetThe values are unchanged, i.e. trainY and testY are unchanged;
s6, setting a deep learning neural network long-term and short-term memory LSTM structure, adding a sample for training, and predicting test data by using a trained neural network model, wherein the method comprises the following steps:
s6.1, establishing a Sequential model structure which is a linear stack of a network layer;
s6.2, adding an LSTM layer and an NN layer, and preventing overfitting of neural network model learning by adopting an activation function;
s6.3, selecting an optimizer of the network, a target loss function loss, a batch size and a training time epochs, and adding an input sample train X _ reNew and an output sample train Y into the network for training to obtain a prediction model;
s6.4, predicting the testset testX _ reenw by the trained prediction model, and performing model prediction comparison by combining testY to obtain a prediction result: root mean square error rmse and scoring function score;
the calculation mode of the root mean square error is as follows:
Figure FDA0002826813800000031
in the above formula, n is the number of samples, hiPredicting RUL for remaining lifepreAnd actual remaining life value RULtrueThe error between;
the calculation mode of the scoring function is as follows:
Figure FDA0002826813800000032
likewise, n is the number of samples, hiPredicting RUL for remaining lifepreAnd actual remaining life value RULtrueThe error between.
2. The method as claimed in claim 1, wherein the step S2.3 of setting the maximum remaining lifetime RUL is performed by using a method of predicting the remaining lifetimemaxThe value is 125.
3. A method for feature enhancement for residual life prediction according to claim 1, characterized in that in step S6.2: the LSTM layer structure is generally set to be a 3-4 layer neural network structure, and the NN layer is generally set to be a 1-2 layer neural network structure.
4. A feature enhancement method for predicting remaining life according to claim 1, wherein the activation function in step S6.2 is a reynolds function.
CN202011432003.4A 2020-12-10 2020-12-10 Feature enhancement method for residual life prediction Active CN112613226B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011432003.4A CN112613226B (en) 2020-12-10 2020-12-10 Feature enhancement method for residual life prediction

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011432003.4A CN112613226B (en) 2020-12-10 2020-12-10 Feature enhancement method for residual life prediction

Publications (2)

Publication Number Publication Date
CN112613226A true CN112613226A (en) 2021-04-06
CN112613226B CN112613226B (en) 2022-11-18

Family

ID=75229510

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011432003.4A Active CN112613226B (en) 2020-12-10 2020-12-10 Feature enhancement method for residual life prediction

Country Status (1)

Country Link
CN (1) CN112613226B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113449463A (en) * 2021-06-09 2021-09-28 重庆锦禹云能源科技有限公司 LSTM-DNN-based equipment life prediction method and device
CN113722989A (en) * 2021-08-23 2021-11-30 南京航空航天大学 CPS-DP model-based aircraft engine life prediction method
CN113722833A (en) * 2021-09-09 2021-11-30 湖南工业大学 Turbofan engine residual service life prediction method based on dual-channel long-short time memory network
CN114186337A (en) * 2021-11-30 2022-03-15 大连理工大学 Gas compressor rotating stall prediction method based on multi-source data fusion
CN114944057A (en) * 2022-04-21 2022-08-26 中山大学 Road network traffic flow data restoration method and system
WO2024020960A1 (en) * 2022-07-28 2024-02-01 西门子股份公司 Method for predicting remaining useful life of circuit breaker, and electronic device and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110555230A (en) * 2019-07-12 2019-12-10 北京交通大学 rotary machine residual life prediction method based on integrated GMDH framework
CN110807257A (en) * 2019-11-04 2020-02-18 中国人民解放军国防科技大学 Method for predicting residual life of aircraft engine

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110555230A (en) * 2019-07-12 2019-12-10 北京交通大学 rotary machine residual life prediction method based on integrated GMDH framework
CN110807257A (en) * 2019-11-04 2020-02-18 中国人民解放军国防科技大学 Method for predicting residual life of aircraft engine

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
王旭等: "基于CAE与LSTM的航空发动机剩余寿命预测", 《北京信息科技大学学报(自然科学版)》 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113449463A (en) * 2021-06-09 2021-09-28 重庆锦禹云能源科技有限公司 LSTM-DNN-based equipment life prediction method and device
CN113722989A (en) * 2021-08-23 2021-11-30 南京航空航天大学 CPS-DP model-based aircraft engine life prediction method
CN113722989B (en) * 2021-08-23 2023-04-28 南京航空航天大学 CPS-DP model-based aeroengine service life prediction method
CN113722833A (en) * 2021-09-09 2021-11-30 湖南工业大学 Turbofan engine residual service life prediction method based on dual-channel long-short time memory network
CN114186337A (en) * 2021-11-30 2022-03-15 大连理工大学 Gas compressor rotating stall prediction method based on multi-source data fusion
WO2023097705A1 (en) * 2021-11-30 2023-06-08 大连理工大学 Air compressor rotation stall prediction method based on multi-source data fusion
CN114944057A (en) * 2022-04-21 2022-08-26 中山大学 Road network traffic flow data restoration method and system
CN114944057B (en) * 2022-04-21 2023-07-25 中山大学 Road network traffic flow data restoration method and system
WO2024020960A1 (en) * 2022-07-28 2024-02-01 西门子股份公司 Method for predicting remaining useful life of circuit breaker, and electronic device and storage medium

Also Published As

Publication number Publication date
CN112613226B (en) 2022-11-18

Similar Documents

Publication Publication Date Title
CN112613226B (en) Feature enhancement method for residual life prediction
CN108805185B (en) Face recognition method and device, storage medium and computer equipment
CN108399201B (en) Web user access path prediction method based on recurrent neural network
CN112149316B (en) Aero-engine residual life prediction method based on improved CNN model
CN110298264B (en) Human body daily behavior activity recognition optimization method based on stacked noise reduction self-encoder
CN106203534A (en) A kind of cost-sensitive Software Defects Predict Methods based on Boosting
CN110688288A (en) Automatic testing method, device, equipment and storage medium based on artificial intelligence
CN111460728A (en) Method and device for predicting residual life of industrial equipment, storage medium and equipment
Li et al. Domain adaptation remaining useful life prediction method based on AdaBN-DCNN
CN112149373A (en) Complex analog circuit fault identification and estimation method and system
CN110210495A (en) The XGBoost soft-measuring modeling method extracted based on parallel LSTM self-encoding encoder behavioral characteristics
CN112001110A (en) Structural damage identification monitoring method based on vibration signal space real-time recursive graph convolutional neural network
CN114297918A (en) Aero-engine residual life prediction method based on full-attention depth network and dynamic ensemble learning
CN113268833A (en) Migration fault diagnosis method based on deep joint distribution alignment
Sadoughi et al. A deep learning approach for failure prognostics of rolling element bearings
CN111079348A (en) Method and device for detecting slowly-varying signal
CN114462459A (en) Hydraulic machine fault diagnosis method based on 1DCNN-LSTM network model
JP2019095600A (en) Acoustic model learning device, speech recognition device, and method and program for them
CN114169091A (en) Method for establishing prediction model of residual life of engineering mechanical part and prediction method
CN112185423B (en) Voice emotion recognition method based on multi-head attention mechanism
CN109493975B (en) Chronic disease recurrence prediction method, device and computer equipment based on xgboost model
CN116720079A (en) Wind driven generator fault mode identification method and system based on multi-feature fusion
Arifin et al. Automatic essay scoring for Indonesian short answers using siamese Manhattan long short-term memory
CN113052060B (en) Bearing residual life prediction method and device based on data enhancement and electronic equipment
CN113962431A (en) Bus load prediction method for two-stage feature processing

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant