CN111680454A - Fan blade icing fault prediction method based on double attention mechanism - Google Patents
Fan blade icing fault prediction method based on double attention mechanism Download PDFInfo
- Publication number
- CN111680454A CN111680454A CN202010551118.9A CN202010551118A CN111680454A CN 111680454 A CN111680454 A CN 111680454A CN 202010551118 A CN202010551118 A CN 202010551118A CN 111680454 A CN111680454 A CN 111680454A
- Authority
- CN
- China
- Prior art keywords
- data
- attention
- fan
- icing
- attention mechanism
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 35
- 230000007246 mechanism Effects 0.000 title claims abstract description 24
- 238000012549 training Methods 0.000 claims abstract description 11
- 238000012360 testing method Methods 0.000 claims abstract description 9
- 238000007781 pre-processing Methods 0.000 claims abstract description 5
- 239000013598 vector Substances 0.000 claims description 41
- 239000011159 matrix material Substances 0.000 claims description 10
- 238000000605 extraction Methods 0.000 claims description 9
- 230000004913 activation Effects 0.000 claims description 7
- 230000008014 freezing Effects 0.000 claims description 7
- 238000007710 freezing Methods 0.000 claims description 7
- 238000004364 calculation method Methods 0.000 claims description 6
- 238000012795 verification Methods 0.000 claims description 6
- 210000002569 neuron Anatomy 0.000 claims description 5
- 230000009977 dual effect Effects 0.000 claims 6
- 238000010248 power generation Methods 0.000 abstract description 6
- 238000001514 detection method Methods 0.000 abstract description 4
- 230000001681 protective effect Effects 0.000 abstract description 2
- 238000013527 convolutional neural network Methods 0.000 description 16
- 238000011176 pooling Methods 0.000 description 14
- 230000006870 function Effects 0.000 description 12
- 230000007774 longterm Effects 0.000 description 10
- 210000004027 cell Anatomy 0.000 description 4
- 238000005259 measurement Methods 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 230000001133 acceleration Effects 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 2
- 238000010276 construction Methods 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 238000003745 diagnosis Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000011156 evaluation Methods 0.000 description 2
- 230000004807 localization Effects 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 230000002159 abnormal effect Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000002427 irreversible effect Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 230000000704 physical effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F30/00—Computer-aided design [CAD]
- G06F30/20—Design optimisation, verification or simulation
- G06F30/27—Design optimisation, verification or simulation using machine learning, e.g. artificial intelligence, neural networks, support vector machines [SVM] or training a model
-
- F—MECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
- F03—MACHINES OR ENGINES FOR LIQUIDS; WIND, SPRING, OR WEIGHT MOTORS; PRODUCING MECHANICAL POWER OR A REACTIVE PROPULSIVE THRUST, NOT OTHERWISE PROVIDED FOR
- F03D—WIND MOTORS
- F03D80/00—Details, components or accessories not provided for in groups F03D1/00 - F03D17/00
- F03D80/40—Ice detection; De-icing means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02E—REDUCTION OF GREENHOUSE GAS [GHG] EMISSIONS, RELATED TO ENERGY GENERATION, TRANSMISSION OR DISTRIBUTION
- Y02E10/00—Energy generation through renewable energy sources
- Y02E10/70—Wind energy
- Y02E10/72—Wind turbines with rotation axis in wind direction
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Software Systems (AREA)
- General Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Data Mining & Analysis (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Biomedical Technology (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Medical Informatics (AREA)
- Computer Hardware Design (AREA)
- Geometry (AREA)
- Sustainable Development (AREA)
- Sustainable Energy (AREA)
- Chemical & Material Sciences (AREA)
- Combustion & Propulsion (AREA)
- Mechanical Engineering (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
The invention relates to a fan blade icing fault prediction method based on a double attention mechanism, which is used for solving the problem that the icing state of a fan is difficult to predict in the prior art by combining a CNN + LSTM network with the double attention mechanism, and specifically comprises the following steps: acquiring an icing original data set of a fan; preprocessing an original data set to obtain a training set and a test set; the CNN network combines an attention mechanism to extract features; further extracting time information and predicting the icing state of the fan by using an LSTM network in combination with an attention mechanism; and predicting the icing of the fan blade by using the trained model. The method of the invention predicts the icing phenomenon of the fan blade by utilizing a large amount of time sequence detection variables acquired by the SCADA system, can predict when the blade will be iced before the blade is iced, takes protective measures in advance to prevent the blade from being iced, and plays a greater role in the safe, reliable and continuous operation of the wind power generation system.
Description
Technical Field
The invention relates to a fan blade icing fault prediction method based on a double attention mechanism, and belongs to the field of industrial system fault prediction.
Background
Wind power generation is currently the most mature and potentially renewable energy technology. The development of wind power in China draws attention. However, the particularity of wind energy acquisition determines that a large number of fans need to be arranged in cold areas with high latitude and high altitude. The fan working in cold areas is affected by weather conditions such as frost ice, rime, wet snow and the like, so that the blade icing phenomenon is easy to occur, and a series of consequences are caused. The wind energy capture capacity is reduced, the power generation power is lost, the fan blade is broken, and even safety accidents are caused.
Therefore, the blade icing fault can be predicted and prevented in time, and the method has important significance for prolonging the service life of the wind power equipment and preventing major safety accidents. In actual operation, severe icing can generally be easily detected and automatically deiced by the fan deicing system. However, severe icing, once formed, has an irreversible effect on blade performance. At present, one of the main methods is to use a sensor to acquire blade temperature data to judge whether the blade has an early icing phenomenon or not, and then to process the blade through a deicing system, but the method has two obvious disadvantages: firstly, need additionally to arrange the sensor and lead to the cost to fly to rise, secondly often cause the damage to the blade when detecting the blade and freeze, can extend blade life-span to a certain extent but the effect is limited. In addition, a data-driven method is adopted, machine learning and deep learning algorithm models are applied, the complex nonlinear relation among signals of the industrial equipment is usually extracted by using industrial equipment operation data acquired by an SCADA system, abnormal measurement values are searched for diagnosing the frozen fan blade, the data-based diagnosis method can effectively detect the frozen blade, and the cost of a wind field is greatly reduced. Although the early icing phenomenon of the fan blade can be accurately detected so as to maintain the blade, the icing still influences the physical parameters of the fan blade and causes certain influence on fan power generation equipment. The method of the invention predicts the icing phenomenon of the fan blade by utilizing a large amount of time sequence detection variables collected by the SCADA system and combining a double attention mechanism based on the CNN + LSTM network, can predict when the blade will be iced before the blade is iced, takes protective measures in advance to prevent the blade from being iced, and plays a greater role in the safe, reliable and continuous operation of the wind power generation system.
Disclosure of Invention
The invention aims to provide a fan blade icing fault prediction method based on a double attention mechanism, which has the innovation points that two different attention mechanisms are used for respectively extracting the characteristics of time and channel dimensions in data, and a fault prediction model with higher accuracy is provided for solving the problem that in a common blade icing diagnosis method based on data driving, the detection can be carried out only after blades are iced, the blade icing cannot be prevented in advance, so that the physical properties of the blades and a wind driven generator are influenced. The method includes the steps that firstly, preprocessed data are input into a CNN network, local features are learned through the CNN, an attention layer is added to help the CNN network focus on the importance degree of local information in the data, high-dimensional features output by the CNN network pass through an LSTM network, the dependency relationship among time of sequence data is learned through an LSTM neural network, the importance of the data in a corresponding time period in icing forecasting is measured through the attention layer, and finally whether a fan blade in the next time period is iced or not is forecasted through data information in historical time. The method can adaptively extract fault characteristics and time sequence information from the original data acquired by the SCADA system, can adaptively focus on characteristics of time and channel dimensions in the fan blade data, and can predict the icing state of the fan blade more accurately.
In order to achieve the purpose, the technical scheme adopted by the invention is a fan blade icing fault prediction method based on a double attention mechanism: as shown in fig. 1, the method is implemented as follows:
step (1): acquiring an icing original data set of a fan
The original data set is from a fan icing data set of 'the first China industry big data competition', the data is acquired from an industrial SCADA system, the total length is 2 months, the data comprises about 58 ten thousand pieces of data, each piece of data comprises 28 dimensions, the characteristic dimensions comprise wind speed, generator rotating speed, net side active power, wind direction angle, each blade angle, variable pitch motor temperature and the like, and the data are subjected to standardization processing. The time in the data set is in minutes as the minimum time unit, and each minute contains about 6-10 measurement sets.
Step (2): preprocessing a data set
Dividing original data into normal data, namely positive samples, labels are normal, fault data, namely negative samples, labels are frozen, and invalid data, namely label-free data, according to the time periods of freezing and non-freezing in the data; when constructing the training set, the non-labeled invalid data is first deleted. Taking several groups of measured values in the same unit time by taking each minute as a unit time, expressing the characteristics by using a plurality of statistics, and performing characteristic expansion by using m statistical representations for each characteristic quantity. Obtaining data of (m, n) -dimension for each group of unit time intervals, i.e.M represents the number of statistical representations, and n represents the number of features. The data after pretreatment are divided into a training set, a verification set and a test set according to the proportion of 0.7: 0.1: 0.2.
And (3): feature extraction is performed using CNN, convolution layers are convolved with training set data through convolution kernels and generate features, and then pooling layers extract the most important local features. And then the importance of different features in the extracted features is concerned through an attention layer. Here, Location-based Attention is used, focusing on the concealment vector itself.
Input data is divided into x ═ x per unit time1,x2,......,xN]And N represents the sequence length. The convolution kernel of the convolution operation is defined as w. Feature map of convolutional layer when inputting data xCan be expressed as:
wherein, b represents an offset amount,it is shown that the activation function is,can be seen as the convolution kernel w at the corresponding input xiThe feature vector of the upper learning, l, indicates the l-th network. The ReLU is used as the activation function, the advantages that the output of a part of neurons is 0, the sparsity of the network is improved, the interdependence relation of parameters is reduced, and overfitting is relieved. The ReLU formula is expressed as follows:
the pooling layer is used to perform downsampling, which in the present method is in the form of max pooling:
down (-) represents the downsampling function of the pooling layer,an output feature vector representing the pooling layer,is the feature vector of the previous convolution layer and s is the pool size.
Attention in CNN focuses on the feature vector itself, and the designed attention model is as follows:
tanh (-) represents the activation function,represents the weight of the attention layer, piRepresenting input data xiHigh-dimensional representation after feature extraction by convolution pooling, battIndicating the offset αiThe attention weight is expressed and can be considered as the importance of the corresponding feature vector in all features, and c represents the feature dimension. The attentions are then distributed with a fully connected layer.
vi=∑i=1αipi(6)
Where v isiRepresents x after the original data is subjected to feature extraction by using CNN + attributeiIs represented in high dimension.
And (4): and (4) continuing the network construction in the step (3) by using the LSTM, wherein the structure of the LSTM is mainly divided into three main stages: forgetting of information, long-term status update, and short-term status update. In conjunction with localization-based Attention, context-related information is captured to help predict the future.
Firstly, an information forgetting stage: forget door ftReading the previous short-term state ht-1And the current state xt(where v is represented by the high dimension obtained in step 3t) And goes one unit c upwards through sigmoid functiont-1Outputting a value between 0 and 1.
ft=σ(wf·ht-1+wf·vt+bf) (7)
Wherein (w)f,bf) Is the weight vector and offset value of the forgetting gate, σ is the sigmoid function, vtRepresenting the raw data x at time t obtained by step 3iHigh dimensional feature v of (i ═ t)i(i=t)。
Long-term state (Long-term state) update phase: it is determined which new information is stored in a long-term cell state. Input door itDetermining updatesValue of (1), input unit gtInputting the current state:
it=σ(wi·ht-1+wi·vt+bi) (8)
gt=tanh(wg·ht-1+wg·vt+bg) (9)
a new cell state c can then be obtainedt.
Short-term state (Short-term state) update phase: through an output gate otPerforming filter on the long-term state to obtain a short-term state ht。
ot=σ(wo·ht-i+wo·vt+bo) (11)
In the failure prediction task, the final goal is to be according to x1To xtFan state prediction xt+1State of (i) yt. An attention mechanism is utilized to capture context-related information to help predict the future. First, the current state h is calculatedtWeights with all previous states, wherein the current state htAnd the previous ith state hiWeight α oftiThe specific calculation is as follows:
wherein, thetaαIs the parameter to be learned, wαIs a weight matrix.
And then α as described abovetiGet attention vector αtThe specific calculation formula is as follows:
αi=Soft max([αt1,αt2,...,αti]) (14)
finally α according to the attention vectortAnd hidden state h1To ht-1Calculating a vector v containing the context attention informationt:
A prediction stage: binding context vector vtAnd a current hidden state htInformation of two vectors, generating one vectorThis vector contains both historical information passed over the LSTM network and information on the importance of the data for the corresponding time period in predicting icing.
wcIs a weight matrix. Attention vectorThe fan state (icing or not) for the t +1 time period is generated by softmax layer input, and is defined as:
wsis a weight matrix, bsIs the offset.
An objective function: calculating the true state ytAnd predicted stateBinary cross entropy between.
The network continuously modifies and updates the parameter values and the connection weights of neurons in each layer through forward propagation and backward propagation, and the verification set is used for determining the optimal values of the parameters to minimize the error value until an iteration stop condition is met, so that a trained model is obtained.
And (3) testing the model by using the test set subjected to pretreatment in the step (2), and selecting Accuracy (Accuracy) as an evaluation criterion to verify the prediction Accuracy of the model.
And (5) predicting the fault by using the trained model.
And acquiring a data set to be predicted from the SCADA system, inputting the data set to be predicted into a model, and predicting whether the fan is frozen in a future time period according to historical fan state information.
The invention has the beneficial effects that:
the method utilizes a large amount of detection variables collected by the SCADA system to predict the icing of the fan blade by combining a convolutional neural network and a long-short term memory network with a double attention mechanism, and predicts when the icing occurs before the blade is iced, so that protection measures are taken in advance to prevent the blade from icing, and a greater effect is achieved on the safe, reliable and continuous operation of a wind power generation system. The method includes the steps that firstly, preprocessed data are input into a CNN network, local features are learned through the CNN, an attention layer is added to help the CNN network focus on the importance degree of local information in the data, high-dimensional features output by the CNN network pass through an LSTM network, the dependency relationship among time of sequence data is learned through an LSTM neural network, the importance of the data in a corresponding time period in icing forecasting is measured through the attention layer, and finally whether a fan blade in the next time period is iced or not is forecasted through data information in historical time. The method can adaptively extract fault characteristics and time sequence information from the original data acquired by the SCADA system, can adaptively focus on characteristics of time and channel dimensions in the fan blade data, and can predict the icing state of the fan blade more accurately.
Drawings
Fig. 1 is an overall flowchart.
Fig. 2 is a flow chart of the training process.
FIG. 3 is a view showing a model structure
Detailed Description
The method adopts a fan blade icing prediction method based on a double-attention machine system, and the implementation process of the method is as follows:
step (1): acquiring an icing original data set of a fan
The original data set is from a fan icing data set of 'the first China industry big data competition', the data is acquired from an industrial SCADA system, the total length is 2 months, the data comprises about 58 ten thousand pieces of data, each piece of data comprises 28 dimensions, the characteristic dimensions comprise wind speed, generator rotating speed, net side active power, wind direction angle, each blade angle, variable pitch motor temperature and the like, and the data are subjected to standardization processing. The time in the data set is in minutes as the minimum time unit, and each minute contains about 6-10 measurement sets.
Step (2): preprocessing a data set
Dividing original data into normal data, namely positive samples, labels are normal, fault data, namely negative samples, labels are frozen, and invalid data, namely label-free data, according to the time periods of freezing and non-freezing in the data; when constructing the training set, the non-labeled invalid data is first deleted. Taking several groups of measured values in the same unit time as a unit time every minute, replacing original measured data with a Mean value (Mean), a Variance (Sample Variance), a Standard Deviation (Standard development), a Coefficient of Variation (CV), a Standard error (Standard error), a Skewness (Skewness), a Kurtosis (Kurtosis), a median (Mean) and a Quartile (Quartile) for the several groups of measured values in each unit time, and performing feature expansion by using m statistical representations for each feature quantity. Obtaining data of (m, n) -dimension for each group of unit time intervals, i.e.m represents the number of statistical representations (here m is 9), and n represents the number of features. The data after pretreatment are divided into a training set, a verification set and a test set according to the proportion of 0.7: 0.1: 0.2.
And (3): feature extraction is performed using CNN, convolution layers are convolved with training set data through convolution kernels and generate features, and then pooling layers extract the most important local features. And then the importance of different features in the extracted features is concerned through an attention layer. Here, Location-based Attention is used, focusing on the concealment vector itself.
Input data is divided into x ═ x per unit time1,x2,......,xN]And N represents the sequence length. The convolution kernel of the convolution operation is defined as w. Feature map of convolutional layer when inputting data xCan be expressed as:
wherein, b represents an offset amount,it is shown that the activation function is,can be seen as the convolution kernel w at the corresponding input xiThe feature vector of the upper learning, l, indicates the l-th network. The ReLU is used as the activation function, the advantages that the output of a part of neurons is 0, the sparsity of the network is improved, the interdependence relation of parameters is reduced, and overfitting is relieved. The ReLU formula is expressed as follows:
the pooling layer is used to perform downsampling, which in the present method is in the form of max pooling:
down (-) represents the downsampling function of the pooling layer,an output feature vector representing the pooling layer,is the feature vector of the previous convolution layer and s is the pool size.
Attention in CNN focuses on the feature vector itself, and the designed attention model is as follows:
tanh (-) represents the activation function,represents the weight of the attention layer, piRepresenting input data xiHigh-dimensional representation after feature extraction by convolution pooling, battIndicating the offset αiThe attention weight is expressed and can be considered as the importance of the corresponding feature vector in all features, and c represents the feature dimension. The attentions are then distributed with a fully connected layer.
vi=∑i=1αipi(6)
Where v isiRepresents x after the original data is subjected to feature extraction by using CNN + attributeiIs represented in high dimension.
And (4): and (4) continuing the network construction in the step (3) by using the LSTM, wherein the structure of the LSTM is mainly divided into three main stages: forgetting of information, long-term status update, and short-term status update. In conjunction with localization-based Attention, context-related information is captured to help predict the future.
Firstly, an information forgetting stage: forget door ftReading the previous short-term state ht-1And the current state xt(obtained here by step 3)High-dimensional representation of vt) And goes one unit c upwards through sigmoid functiont-1Outputting a value between 0 and 1.
ft=σ(wf·ht-1+wf·vt+bf) (7)
Wherein (w)f,bf) Is the weight vector and offset value of the forgetting gate, σ is the sigmoid function, vtRepresenting the raw data x at time t obtained by step 3iHigh dimensional feature v of (i ═ t)i(i=t)。
Long-term state (Long-term state) update phase: it is determined which new information is stored in a long-term cell state. Input door itDetermining updated values, input unit gtInputting the current state:
it=σ(wi·ht-1+wi·vt+bi) (8)
gt=tanh(wg·ht-1+wg·vt+bg) (9)
a new cell state c can then be obtainedt.
Short-term state (Short-term state) update phase: through an output gate otPerforming filter on the long-term state to obtain a short-term state ht。
ot=σ(wo·ht-1+wo·vt+bo) (11)
In the failure prediction task, the final goal is to be according to x1To xtFan state prediction xt+1State of (i) yt. An attention mechanism is utilized to capture context-related information to help predict the future. In the task of the failure prediction,the final goal is according to x1To xtFan state prediction xt+1State of (i) yt. An attention mechanism is utilized to capture context-related information to help predict the future. First, the current state h is calculatedtWeights with all previous states, wherein the current state htAnd the previous ith state hiWeight α oftiThe specific calculation is as follows:
wherein, thetaαIs the parameter to be learned, wαIs a weight matrix.
And then α as described abovetiGet attention vector αtThe specific calculation formula is as follows:
αi=Soft max([αt1,αt2,...,αti]) (14)
finally α according to the attention vectortAnd hidden state h1To ht-1Calculating a vector v containing the context attention informationt:
A prediction stage: binding context vector vtAnd a current hidden state htInformation of two vectors, generating one vectorThis vector contains both historical information passed over the LSTM network and information on the importance of the data for the corresponding time period in predicting icing.
wcIs a weight matrix, attention vectorThe fan state (icing or not) for the t +1 time period is generated by softmax layer input, and is defined as:
wsis a weight matrix, bsIs the offset.
An objective function: calculating the true state ytAnd predicted stateBinary cross entropy between.
The technical scheme of the invention is described in the aspect of network structure as follows: the input of the convolution network is a matrix of 26x9, and the convolution kernel is set to be 2x9 through convolution → pooling → convolution → pooling, and then the LSTM network is set to have an LSTM hidden layer characteristic number of 128 and an output layer of 2. The network continuously modifies and updates the parameter values and the connection weights of the neurons in each layer through forward propagation and backward propagation. And the verification set is used for determining the optimal value of the parameter, the iteration times are set to be 100 times, the learning rate is set to be 0.001, and the error value is minimized until the iteration stop condition is met to obtain the trained model.
And (3) testing the model by using the test set subjected to pretreatment in the step (2), and selecting Accuracy (Accuracy) as an evaluation criterion to verify the prediction Accuracy of the model.
And (5) detecting early icing faults of the blades by using the trained model.
The data set to be predicted is collected from an SCADA system of an actual wind force field, wherein the data set to be predicted comprises a timestamp, a wind speed, a generator rotating speed, net side active power (kw), a wind angle (°), a 25-second average wind direction angle, a yaw position, a yaw speed, a blade 1 angle, a blade 2 angle, a blade 3 angle, a blade 1 speed, a blade 2 speed, a blade 3 speed, a pitch motor 1 temperature, a pitch motor 2 temperature, a pitch motor 3 temperature, an x-direction acceleration, a y-direction acceleration, an environment temperature, a cabin temperature, an ng 51 temperature, an ng 52 temperature, an ng 53 temperature, an ng 51 charger direct current, an ng 52 charger direct current and an ng 53 charger direct current.
And inputting the data set to be predicted into the model, and predicting whether the fan is frozen in the future time period according to historical fan state information.
Claims (7)
1. The method for predicting the icing fault of the fan blade based on the double attention mechanism is characterized by comprising the following steps of:
step (1): acquiring an icing original data set of a fan;
step (2): preprocessing an original data set to obtain a training set and a test set;
and (3): performing local feature extraction on the preprocessed data by using the CNN; then, focusing attention on the importance of different features in the extracted local features through an attention layer to obtain a local feature vector containing attention weight;
and (4): inputting the feature vector containing attention weight obtained in the step (3), and extracting the feature h of the time dimension through LSTMtCapturing context related information through an attention mechanism, paying attention to feature importance of different time, and finally forming a prediction model;
and (5): and predicting the icing fault of the fan blade.
And acquiring a data set to be predicted from the SCADA system, inputting the data set to be predicted into a model, and predicting whether the fan is frozen in a future time period according to historical fan state information.
2. The dual attention mechanism-based wind turbine blade icing fault prediction method of claim 1, wherein: the freezing original data of the fan comprises but is not limited to wind speed, rotating speed of a generator, net side active power, wind direction angle, angle of each blade and temperature of a variable pitch motor.
3. The dual attention mechanism-based wind turbine blade icing fault prediction method of claim 1, wherein: the pretreatment comprises the following steps: dividing original data into normal data, namely positive samples, labels are normal, fault data, namely negative samples, labels are frozen, and invalid data, namely label-free data, according to the time periods of freezing and non-freezing in the data; when constructing the training set, the non-labeled invalid data is first deleted. Taking several groups of measured values in the same unit time by taking each minute as a unit time, expressing the characteristics by using a plurality of statistics, and performing characteristic expansion by using m statistical representations for each characteristic quantity. Obtaining data of (m, n) -dimension for each group of unit time intervals, i.e.m represents the number of statistical representations and n represents the number of features. And dividing the data after preprocessing into a training set, a verification set and a test set.
4. The dual attention mechanism-based wind turbine blade icing fault prediction method of claim 1, wherein: the attention model in the step (3) is as follows:
tanh (-) represents the activation function,represents the weight of the attention layer, piRepresenting input data xiHigh-dimensional representation after feature extraction by CNN, battDenotes an offset amount, αiExpressing attention weight, expressing the importance degree of the corresponding feature vector in all features, and c expressing feature dimension;
and then the attentions are distributed by using a full connection layer,
vi=∑i=1αipi(6)
where v isiRepresents x after the original data is subjected to feature extraction by using CNN + attributeiIs represented in high dimension.
5. The method for predicting the icing fault of the fan blade based on the dual attention mechanism according to claim 1, wherein the attention mechanism in the step 4 is specifically as follows:
first, the current state h is calculatedtWeights with all previous states, wherein the current state htAnd the previous ith state hiWeight α oftiThe specific calculation is as follows:
wherein, thetaαIs the parameter to be learned, wαIs a weight matrix.
And then α as described abovetiGet attention vector αtThe specific calculation formula is as follows:
αi=Softmax([αt1,αt2,...,αti]) (14)
finally α according to the attention vectortAnd hidden state h1To ht-1Calculating a vector v containing the context attention informationt:
6. The method for predicting the icing fault of the fan blade based on the dual attention mechanism according to claim 1, wherein the prediction model in the step 4 is specifically as follows:
binding context vector vtAnd a current hidden state htInformation of two vectors, generating one vectorThis vector contains both historical information passed over the LSTM network, and information on the importance of the data for the corresponding time period in predicting icing,
wcis a weight matrix, vectorThe input through softmax layer, the state of the fan generating t +1 time period, i.e. whether the fan is frozen or not, is defined as:
wsis a weight matrix, bsIs the offset.
7. The dual attention mechanism-based wind turbine blade icing fault prediction method according to claim 1, wherein the objective function of the prediction model in step 4 is:
the network continuously modifies and updates the parameter values and the connection weights of neurons in each layer through forward propagation and backward propagation, and the verification set is used for determining the optimal values of the parameters to minimize the error value until an iteration stop condition is met, so that a trained model is obtained.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010551118.9A CN111680454A (en) | 2020-06-16 | 2020-06-16 | Fan blade icing fault prediction method based on double attention mechanism |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010551118.9A CN111680454A (en) | 2020-06-16 | 2020-06-16 | Fan blade icing fault prediction method based on double attention mechanism |
Publications (1)
Publication Number | Publication Date |
---|---|
CN111680454A true CN111680454A (en) | 2020-09-18 |
Family
ID=72455202
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010551118.9A Pending CN111680454A (en) | 2020-06-16 | 2020-06-16 | Fan blade icing fault prediction method based on double attention mechanism |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111680454A (en) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112200032A (en) * | 2020-09-28 | 2021-01-08 | 辽宁石油化工大学 | Attention mechanism-based high-voltage circuit breaker mechanical state online monitoring method |
CN112633317A (en) * | 2020-11-02 | 2021-04-09 | 国能信控互联技术有限公司 | CNN-LSTM fan fault prediction method and system based on attention mechanism |
CN112734131A (en) * | 2021-01-22 | 2021-04-30 | 国家电投集团四川电力有限公司 | Fan blade icing state prediction method based on deep learning algorithm |
CN113323823A (en) * | 2021-06-08 | 2021-08-31 | 云南大学 | AWKELM-based fan blade icing fault detection method and system |
CN113327033A (en) * | 2021-05-28 | 2021-08-31 | 广西电网有限责任公司电力科学研究院 | Power distribution network fault diagnosis method and system |
CN113688822A (en) * | 2021-09-07 | 2021-11-23 | 河南工业大学 | Time sequence attention mechanism scene image identification method |
CN114165392A (en) * | 2021-11-03 | 2022-03-11 | 华能射阳新能源发电有限公司 | Wind turbine generator set power abnormity diagnosis method and device and storage medium |
CN114330881A (en) * | 2021-12-29 | 2022-04-12 | 南京邮电大学 | Data-driven fan blade icing prediction method and device |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109751206A (en) * | 2019-02-25 | 2019-05-14 | 宜通世纪物联网研究院(广州)有限公司 | fan blade icing failure prediction method, device and storage medium |
CN109802430A (en) * | 2018-12-29 | 2019-05-24 | 上海电力学院 | A kind of wind-powered electricity generation power grid control method based on LSTM-Attention network |
CN109886492A (en) * | 2019-02-26 | 2019-06-14 | 浙江鑫升新能源科技有限公司 | Photovoltaic power generation power prediction model and its construction method based on Attention LSTM |
CN110909919A (en) * | 2019-11-07 | 2020-03-24 | 哈尔滨工程大学 | Photovoltaic power prediction method of depth neural network model with attention mechanism fused |
CN111144499A (en) * | 2019-12-27 | 2020-05-12 | 北京工业大学 | Fan blade early icing fault detection method based on deep neural network |
-
2020
- 2020-06-16 CN CN202010551118.9A patent/CN111680454A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109802430A (en) * | 2018-12-29 | 2019-05-24 | 上海电力学院 | A kind of wind-powered electricity generation power grid control method based on LSTM-Attention network |
CN109751206A (en) * | 2019-02-25 | 2019-05-14 | 宜通世纪物联网研究院(广州)有限公司 | fan blade icing failure prediction method, device and storage medium |
CN109886492A (en) * | 2019-02-26 | 2019-06-14 | 浙江鑫升新能源科技有限公司 | Photovoltaic power generation power prediction model and its construction method based on Attention LSTM |
CN110909919A (en) * | 2019-11-07 | 2020-03-24 | 哈尔滨工程大学 | Photovoltaic power prediction method of depth neural network model with attention mechanism fused |
CN111144499A (en) * | 2019-12-27 | 2020-05-12 | 北京工业大学 | Fan blade early icing fault detection method based on deep neural network |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112200032A (en) * | 2020-09-28 | 2021-01-08 | 辽宁石油化工大学 | Attention mechanism-based high-voltage circuit breaker mechanical state online monitoring method |
CN112200032B (en) * | 2020-09-28 | 2023-05-30 | 辽宁石油化工大学 | Online monitoring method for mechanical state of high-voltage circuit breaker based on attention mechanism |
CN112633317A (en) * | 2020-11-02 | 2021-04-09 | 国能信控互联技术有限公司 | CNN-LSTM fan fault prediction method and system based on attention mechanism |
CN112734131A (en) * | 2021-01-22 | 2021-04-30 | 国家电投集团四川电力有限公司 | Fan blade icing state prediction method based on deep learning algorithm |
CN113327033A (en) * | 2021-05-28 | 2021-08-31 | 广西电网有限责任公司电力科学研究院 | Power distribution network fault diagnosis method and system |
CN113323823A (en) * | 2021-06-08 | 2021-08-31 | 云南大学 | AWKELM-based fan blade icing fault detection method and system |
CN113688822A (en) * | 2021-09-07 | 2021-11-23 | 河南工业大学 | Time sequence attention mechanism scene image identification method |
CN114165392A (en) * | 2021-11-03 | 2022-03-11 | 华能射阳新能源发电有限公司 | Wind turbine generator set power abnormity diagnosis method and device and storage medium |
CN114330881A (en) * | 2021-12-29 | 2022-04-12 | 南京邮电大学 | Data-driven fan blade icing prediction method and device |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111680454A (en) | Fan blade icing fault prediction method based on double attention mechanism | |
Xiang et al. | Condition monitoring and anomaly detection of wind turbine based on cascaded and bidirectional deep learning networks | |
CN111144499A (en) | Fan blade early icing fault detection method based on deep neural network | |
Kreutz et al. | Machine learning-based icing prediction on wind turbines | |
CN109284863A (en) | A kind of power equipment temperature predicting method based on deep neural network | |
Chen et al. | Learning deep representation for blades icing fault detection of wind turbines | |
Remadna et al. | Leveraging the power of the combination of CNN and bi-directional LSTM networks for aircraft engine RUL estimation | |
CN110472684B (en) | Method and device for monitoring icing of fan blade and readable storage medium | |
Guo et al. | Wind turbine blade icing detection with multi-model collaborative monitoring method | |
WO2024045246A1 (en) | Spike echo state network model for aero engine fault prediction | |
CN115809405A (en) | Fan main shaft gear box temperature anomaly detection method based on multi-feature fusion | |
CN112832960A (en) | Fan blade icing detection method based on deep learning and storage medium | |
CN111814849B (en) | DA-RNN-based wind turbine generator set key component fault early warning method | |
Chen et al. | Prediction of icing fault of wind turbine blades based on deep learning | |
CN116738868A (en) | Rolling bearing residual life prediction method | |
CN116021981A (en) | Method, device, equipment and storage medium for predicting ice coating faults of power distribution network line | |
CN117272753A (en) | Power transmission line icing galloping early warning method based on process parameters | |
CN118188342A (en) | Fan-oriented fault early warning and life prediction method | |
CN112215281A (en) | Fan blade icing fault detection method | |
Peter et al. | Wind turbine generator prognostics using field SCADA data | |
CN114962186A (en) | Method for predicting icing of wind turbine generator blade | |
Song et al. | Anomaly detection of wind turbine generator based on temporal information | |
CN114049014A (en) | Method, device and system for evaluating operation state of offshore wind turbine generator | |
Zhenhao et al. | Prediction of wind power ramp events based on deep neural network | |
Wei et al. | Research on transmission line trip prediction based on logistic regression algorithm under icing condition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20200918 |