CN111680454A - Fan blade icing fault prediction method based on double attention mechanism - Google Patents

Fan blade icing fault prediction method based on double attention mechanism Download PDF

Info

Publication number
CN111680454A
CN111680454A CN202010551118.9A CN202010551118A CN111680454A CN 111680454 A CN111680454 A CN 111680454A CN 202010551118 A CN202010551118 A CN 202010551118A CN 111680454 A CN111680454 A CN 111680454A
Authority
CN
China
Prior art keywords
data
attention
fan
icing
attention mechanism
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010551118.9A
Other languages
Chinese (zh)
Inventor
朱玉婷
于海阳
杨震
郑忠斌
王朝栋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Technology
Original Assignee
Beijing University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Technology filed Critical Beijing University of Technology
Priority to CN202010551118.9A priority Critical patent/CN111680454A/en
Publication of CN111680454A publication Critical patent/CN111680454A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F30/00Computer-aided design [CAD]
    • G06F30/20Design optimisation, verification or simulation
    • G06F30/27Design optimisation, verification or simulation using machine learning, e.g. artificial intelligence, neural networks, support vector machines [SVM] or training a model
    • FMECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
    • F03MACHINES OR ENGINES FOR LIQUIDS; WIND, SPRING, OR WEIGHT MOTORS; PRODUCING MECHANICAL POWER OR A REACTIVE PROPULSIVE THRUST, NOT OTHERWISE PROVIDED FOR
    • F03DWIND MOTORS
    • F03D80/00Details, components or accessories not provided for in groups F03D1/00 - F03D17/00
    • F03D80/40Ice detection; De-icing means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02EREDUCTION OF GREENHOUSE GAS [GHG] EMISSIONS, RELATED TO ENERGY GENERATION, TRANSMISSION OR DISTRIBUTION
    • Y02E10/00Energy generation through renewable energy sources
    • Y02E10/70Wind energy
    • Y02E10/72Wind turbines with rotation axis in wind direction

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Medical Informatics (AREA)
  • Computer Hardware Design (AREA)
  • Geometry (AREA)
  • Sustainable Development (AREA)
  • Sustainable Energy (AREA)
  • Chemical & Material Sciences (AREA)
  • Combustion & Propulsion (AREA)
  • Mechanical Engineering (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention relates to a fan blade icing fault prediction method based on a double attention mechanism, which is used for solving the problem that the icing state of a fan is difficult to predict in the prior art by combining a CNN + LSTM network with the double attention mechanism, and specifically comprises the following steps: acquiring an icing original data set of a fan; preprocessing an original data set to obtain a training set and a test set; the CNN network combines an attention mechanism to extract features; further extracting time information and predicting the icing state of the fan by using an LSTM network in combination with an attention mechanism; and predicting the icing of the fan blade by using the trained model. The method of the invention predicts the icing phenomenon of the fan blade by utilizing a large amount of time sequence detection variables acquired by the SCADA system, can predict when the blade will be iced before the blade is iced, takes protective measures in advance to prevent the blade from being iced, and plays a greater role in the safe, reliable and continuous operation of the wind power generation system.

Description

Fan blade icing fault prediction method based on double attention mechanism
Technical Field
The invention relates to a fan blade icing fault prediction method based on a double attention mechanism, and belongs to the field of industrial system fault prediction.
Background
Wind power generation is currently the most mature and potentially renewable energy technology. The development of wind power in China draws attention. However, the particularity of wind energy acquisition determines that a large number of fans need to be arranged in cold areas with high latitude and high altitude. The fan working in cold areas is affected by weather conditions such as frost ice, rime, wet snow and the like, so that the blade icing phenomenon is easy to occur, and a series of consequences are caused. The wind energy capture capacity is reduced, the power generation power is lost, the fan blade is broken, and even safety accidents are caused.
Therefore, the blade icing fault can be predicted and prevented in time, and the method has important significance for prolonging the service life of the wind power equipment and preventing major safety accidents. In actual operation, severe icing can generally be easily detected and automatically deiced by the fan deicing system. However, severe icing, once formed, has an irreversible effect on blade performance. At present, one of the main methods is to use a sensor to acquire blade temperature data to judge whether the blade has an early icing phenomenon or not, and then to process the blade through a deicing system, but the method has two obvious disadvantages: firstly, need additionally to arrange the sensor and lead to the cost to fly to rise, secondly often cause the damage to the blade when detecting the blade and freeze, can extend blade life-span to a certain extent but the effect is limited. In addition, a data-driven method is adopted, machine learning and deep learning algorithm models are applied, the complex nonlinear relation among signals of the industrial equipment is usually extracted by using industrial equipment operation data acquired by an SCADA system, abnormal measurement values are searched for diagnosing the frozen fan blade, the data-based diagnosis method can effectively detect the frozen blade, and the cost of a wind field is greatly reduced. Although the early icing phenomenon of the fan blade can be accurately detected so as to maintain the blade, the icing still influences the physical parameters of the fan blade and causes certain influence on fan power generation equipment. The method of the invention predicts the icing phenomenon of the fan blade by utilizing a large amount of time sequence detection variables collected by the SCADA system and combining a double attention mechanism based on the CNN + LSTM network, can predict when the blade will be iced before the blade is iced, takes protective measures in advance to prevent the blade from being iced, and plays a greater role in the safe, reliable and continuous operation of the wind power generation system.
Disclosure of Invention
The invention aims to provide a fan blade icing fault prediction method based on a double attention mechanism, which has the innovation points that two different attention mechanisms are used for respectively extracting the characteristics of time and channel dimensions in data, and a fault prediction model with higher accuracy is provided for solving the problem that in a common blade icing diagnosis method based on data driving, the detection can be carried out only after blades are iced, the blade icing cannot be prevented in advance, so that the physical properties of the blades and a wind driven generator are influenced. The method includes the steps that firstly, preprocessed data are input into a CNN network, local features are learned through the CNN, an attention layer is added to help the CNN network focus on the importance degree of local information in the data, high-dimensional features output by the CNN network pass through an LSTM network, the dependency relationship among time of sequence data is learned through an LSTM neural network, the importance of the data in a corresponding time period in icing forecasting is measured through the attention layer, and finally whether a fan blade in the next time period is iced or not is forecasted through data information in historical time. The method can adaptively extract fault characteristics and time sequence information from the original data acquired by the SCADA system, can adaptively focus on characteristics of time and channel dimensions in the fan blade data, and can predict the icing state of the fan blade more accurately.
In order to achieve the purpose, the technical scheme adopted by the invention is a fan blade icing fault prediction method based on a double attention mechanism: as shown in fig. 1, the method is implemented as follows:
step (1): acquiring an icing original data set of a fan
The original data set is from a fan icing data set of 'the first China industry big data competition', the data is acquired from an industrial SCADA system, the total length is 2 months, the data comprises about 58 ten thousand pieces of data, each piece of data comprises 28 dimensions, the characteristic dimensions comprise wind speed, generator rotating speed, net side active power, wind direction angle, each blade angle, variable pitch motor temperature and the like, and the data are subjected to standardization processing. The time in the data set is in minutes as the minimum time unit, and each minute contains about 6-10 measurement sets.
Step (2): preprocessing a data set
Dividing original data into normal data, namely positive samples, labels are normal, fault data, namely negative samples, labels are frozen, and invalid data, namely label-free data, according to the time periods of freezing and non-freezing in the data; when constructing the training set, the non-labeled invalid data is first deleted. Taking several groups of measured values in the same unit time by taking each minute as a unit time, expressing the characteristics by using a plurality of statistics, and performing characteristic expansion by using m statistical representations for each characteristic quantity. Obtaining data of (m, n) -dimension for each group of unit time intervals, i.e.
Figure BDA0002541304530000035
M represents the number of statistical representations, and n represents the number of features. The data after pretreatment are divided into a training set, a verification set and a test set according to the proportion of 0.7: 0.1: 0.2.
And (3): feature extraction is performed using CNN, convolution layers are convolved with training set data through convolution kernels and generate features, and then pooling layers extract the most important local features. And then the importance of different features in the extracted features is concerned through an attention layer. Here, Location-based Attention is used, focusing on the concealment vector itself.
Input data is divided into x ═ x per unit time1,x2,......,xN]And N represents the sequence length. The convolution kernel of the convolution operation is defined as w. Feature map of convolutional layer when inputting data x
Figure BDA0002541304530000031
Can be expressed as:
Figure BDA0002541304530000032
wherein, b represents an offset amount,
Figure BDA0002541304530000033
it is shown that the activation function is,
Figure BDA0002541304530000034
can be seen as the convolution kernel w at the corresponding input xiThe feature vector of the upper learning, l, indicates the l-th network. The ReLU is used as the activation function, the advantages that the output of a part of neurons is 0, the sparsity of the network is improved, the interdependence relation of parameters is reduced, and overfitting is relieved. The ReLU formula is expressed as follows:
Figure BDA0002541304530000041
the pooling layer is used to perform downsampling, which in the present method is in the form of max pooling:
Figure BDA0002541304530000042
down (-) represents the downsampling function of the pooling layer,
Figure BDA0002541304530000043
an output feature vector representing the pooling layer,
Figure BDA0002541304530000044
is the feature vector of the previous convolution layer and s is the pool size.
Attention in CNN focuses on the feature vector itself, and the designed attention model is as follows:
Figure BDA0002541304530000045
Figure BDA0002541304530000046
tanh (-) represents the activation function,
Figure BDA0002541304530000047
represents the weight of the attention layer, piRepresenting input data xiHigh-dimensional representation after feature extraction by convolution pooling, battIndicating the offset αiThe attention weight is expressed and can be considered as the importance of the corresponding feature vector in all features, and c represents the feature dimension. The attentions are then distributed with a fully connected layer.
vi=∑i=1αipi(6)
Where v isiRepresents x after the original data is subjected to feature extraction by using CNN + attributeiIs represented in high dimension.
And (4): and (4) continuing the network construction in the step (3) by using the LSTM, wherein the structure of the LSTM is mainly divided into three main stages: forgetting of information, long-term status update, and short-term status update. In conjunction with localization-based Attention, context-related information is captured to help predict the future.
Firstly, an information forgetting stage: forget door ftReading the previous short-term state ht-1And the current state xt(where v is represented by the high dimension obtained in step 3t) And goes one unit c upwards through sigmoid functiont-1Outputting a value between 0 and 1.
ft=σ(wf·ht-1+wf·vt+bf) (7)
Wherein (w)f,bf) Is the weight vector and offset value of the forgetting gate, σ is the sigmoid function, vtRepresenting the raw data x at time t obtained by step 3iHigh dimensional feature v of (i ═ t)i(i=t)。
Long-term state (Long-term state) update phase: it is determined which new information is stored in a long-term cell state. Input door itDetermining updatesValue of (1), input unit gtInputting the current state:
it=σ(wi·ht-1+wi·vt+bi) (8)
gt=tanh(wg·ht-1+wg·vt+bg) (9)
a new cell state c can then be obtainedt.
Figure BDA0002541304530000051
Short-term state (Short-term state) update phase: through an output gate otPerforming filter on the long-term state to obtain a short-term state ht
ot=σ(wo·ht-i+wo·vt+bo) (11)
Figure BDA0002541304530000052
In the failure prediction task, the final goal is to be according to x1To xtFan state prediction xt+1State of (i) yt. An attention mechanism is utilized to capture context-related information to help predict the future. First, the current state h is calculatedtWeights with all previous states, wherein the current state htAnd the previous ith state hiWeight α oftiThe specific calculation is as follows:
Figure BDA0002541304530000053
wherein, thetaαIs the parameter to be learned, wαIs a weight matrix.
And then α as described abovetiGet attention vector αtThe specific calculation formula is as follows:
αi=Soft max([αt1,αt2,...,αti]) (14)
finally α according to the attention vectortAnd hidden state h1To ht-1Calculating a vector v containing the context attention informationt
Figure BDA0002541304530000061
A prediction stage: binding context vector vtAnd a current hidden state htInformation of two vectors, generating one vector
Figure BDA0002541304530000062
This vector contains both historical information passed over the LSTM network and information on the importance of the data for the corresponding time period in predicting icing.
Figure BDA0002541304530000063
wcIs a weight matrix. Attention vector
Figure BDA0002541304530000064
The fan state (icing or not) for the t +1 time period is generated by softmax layer input, and is defined as:
Figure BDA0002541304530000065
wsis a weight matrix, bsIs the offset.
An objective function: calculating the true state ytAnd predicted state
Figure BDA0002541304530000066
Binary cross entropy between.
Figure BDA0002541304530000067
The network continuously modifies and updates the parameter values and the connection weights of neurons in each layer through forward propagation and backward propagation, and the verification set is used for determining the optimal values of the parameters to minimize the error value until an iteration stop condition is met, so that a trained model is obtained.
And (3) testing the model by using the test set subjected to pretreatment in the step (2), and selecting Accuracy (Accuracy) as an evaluation criterion to verify the prediction Accuracy of the model.
And (5) predicting the fault by using the trained model.
And acquiring a data set to be predicted from the SCADA system, inputting the data set to be predicted into a model, and predicting whether the fan is frozen in a future time period according to historical fan state information.
The invention has the beneficial effects that:
the method utilizes a large amount of detection variables collected by the SCADA system to predict the icing of the fan blade by combining a convolutional neural network and a long-short term memory network with a double attention mechanism, and predicts when the icing occurs before the blade is iced, so that protection measures are taken in advance to prevent the blade from icing, and a greater effect is achieved on the safe, reliable and continuous operation of a wind power generation system. The method includes the steps that firstly, preprocessed data are input into a CNN network, local features are learned through the CNN, an attention layer is added to help the CNN network focus on the importance degree of local information in the data, high-dimensional features output by the CNN network pass through an LSTM network, the dependency relationship among time of sequence data is learned through an LSTM neural network, the importance of the data in a corresponding time period in icing forecasting is measured through the attention layer, and finally whether a fan blade in the next time period is iced or not is forecasted through data information in historical time. The method can adaptively extract fault characteristics and time sequence information from the original data acquired by the SCADA system, can adaptively focus on characteristics of time and channel dimensions in the fan blade data, and can predict the icing state of the fan blade more accurately.
Drawings
Fig. 1 is an overall flowchart.
Fig. 2 is a flow chart of the training process.
FIG. 3 is a view showing a model structure
Detailed Description
The method adopts a fan blade icing prediction method based on a double-attention machine system, and the implementation process of the method is as follows:
step (1): acquiring an icing original data set of a fan
The original data set is from a fan icing data set of 'the first China industry big data competition', the data is acquired from an industrial SCADA system, the total length is 2 months, the data comprises about 58 ten thousand pieces of data, each piece of data comprises 28 dimensions, the characteristic dimensions comprise wind speed, generator rotating speed, net side active power, wind direction angle, each blade angle, variable pitch motor temperature and the like, and the data are subjected to standardization processing. The time in the data set is in minutes as the minimum time unit, and each minute contains about 6-10 measurement sets.
Step (2): preprocessing a data set
Dividing original data into normal data, namely positive samples, labels are normal, fault data, namely negative samples, labels are frozen, and invalid data, namely label-free data, according to the time periods of freezing and non-freezing in the data; when constructing the training set, the non-labeled invalid data is first deleted. Taking several groups of measured values in the same unit time as a unit time every minute, replacing original measured data with a Mean value (Mean), a Variance (Sample Variance), a Standard Deviation (Standard development), a Coefficient of Variation (CV), a Standard error (Standard error), a Skewness (Skewness), a Kurtosis (Kurtosis), a median (Mean) and a Quartile (Quartile) for the several groups of measured values in each unit time, and performing feature expansion by using m statistical representations for each feature quantity. Obtaining data of (m, n) -dimension for each group of unit time intervals, i.e.
Figure BDA0002541304530000081
m represents the number of statistical representations (here m is 9), and n represents the number of features. The data after pretreatment are divided into a training set, a verification set and a test set according to the proportion of 0.7: 0.1: 0.2.
And (3): feature extraction is performed using CNN, convolution layers are convolved with training set data through convolution kernels and generate features, and then pooling layers extract the most important local features. And then the importance of different features in the extracted features is concerned through an attention layer. Here, Location-based Attention is used, focusing on the concealment vector itself.
Input data is divided into x ═ x per unit time1,x2,......,xN]And N represents the sequence length. The convolution kernel of the convolution operation is defined as w. Feature map of convolutional layer when inputting data x
Figure BDA0002541304530000091
Can be expressed as:
Figure BDA0002541304530000092
wherein, b represents an offset amount,
Figure BDA00025413045300000911
it is shown that the activation function is,
Figure BDA0002541304530000093
can be seen as the convolution kernel w at the corresponding input xiThe feature vector of the upper learning, l, indicates the l-th network. The ReLU is used as the activation function, the advantages that the output of a part of neurons is 0, the sparsity of the network is improved, the interdependence relation of parameters is reduced, and overfitting is relieved. The ReLU formula is expressed as follows:
Figure BDA0002541304530000094
the pooling layer is used to perform downsampling, which in the present method is in the form of max pooling:
Figure BDA0002541304530000095
down (-) represents the downsampling function of the pooling layer,
Figure BDA0002541304530000096
an output feature vector representing the pooling layer,
Figure BDA0002541304530000097
is the feature vector of the previous convolution layer and s is the pool size.
Attention in CNN focuses on the feature vector itself, and the designed attention model is as follows:
Figure BDA0002541304530000098
Figure BDA0002541304530000099
tanh (-) represents the activation function,
Figure BDA00025413045300000910
represents the weight of the attention layer, piRepresenting input data xiHigh-dimensional representation after feature extraction by convolution pooling, battIndicating the offset αiThe attention weight is expressed and can be considered as the importance of the corresponding feature vector in all features, and c represents the feature dimension. The attentions are then distributed with a fully connected layer.
vi=∑i=1αipi(6)
Where v isiRepresents x after the original data is subjected to feature extraction by using CNN + attributeiIs represented in high dimension.
And (4): and (4) continuing the network construction in the step (3) by using the LSTM, wherein the structure of the LSTM is mainly divided into three main stages: forgetting of information, long-term status update, and short-term status update. In conjunction with localization-based Attention, context-related information is captured to help predict the future.
Firstly, an information forgetting stage: forget door ftReading the previous short-term state ht-1And the current state xt(obtained here by step 3)High-dimensional representation of vt) And goes one unit c upwards through sigmoid functiont-1Outputting a value between 0 and 1.
ft=σ(wf·ht-1+wf·vt+bf) (7)
Wherein (w)f,bf) Is the weight vector and offset value of the forgetting gate, σ is the sigmoid function, vtRepresenting the raw data x at time t obtained by step 3iHigh dimensional feature v of (i ═ t)i(i=t)。
Long-term state (Long-term state) update phase: it is determined which new information is stored in a long-term cell state. Input door itDetermining updated values, input unit gtInputting the current state:
it=σ(wi·ht-1+wi·vt+bi) (8)
gt=tanh(wg·ht-1+wg·vt+bg) (9)
a new cell state c can then be obtainedt.
Figure BDA0002541304530000101
Short-term state (Short-term state) update phase: through an output gate otPerforming filter on the long-term state to obtain a short-term state ht
ot=σ(wo·ht-1+wo·vt+bo) (11)
Figure BDA0002541304530000102
In the failure prediction task, the final goal is to be according to x1To xtFan state prediction xt+1State of (i) yt. An attention mechanism is utilized to capture context-related information to help predict the future. In the task of the failure prediction,the final goal is according to x1To xtFan state prediction xt+1State of (i) yt. An attention mechanism is utilized to capture context-related information to help predict the future. First, the current state h is calculatedtWeights with all previous states, wherein the current state htAnd the previous ith state hiWeight α oftiThe specific calculation is as follows:
Figure BDA0002541304530000111
wherein, thetaαIs the parameter to be learned, wαIs a weight matrix.
And then α as described abovetiGet attention vector αtThe specific calculation formula is as follows:
αi=Soft max([αt1,αt2,...,αti]) (14)
finally α according to the attention vectortAnd hidden state h1To ht-1Calculating a vector v containing the context attention informationt
Figure BDA0002541304530000112
A prediction stage: binding context vector vtAnd a current hidden state htInformation of two vectors, generating one vector
Figure BDA0002541304530000113
This vector contains both historical information passed over the LSTM network and information on the importance of the data for the corresponding time period in predicting icing.
Figure BDA0002541304530000114
wcIs a weight matrix, attention vector
Figure BDA0002541304530000115
The fan state (icing or not) for the t +1 time period is generated by softmax layer input, and is defined as:
Figure BDA0002541304530000116
wsis a weight matrix, bsIs the offset.
An objective function: calculating the true state ytAnd predicted state
Figure BDA0002541304530000117
Binary cross entropy between.
Figure BDA0002541304530000118
The technical scheme of the invention is described in the aspect of network structure as follows: the input of the convolution network is a matrix of 26x9, and the convolution kernel is set to be 2x9 through convolution → pooling → convolution → pooling, and then the LSTM network is set to have an LSTM hidden layer characteristic number of 128 and an output layer of 2. The network continuously modifies and updates the parameter values and the connection weights of the neurons in each layer through forward propagation and backward propagation. And the verification set is used for determining the optimal value of the parameter, the iteration times are set to be 100 times, the learning rate is set to be 0.001, and the error value is minimized until the iteration stop condition is met to obtain the trained model.
And (3) testing the model by using the test set subjected to pretreatment in the step (2), and selecting Accuracy (Accuracy) as an evaluation criterion to verify the prediction Accuracy of the model.
And (5) detecting early icing faults of the blades by using the trained model.
The data set to be predicted is collected from an SCADA system of an actual wind force field, wherein the data set to be predicted comprises a timestamp, a wind speed, a generator rotating speed, net side active power (kw), a wind angle (°), a 25-second average wind direction angle, a yaw position, a yaw speed, a blade 1 angle, a blade 2 angle, a blade 3 angle, a blade 1 speed, a blade 2 speed, a blade 3 speed, a pitch motor 1 temperature, a pitch motor 2 temperature, a pitch motor 3 temperature, an x-direction acceleration, a y-direction acceleration, an environment temperature, a cabin temperature, an ng 51 temperature, an ng 52 temperature, an ng 53 temperature, an ng 51 charger direct current, an ng 52 charger direct current and an ng 53 charger direct current.
And inputting the data set to be predicted into the model, and predicting whether the fan is frozen in the future time period according to historical fan state information.

Claims (7)

1. The method for predicting the icing fault of the fan blade based on the double attention mechanism is characterized by comprising the following steps of:
step (1): acquiring an icing original data set of a fan;
step (2): preprocessing an original data set to obtain a training set and a test set;
and (3): performing local feature extraction on the preprocessed data by using the CNN; then, focusing attention on the importance of different features in the extracted local features through an attention layer to obtain a local feature vector containing attention weight;
and (4): inputting the feature vector containing attention weight obtained in the step (3), and extracting the feature h of the time dimension through LSTMtCapturing context related information through an attention mechanism, paying attention to feature importance of different time, and finally forming a prediction model;
and (5): and predicting the icing fault of the fan blade.
And acquiring a data set to be predicted from the SCADA system, inputting the data set to be predicted into a model, and predicting whether the fan is frozen in a future time period according to historical fan state information.
2. The dual attention mechanism-based wind turbine blade icing fault prediction method of claim 1, wherein: the freezing original data of the fan comprises but is not limited to wind speed, rotating speed of a generator, net side active power, wind direction angle, angle of each blade and temperature of a variable pitch motor.
3. The dual attention mechanism-based wind turbine blade icing fault prediction method of claim 1, wherein: the pretreatment comprises the following steps: dividing original data into normal data, namely positive samples, labels are normal, fault data, namely negative samples, labels are frozen, and invalid data, namely label-free data, according to the time periods of freezing and non-freezing in the data; when constructing the training set, the non-labeled invalid data is first deleted. Taking several groups of measured values in the same unit time by taking each minute as a unit time, expressing the characteristics by using a plurality of statistics, and performing characteristic expansion by using m statistical representations for each characteristic quantity. Obtaining data of (m, n) -dimension for each group of unit time intervals, i.e.
Figure FDA0002541304520000021
m represents the number of statistical representations and n represents the number of features. And dividing the data after preprocessing into a training set, a verification set and a test set.
4. The dual attention mechanism-based wind turbine blade icing fault prediction method of claim 1, wherein: the attention model in the step (3) is as follows:
Figure FDA0002541304520000022
Figure FDA0002541304520000023
tanh (-) represents the activation function,
Figure FDA0002541304520000024
represents the weight of the attention layer, piRepresenting input data xiHigh-dimensional representation after feature extraction by CNN, battDenotes an offset amount, αiExpressing attention weight, expressing the importance degree of the corresponding feature vector in all features, and c expressing feature dimension;
and then the attentions are distributed by using a full connection layer,
vi=∑i=1αipi(6)
where v isiRepresents x after the original data is subjected to feature extraction by using CNN + attributeiIs represented in high dimension.
5. The method for predicting the icing fault of the fan blade based on the dual attention mechanism according to claim 1, wherein the attention mechanism in the step 4 is specifically as follows:
first, the current state h is calculatedtWeights with all previous states, wherein the current state htAnd the previous ith state hiWeight α oftiThe specific calculation is as follows:
Figure FDA0002541304520000025
wherein, thetaαIs the parameter to be learned, wαIs a weight matrix.
And then α as described abovetiGet attention vector αtThe specific calculation formula is as follows:
αi=Softmax([αt1,αt2,...,αti]) (14)
finally α according to the attention vectortAnd hidden state h1To ht-1Calculating a vector v containing the context attention informationt
Figure FDA0002541304520000031
6. The method for predicting the icing fault of the fan blade based on the dual attention mechanism according to claim 1, wherein the prediction model in the step 4 is specifically as follows:
binding context vector vtAnd a current hidden state htInformation of two vectors, generating one vector
Figure FDA0002541304520000032
This vector contains both historical information passed over the LSTM network, and information on the importance of the data for the corresponding time period in predicting icing,
Figure FDA0002541304520000033
wcis a weight matrix, vector
Figure FDA0002541304520000034
The input through softmax layer, the state of the fan generating t +1 time period, i.e. whether the fan is frozen or not, is defined as:
Figure FDA0002541304520000035
wsis a weight matrix, bsIs the offset.
7. The dual attention mechanism-based wind turbine blade icing fault prediction method according to claim 1, wherein the objective function of the prediction model in step 4 is:
calculating the true state ytAnd predicted state
Figure FDA0002541304520000036
The binary cross-entropy between the two is,
Figure FDA0002541304520000037
the network continuously modifies and updates the parameter values and the connection weights of neurons in each layer through forward propagation and backward propagation, and the verification set is used for determining the optimal values of the parameters to minimize the error value until an iteration stop condition is met, so that a trained model is obtained.
CN202010551118.9A 2020-06-16 2020-06-16 Fan blade icing fault prediction method based on double attention mechanism Pending CN111680454A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010551118.9A CN111680454A (en) 2020-06-16 2020-06-16 Fan blade icing fault prediction method based on double attention mechanism

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010551118.9A CN111680454A (en) 2020-06-16 2020-06-16 Fan blade icing fault prediction method based on double attention mechanism

Publications (1)

Publication Number Publication Date
CN111680454A true CN111680454A (en) 2020-09-18

Family

ID=72455202

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010551118.9A Pending CN111680454A (en) 2020-06-16 2020-06-16 Fan blade icing fault prediction method based on double attention mechanism

Country Status (1)

Country Link
CN (1) CN111680454A (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112200032A (en) * 2020-09-28 2021-01-08 辽宁石油化工大学 Attention mechanism-based high-voltage circuit breaker mechanical state online monitoring method
CN112633317A (en) * 2020-11-02 2021-04-09 国能信控互联技术有限公司 CNN-LSTM fan fault prediction method and system based on attention mechanism
CN112734131A (en) * 2021-01-22 2021-04-30 国家电投集团四川电力有限公司 Fan blade icing state prediction method based on deep learning algorithm
CN113323823A (en) * 2021-06-08 2021-08-31 云南大学 AWKELM-based fan blade icing fault detection method and system
CN113327033A (en) * 2021-05-28 2021-08-31 广西电网有限责任公司电力科学研究院 Power distribution network fault diagnosis method and system
CN113688822A (en) * 2021-09-07 2021-11-23 河南工业大学 Time sequence attention mechanism scene image identification method
CN114165392A (en) * 2021-11-03 2022-03-11 华能射阳新能源发电有限公司 Wind turbine generator set power abnormity diagnosis method and device and storage medium
CN114330881A (en) * 2021-12-29 2022-04-12 南京邮电大学 Data-driven fan blade icing prediction method and device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109751206A (en) * 2019-02-25 2019-05-14 宜通世纪物联网研究院(广州)有限公司 fan blade icing failure prediction method, device and storage medium
CN109802430A (en) * 2018-12-29 2019-05-24 上海电力学院 A kind of wind-powered electricity generation power grid control method based on LSTM-Attention network
CN109886492A (en) * 2019-02-26 2019-06-14 浙江鑫升新能源科技有限公司 Photovoltaic power generation power prediction model and its construction method based on Attention LSTM
CN110909919A (en) * 2019-11-07 2020-03-24 哈尔滨工程大学 Photovoltaic power prediction method of depth neural network model with attention mechanism fused
CN111144499A (en) * 2019-12-27 2020-05-12 北京工业大学 Fan blade early icing fault detection method based on deep neural network

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109802430A (en) * 2018-12-29 2019-05-24 上海电力学院 A kind of wind-powered electricity generation power grid control method based on LSTM-Attention network
CN109751206A (en) * 2019-02-25 2019-05-14 宜通世纪物联网研究院(广州)有限公司 fan blade icing failure prediction method, device and storage medium
CN109886492A (en) * 2019-02-26 2019-06-14 浙江鑫升新能源科技有限公司 Photovoltaic power generation power prediction model and its construction method based on Attention LSTM
CN110909919A (en) * 2019-11-07 2020-03-24 哈尔滨工程大学 Photovoltaic power prediction method of depth neural network model with attention mechanism fused
CN111144499A (en) * 2019-12-27 2020-05-12 北京工业大学 Fan blade early icing fault detection method based on deep neural network

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112200032A (en) * 2020-09-28 2021-01-08 辽宁石油化工大学 Attention mechanism-based high-voltage circuit breaker mechanical state online monitoring method
CN112200032B (en) * 2020-09-28 2023-05-30 辽宁石油化工大学 Online monitoring method for mechanical state of high-voltage circuit breaker based on attention mechanism
CN112633317A (en) * 2020-11-02 2021-04-09 国能信控互联技术有限公司 CNN-LSTM fan fault prediction method and system based on attention mechanism
CN112734131A (en) * 2021-01-22 2021-04-30 国家电投集团四川电力有限公司 Fan blade icing state prediction method based on deep learning algorithm
CN113327033A (en) * 2021-05-28 2021-08-31 广西电网有限责任公司电力科学研究院 Power distribution network fault diagnosis method and system
CN113323823A (en) * 2021-06-08 2021-08-31 云南大学 AWKELM-based fan blade icing fault detection method and system
CN113688822A (en) * 2021-09-07 2021-11-23 河南工业大学 Time sequence attention mechanism scene image identification method
CN114165392A (en) * 2021-11-03 2022-03-11 华能射阳新能源发电有限公司 Wind turbine generator set power abnormity diagnosis method and device and storage medium
CN114330881A (en) * 2021-12-29 2022-04-12 南京邮电大学 Data-driven fan blade icing prediction method and device

Similar Documents

Publication Publication Date Title
CN111680454A (en) Fan blade icing fault prediction method based on double attention mechanism
Xiang et al. Condition monitoring and anomaly detection of wind turbine based on cascaded and bidirectional deep learning networks
CN111144499A (en) Fan blade early icing fault detection method based on deep neural network
Kreutz et al. Machine learning-based icing prediction on wind turbines
CN109284863A (en) A kind of power equipment temperature predicting method based on deep neural network
Chen et al. Learning deep representation for blades icing fault detection of wind turbines
Remadna et al. Leveraging the power of the combination of CNN and bi-directional LSTM networks for aircraft engine RUL estimation
CN110472684B (en) Method and device for monitoring icing of fan blade and readable storage medium
Guo et al. Wind turbine blade icing detection with multi-model collaborative monitoring method
WO2024045246A1 (en) Spike echo state network model for aero engine fault prediction
CN115809405A (en) Fan main shaft gear box temperature anomaly detection method based on multi-feature fusion
CN112832960A (en) Fan blade icing detection method based on deep learning and storage medium
CN111814849B (en) DA-RNN-based wind turbine generator set key component fault early warning method
Chen et al. Prediction of icing fault of wind turbine blades based on deep learning
CN116738868A (en) Rolling bearing residual life prediction method
CN116021981A (en) Method, device, equipment and storage medium for predicting ice coating faults of power distribution network line
CN117272753A (en) Power transmission line icing galloping early warning method based on process parameters
CN118188342A (en) Fan-oriented fault early warning and life prediction method
CN112215281A (en) Fan blade icing fault detection method
Peter et al. Wind turbine generator prognostics using field SCADA data
CN114962186A (en) Method for predicting icing of wind turbine generator blade
Song et al. Anomaly detection of wind turbine generator based on temporal information
CN114049014A (en) Method, device and system for evaluating operation state of offshore wind turbine generator
Zhenhao et al. Prediction of wind power ramp events based on deep neural network
Wei et al. Research on transmission line trip prediction based on logistic regression algorithm under icing condition

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20200918