CN110647867B - Bearing fault diagnosis method and system based on self-adaptive anti-noise neural network - Google Patents

Bearing fault diagnosis method and system based on self-adaptive anti-noise neural network Download PDF

Info

Publication number
CN110647867B
CN110647867B CN201910957332.1A CN201910957332A CN110647867B CN 110647867 B CN110647867 B CN 110647867B CN 201910957332 A CN201910957332 A CN 201910957332A CN 110647867 B CN110647867 B CN 110647867B
Authority
CN
China
Prior art keywords
neural network
data set
network model
training
noise
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910957332.1A
Other languages
Chinese (zh)
Other versions
CN110647867A (en
Inventor
金国强
金�一
王浩璇
陈怀安
竺长安
陈恩红
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Science and Technology of China USTC
Original Assignee
University of Science and Technology of China USTC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Science and Technology of China USTC filed Critical University of Science and Technology of China USTC
Priority to CN201910957332.1A priority Critical patent/CN110647867B/en
Publication of CN110647867A publication Critical patent/CN110647867A/en
Application granted granted Critical
Publication of CN110647867B publication Critical patent/CN110647867B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/02Preprocessing
    • G06F2218/04Denoising
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01MTESTING STATIC OR DYNAMIC BALANCE OF MACHINES OR STRUCTURES; TESTING OF STRUCTURES OR APPARATUS, NOT OTHERWISE PROVIDED FOR
    • G01M13/00Testing of machine parts
    • G01M13/04Bearings
    • G01M13/045Acoustic or vibration analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/16Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Computational Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Pure & Applied Mathematics (AREA)
  • Molecular Biology (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Acoustics & Sound (AREA)
  • Algebra (AREA)
  • Databases & Information Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Signal Processing (AREA)
  • Testing Of Devices, Machine Parts, Or Other Structures Thereof (AREA)

Abstract

The invention discloses a bearing fault diagnosis method and a system based on a self-adaptive anti-noise neural network, wherein the method comprises the following steps: acquiring a data set, and dividing the data set into a training data set and a testing data set according to a preset proportion; constructing a neural network model, wherein the neural network model comprises the following steps: the system comprises an input layer, two convolutional layers, two gating recursion unit layers, an attention mechanism layer and a deep neural network layer; training the constructed neural network model based on training data in a noise-free training data set to obtain a trained neural network model; and adding additive white Gaussian noise into the test data in the test data set, and inputting the test data into the trained neural network model to obtain a bearing fault diagnosis result. The method can automatically extract the features from the original signals without manually selecting the features and denoising, can distinguish different fault types and fault severity degrees, and can adapt to different noise environments.

Description

Bearing fault diagnosis method and system based on self-adaptive anti-noise neural network
Technical Field
The invention relates to the technical field of bearing faults, in particular to a bearing fault diagnosis method and system based on a self-adaptive anti-noise neural network.
Background
The diagnosis of mechanical equipment failure, which causes economic loss and casualties, is a widespread concern in modern industries. Rolling bearings are important targets for fault diagnosis of mechanical equipment, and especially in rotating mechanical equipment, rolling bearing faults account for a large proportion of faults. In the past decades, the diagnosis of faults in rolling bearings has been extensively studied. The data-driven-based method is a common method in bearing fault diagnosis and is mainly divided into a diagnosis method based on signal analysis and a diagnosis method based on deep learning.
The bearing fault diagnosis method based on signal analysis mainly comprises the following steps: fourier analysis, wavelet transform, cepstrum, empirical mode decomposition, and the like. These diagnostic methods require expert knowledge for signal preprocessing and feature selection, and are less effective in complex environmental conditions.
The traditional bearing fault diagnosis method based on deep learning comprises the following steps: the system comprises an artificial neural network, a support vector machine, a stacked self-encoder, a deep Boltzmann machine, a deep belief network and the like, and has poor diagnosis effect in a complex environment, particularly a noise environment.
In recent years, with the development of deep neural network training optimization algorithms, various deep neural networks are beginning to be widely researched. CNNs (Convolutional Neural Networks) and RNNs (Recurrent Neural Networks) are the most widely used deep learning Networks at present, and have been widely used in various tasks including bearing fault diagnosis. CNNs use convolution filters to extract local features, and RNNs are widely used in time series data or sequence modeling to capture time-related features from signals. However, when the bearing fault diagnosis is carried out by early experiments of CNNs or RNNs, complicated manual operation is required, so that the fault diagnosis of the bearing under the complex noise environment is still a challenging task, and particularly under the condition of high noise, the adaptability of a neural network to different noise levels still needs to be improved.
Therefore, how to effectively diagnose the bearing fault is an urgent problem to be solved.
Disclosure of Invention
In view of the above, the invention provides a bearing fault diagnosis method based on a self-adaptive anti-noise neural network, which can automatically extract features from an original signal without manually selecting the features and denoising, can distinguish different fault types and fault severity degrees, and can adapt to different noise environments.
The invention provides a bearing fault diagnosis method based on a self-adaptive anti-noise neural network, which comprises the following steps:
acquiring a data set, and dividing the data set into a training data set and a testing data set according to a preset proportion;
constructing a neural network model, wherein the neural network model comprises: the system comprises an input layer, two convolutional layers, two gating recursion unit layers, an attention mechanism layer and a deep neural network layer;
training the constructed neural network model based on training data in a noise-free training data set to obtain a trained neural network model;
and adding additive white Gaussian noise into the test data in the test data set, and inputting the test data into the trained neural network model to obtain a bearing fault diagnosis result.
Preferably, the acquiring the data set and dividing the data set into a training data set and a testing data set according to a preset proportion includes:
acquiring a data set, and enabling the data set to be as follows: the scale of 2 is divided into a training data set and a test data set.
Preferably, two of the convolutional layers act as feature extractors using the ELU as an activation function.
Preferably, the training data in the noise-free training data set is used to train the constructed neural network model to obtain a trained neural network model, and the training includes:
inputting the signals of the training data in the noise-free training data set in a random sampling mode;
setting a hyper-parameter, a weight initialization mode and a loss function of a neural network model;
and training the neural network model by using a random gradient descent method and a back propagation algorithm to obtain the trained neural network model.
Preferably, the hyper-parameters comprise: learning rate, attenuation rate and training times of the neural network model.
A bearing fault diagnostic system based on an adaptive noise immune neural network, comprising:
the acquisition module is used for acquiring a data set and dividing the data set into a training data set and a testing data set according to a preset proportion;
a building module configured to build a neural network model, wherein the neural network model comprises: the system comprises an input layer, two convolutional layers, two gating recursion unit layers, an attention mechanism layer and a deep neural network layer;
the training module is used for training the built neural network model based on training data in the noiseless training data set to obtain a trained neural network model;
and the diagnosis module is used for adding additive white Gaussian noise into the test data in the test data set and inputting the test data into the trained neural network model to obtain a bearing fault diagnosis result.
Preferably, when the obtaining module executes obtaining of a data set and divides the data set into a training data set and a testing data set according to a preset proportion, the obtaining module is specifically configured to:
acquiring a data set, and enabling the data set to be as follows: the scale of 2 is divided into a training data set and a test data set.
Preferably, the two convolutional layers act as feature extractors using the ELU as an activation function.
Preferably, when the training module executes training data in the noise-free training data set to train the constructed neural network model, so as to obtain a trained neural network model, the training module is specifically configured to:
inputting the signals of the training data in the noise-free training data set in a random sampling mode;
setting a hyper-parameter, a weight initialization mode and a loss function of a neural network model;
and training the neural network model by using a random gradient descent method and a back propagation algorithm to obtain the trained neural network model.
Preferably, the hyper-parameters comprise: learning rate, attenuation rate and training times of the neural network model.
In summary, the invention discloses a bearing fault diagnosis method based on a self-adaptive anti-noise neural network, when a bearing fault needs to be diagnosed, a data set is firstly obtained, the data set is divided into a training data set and a testing data set according to a preset proportion, and then a neural network model is constructed, wherein the neural network model comprises: the system comprises an input layer, two convolutional layers, two gating recursion unit layers, an attention mechanism layer and a deep neural network layer; training the constructed neural network model based on training data in the noise-free training data set to obtain a trained neural network model; and adding additive white Gaussian noise into the test data in the test data set, and inputting the test data into the trained neural network model to obtain a bearing fault diagnosis result. The neural network obtained through training can automatically extract features from the original signal without manually selecting the features and denoising, can distinguish different fault types and fault severity degrees, and can adapt to different noise environments.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a flowchart of a bearing fault diagnosis method based on an adaptive anti-noise neural network according to embodiment 1 of the present disclosure;
FIG. 2 is a flowchart of a method of embodiment 2 of a bearing fault diagnosis method based on an adaptive anti-noise neural network disclosed in the present invention;
FIG. 3 is a schematic structural diagram of an embodiment 1 of a bearing fault diagnosis system based on an adaptive anti-noise neural network, which is disclosed by the invention;
FIG. 4 is a schematic structural diagram of an embodiment 2 of a bearing fault diagnosis system based on an adaptive anti-noise neural network disclosed in the present invention;
FIG. 5 is a schematic diagram of a CWRU data set selected by a neural network according to the present disclosure;
FIG. 6 is a schematic diagram of a neural network structure disclosed in the present invention;
FIG. 7 is a schematic diagram of an attention mechanism in a neural network according to the present disclosure;
fig. 8 is a schematic diagram of specific structural parameters of the neural network disclosed in the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
As shown in fig. 1, which is a flowchart of a method of embodiment 1 of a bearing fault diagnosis method based on an adaptive anti-noise neural network disclosed in the present invention, the method may include the following steps:
s101, acquiring a data set, and dividing the data set into a training data set and a testing data set according to a preset proportion;
when bearing faults need to be detected, firstly, a data set is obtained and divided into a training data set and a testing data set according to a preset proportion. And preprocessing and amplitude normalization are carried out on signals in the data set, and a data input strategy for improving the anti-noise performance of the network by adopting random sampling is adopted.
S102, constructing a neural network model, wherein the neural network model comprises the following steps: the system comprises an input layer, two convolutional layers, two gating recursion unit layers, an attention mechanism layer and a deep neural network layer;
then, a neural network model is established, and the typical structure of the neural network model mainly comprises an input layer, two convolution layers, two GRU layers, an attention mechanism layer and a DNN layer. The neural network model takes a signal with a fixed length as an input, and realizes random sampling of input data by using dropout at an input layer. After the input layers, the two convolutional layers act as feature extractors using the ELU as an activation function. After each convolution layer, there is a batch normalization layer and a dropout layer, respectively, for stability and versatility. Then, the CNN part is followed by two GRU layers, and then another dropout layer. Note that the mechanism layer follows the dropout layer. Finally, the classification result is generated and stabilized by the two fully connected layers and the softmax layer.
S103, training the constructed neural network model based on training data in the noise-free training data set to obtain a trained neural network model;
then, training all neural network models on the noise-free original data, testing the data added with different noise levels, inputting the original signals of the training data in the noise-free training data set in a random sampling mode, and performing model training to obtain the trained neural network models.
And S104, adding additive white Gaussian noise into the test data in the test data set, and inputting the test data into the trained neural network model to obtain a bearing fault diagnosis result.
And finally, adding additive white Gaussian noise into the test data in the test data set, and inputting the test data into the trained neural network model to obtain a bearing fault diagnosis result.
In summary, in the above embodiment, when a bearing fault needs to be diagnosed, a data set is first obtained, and the data set is divided into a training data set and a testing data set according to a preset proportion, and then a neural network model is constructed, where the neural network model includes: the system comprises an input layer, two convolutional layers, two gating recursion unit layers, an attention mechanism layer and a deep neural network layer; training the constructed neural network model based on training data in the noise-free training data set to obtain a trained neural network model; and adding additive white Gaussian noise into the test data in the test data set, and inputting the test data into the trained neural network model to obtain a bearing fault diagnosis result. The neural network obtained through training can automatically extract features from the original signal without manually selecting the features and denoising, can distinguish different fault types and fault severity degrees, and can adapt to different noise environments.
As shown in fig. 2, which is a flowchart of a method of embodiment 2 of the bearing fault diagnosis method based on the adaptive anti-noise neural network disclosed in the present invention, the method may include the following steps:
s201, acquiring a data set, and enabling the data set to be as 8:2 into a training data set and a test data set;
this example uses two different data sets, one from the bearing data center at the university of Keiss Sige (CWRU) and the second from the bearing fault diagnosis test platform QPZZ-II. Through bearing fault diagnosis experiments, four types of bearing vibration signals are collected. The two data sets were divided into a training data set and a test data set on an 8:2 scale.
Laboratory table for CWRU data set as shown in fig. 5, the vibration signal used was collected from the b-side of the accelerometer driver at a 48000Hz sampling rate. The motor load is 3hp, and the motor rotating speed is changed between 1719 rpm and 1729rpm under different bearing health states. In the embodiment, the average rotating speed of the motor is 1724rpm, and the driving end bearing is a deep groove ball bearing (6205-2RS JEM SKF). The test bearings have four states: normal state, inner ring fault, rolling element fault and outer ring fault. Each fault condition has three types of fault diameters, 7mils, 14mils, and 21mils (1 mil-0.001 inch), which are produced by electrical discharge machining. Thus, there is a normal condition and three fault types, each with three severity levels.
The signal of the data set is preprocessed, since the input length of the algorithm is a fixed value, while the vibration signal is periodic, it is necessary to segment the signal according to the period and normalize the amplitude of the input signal to [ -1, +1 ].
S202, constructing a neural network model, wherein the neural network model comprises the following steps: the system comprises an input layer, two convolutional layers, two gating recursion unit layers, an attention mechanism layer and a deep neural network layer;
as shown in fig. 6, the typical structure of the neural network model is mainly composed of one input layer, two convolutional layers, two GRU layers, an attention mechanism layer, and a DNN layer. The neural network model takes a signal with a fixed length as an input, and realizes random sampling of input data by using dropout at an input layer. After the input layers, the two convolutional layers act as feature extractors, using the ELU as an activation function, the formula is as follows:
Figure BDA0002226762390000081
after each convolution layer, there is a batch normalization layer and a dropout layer, respectively, for stability and versatility. Then, following the CNN part by two GRU layers, a Gated Recursion Unit (GRU) is a modified RNN that can capture the long-term dependence of the gate structure. These gates are used to delete or add information to the hidden state to decide whether to remember long term dependencies or to use short term information. The work flow of the GRU is as follows: reset gate r at the t-th time steptThe formula is as follows:
rt=σ(Wrxt+Urht-1+br);
wherein xtAs an input vector, WrAnd UrAs a weight matrix, brTo bias the weights, ht-1Is the previous activation. σ is a logistic sigmoid function. Reset door rtDetermining how many past activations to retain in computing candidate activations
Figure BDA0002226762390000082
Figure BDA0002226762390000083
Figure BDA0002226762390000084
Wherein |, indicates an element multiplication. Candidate activation
Figure BDA0002226762390000085
From an input vector xtAnd a preceding step ht-1Calculated by resetting the gate rtAnd (5) modulating.
zt=σ(Wzxt+Uzht-1+bz);
Updating the door ztControlling the ratio of past activations and candidate activations to calculate a new activation:
Figure BDA0002226762390000086
the GRU layer is followed by another dropout layer. Note that the mechanism layer follows the dropout layer.
The attention mechanism obtains a final result by adopting a weighting method for each time step, and the utilization rate of information is improved. Note that the main calculation process of the institution is shown in FIG. 7, and that the weight vectors are calculated as follows: given all hidden source states
Figure BDA0002226762390000091
Current target state htCalculating an attention weight vector at:
Figure BDA0002226762390000092
Wherein the function score compares the current target hidden state htAnd per source hidden state
Figure BDA0002226762390000093
The contribution of each time step to the final output is calculated, as:
Figure BDA0002226762390000094
wherein, WSIs a trainable weighting matrix. Then, by noticing the weight vector atAnd each step
Figure BDA0002226762390000095
Computing context vector ctI.e., weighted time step vector:
Figure BDA0002226762390000096
finally, the context vector ctWith the target hidden state htConnected to obtain an output attention-hiding state
Figure BDA0002226762390000097
Wherein WcFor a trainable weighting matrix:
Figure BDA0002226762390000098
finally, the classification result is generated and stabilized by the two fully connected layers and the softmax layer. Specific structural parameters are shown in fig. 8.
The Dropout rate is fixed to 0.5. The convolution kernel size is 128 x 1 and the number of first and second layers of filters is 64 and 72, respectively. The step size of the convolutional layer is 1, and no padding is used for the convolutional layer. GRU first layer neurons 64, second layer neurons 128. There are 128 neurons in the hidden state vector of note. Then two fully connected layers, 64 and 32 neurons respectively. Finally, a softmax layer with 10 neurons.
S203, inputting a signal of training data in a noise-free training data set in a random sampling mode, setting a hyper-parameter, a weight initialization mode and a loss function of a neural network model, and training the neural network model by using a random gradient descent method and a back propagation algorithm to obtain a trained neural network model;
training all neural network models on noiseless original data, testing data added with different noise levels, inputting original signals of training data in the noiseless training data set in a random sampling mode, setting hyper-parameters such as learning rate, attenuation rate and training times of the neural network models, and using an Adam optimization algorithm as an optimizer. And (3) adopting a power attenuation learning rate, setting an initial value of 0.0001 and a weight of 0.9, setting a weight initialization mode, setting a loss function of the network model, training the model by using a back propagation algorithm, and storing the trained model weight.
And S204, adding additive white Gaussian noise into the test data in the test data set, and inputting the test data into the trained neural network model to obtain a bearing fault diagnosis result.
Preparing a test data set, adding additive white Gaussian noise to the test data in the test data set, inputting the test data into a trained neural network model, simulating different noise conditions, and finally measuring by a signal-to-noise ratio (SNR), wherein the SNR is defined as the ratio of the power of a meaningful signal to the power of background noise [63 ]. The decibel form (dB) of the signal-to-noise ratio is expressed as:
Figure BDA0002226762390000101
wherein, P is the average power of the periodic signal c (t) in the period t, and is defined as:
Figure BDA0002226762390000102
when the signal is in discrete form, the average power can be calculated as:
Figure BDA0002226762390000103
for a signal with zero mean and known variance, the power can be measured by the variance
Figure BDA0002226762390000111
Indicating that, therefore, for standard normally distributed noise, the power is 1. Thus, the power of the original signal is first calculated, and then the noise signal P generated at the desired signal-to-noise ratio is calculatednoiseOf the power of (c). Finally, an additive white gaussian noise is generated by the following formula and then added to the original signal to make the signal have a desired signal-to-noise ratio. The standard normally distributed noise can be generated by the following equation:
Figure BDA0002226762390000112
wherein randn represents a function of a random number or matrix that produces a standard normal distribution;
and finally, obtaining a bearing fault diagnosis result.
In summary, the neural network architecture of the present invention mainly comprises four parts: random sample data input, CNN-based enhanced feature extractor, GRU-based feature classifier, attention-based mechanism and DNN-based feature post-processing. The input signal adopts a simple random sampling strategy, so that the adaptability of the network to noise is improved. In the feature extraction part, the convolutional layer adopts an exponential linear unit activation function and a dropout to improve the adaptability of the neural network under the noise condition. The function of the feature extractor is similar to that of a time-domain frequency-domain converter, a noise reduction filter and a feature selector, the feature extractor is similar to that of the time-domain frequency-domain converter, the noise reduction filter, the feature selector and the like, features can be automatically extracted from noise signals without manual feature selection and denoising processes, and the feature extractor has strong adaptability to noise. The features conveyed by the CNN part are further learned and classified using the famous GRU as a decoder or classifier when processing time-varying sequences. Finally, using the attention mechanism and DNN component as a post processor, appropriate features are selected from the GRU output as the final classification result. The neural network can distinguish not only different fault types, but also severity of corresponding faults; the method has stronger processing adaptability to signals with large noise and better universality to noise signals with different grades.
As shown in fig. 3, which is a schematic structural diagram of an embodiment 1 of the bearing fault diagnosis system based on the adaptive anti-noise neural network disclosed in the present invention, the system may include:
an obtaining module 301, configured to obtain a data set, and divide the data set into a training data set and a testing data set according to a preset ratio;
when bearing faults need to be detected, firstly, a data set is obtained and divided into a training data set and a testing data set according to a preset proportion. And preprocessing and amplitude normalization are carried out on signals in the data set, and a data input strategy for improving the anti-noise performance of the network by adopting random sampling is adopted.
A building module 302, configured to build a neural network model, where the neural network model includes: the system comprises an input layer, two convolutional layers, two gating recursion unit layers, an attention mechanism layer and a deep neural network layer;
then, a neural network model is established, and the typical structure of the neural network model mainly comprises an input layer, two convolution layers, two GRU layers, an attention mechanism layer and a DNN layer. The neural network model takes a signal with a fixed length as an input, and realizes random sampling of input data by using dropout at an input layer. After the input layers, the two convolutional layers act as feature extractors using the ELU as an activation function. After each convolution layer, there is a batch normalization layer and a dropout layer, respectively, for stability and versatility. Then, the CNN part is followed by two GRU layers, and then another dropout layer. Note that the mechanism layer follows the dropout layer. Finally, the classification result is generated and stabilized by the two fully connected layers and the softmax layer.
The training module 303 is configured to train the constructed neural network model based on training data in the noise-free training data set to obtain a trained neural network model;
then, training all neural network models on the noise-free original data, testing the data added with different noise levels, inputting the original signals of the training data in the noise-free training data set in a random sampling mode, and performing model training to obtain the trained neural network models.
And the diagnosis module 304 is used for adding additive white gaussian noise into the test data in the test data set, inputting the additive white gaussian noise into the trained neural network model, and obtaining a bearing fault diagnosis result.
And finally, adding additive white Gaussian noise into the test data in the test data set, and inputting the test data into the trained neural network model to obtain a bearing fault diagnosis result.
In summary, in the above embodiments, when a bearing fault needs to be diagnosed, a data set is first obtained, and the data set is divided into a training data set and a testing data set according to a preset proportion, and then a neural network model is constructed, where the neural network model includes: the system comprises an input layer, two convolutional layers, two gating recursion unit layers, an attention mechanism layer and a deep neural network layer; training the constructed neural network model based on training data in the noise-free training data set to obtain a trained neural network model; and adding additive white Gaussian noise into the test data in the test data set, and inputting the test data into the trained neural network model to obtain a bearing fault diagnosis result. The neural network obtained through training can automatically extract features from the original signal without manually selecting the features and denoising, can distinguish different fault types and fault severity degrees, and can adapt to different noise environments.
As shown in fig. 4, which is a schematic structural diagram of an embodiment 2 of the bearing fault diagnosis system based on the adaptive anti-noise neural network disclosed in the present invention, the system may include:
an obtaining module 401, configured to obtain a data set, and compare the data set with a data set of 8:2 into a training data set and a test data set;
this example uses two different data sets, one from the bearing data center at the university of Keiss Sige (CWRU) and the second from the bearing fault diagnosis test platform QPZZ-II. Through bearing fault diagnosis experiments, four types of bearing vibration signals are collected. The two data sets were divided into a training data set and a test data set on an 8:2 scale.
Laboratory table for CWRU data set as shown in fig. 5, the vibration signal used was collected from the b-side of the accelerometer driver at a 48000Hz sampling rate. The motor load is 3hp, and the motor rotating speed is changed between 1719 rpm and 1729rpm under different bearing health states. In the embodiment, the average rotating speed of the motor is 1724rpm, and the driving end bearing is a deep groove ball bearing (6205-2RS JEM SKF). The test bearings have four states: normal state, inner ring fault, rolling element fault and outer race fault. Each fault condition has three types of fault diameters, 7mils, 14mils, and 21mils (1 mil-0.001 inch), which are produced by electrical discharge machining. Thus, there is a normal condition and three fault types, each with three severity levels.
The signal of the data set is preprocessed, since the input length of the algorithm is a fixed value, while the vibration signal is periodic, it is necessary to segment the signal according to the period and normalize the amplitude of the input signal to [ -1, +1 ].
A building module 402, configured to build a neural network model, where the neural network model includes: the system comprises an input layer, two convolutional layers, two gating recursion unit layers, an attention mechanism layer and a deep neural network layer;
as shown in fig. 6, the typical structure of the neural network model is mainly composed of one input layer, two convolutional layers, two GRU layers, an attention mechanism layer, and a DNN layer. The neural network model takes a signal with a fixed length as an input, and realizes random sampling of input data by using dropout at an input layer. After the input layers, the two convolutional layers act as feature extractors, using the ELU as an activation function, the formula is as follows:
Figure BDA0002226762390000141
after each convolution layer, there is a batch normalization layer and a dropout layer, respectively, for stability and versatility. However, the device is not suitable for use in a kitchenLater, two GRU layers follow the CNN part, and the Gated Recursion Unit (GRU) is a modified RNN that can capture the long-term dependence of the gate structure. These gates are used to delete or add information to the hidden state to decide whether to remember long term dependencies or to use short term information. The work flow of the GRU is as follows: reset gate r at the t-th time steptThe formula is as follows:
rt=σ(Wrxt+Urht-1+br);
wherein xtAs an input vector, WrAnd UrAs a weight matrix, brTo bias the weights, ht-1Is the previous activation. σ is a logistic sigmoid function. Reset door rtDetermining how many past activations to retain in computing candidate activations
Figure BDA0002226762390000142
Figure BDA0002226762390000143
Figure BDA0002226762390000144
Wherein |, indicates an element multiplication. Candidate activation
Figure BDA0002226762390000151
From an input vector xtAnd a preceding step ht-1Calculated by resetting the gate rtAnd (5) modulating.
zt=σ(Wzxt+Uzht-1+bz);
Updating the door ztControlling the ratio of past activations and candidate activations to calculate a new activation:
Figure BDA0002226762390000152
the GRU layer is followed by another dropout layer. Note that the mechanism layer follows the dropout layer.
The attention mechanism obtains a final result by adopting a weighting method for each time step, and the utilization rate of information is improved. Note that the main calculation process of the institution is shown in FIG. 7, and that the weight vectors are calculated as follows: given all hidden source states
Figure BDA0002226762390000153
Current target state htCalculating an attention weight vector at:
Figure BDA0002226762390000154
Wherein the function score compares the current target hidden state htAnd per source hidden state
Figure BDA0002226762390000155
The contribution of each time step to the final output is calculated, as:
Figure BDA0002226762390000156
wherein, WSIs a trainable weighting matrix. Then, by noticing the weight vector atAnd each step
Figure BDA0002226762390000157
Computing context vector ctI.e., weighted time step vector:
Figure BDA0002226762390000158
finally, the context vector ctWith the target hidden state htConnected to obtain an output attention-hiding state
Figure BDA0002226762390000159
Wherein WcFor a trainable weighting matrix:
Figure BDA00022267623900001510
finally, the classification result is generated and stabilized by the two fully connected layers and the softmax layer. Specific structural parameters are shown in fig. 8.
The Dropout rate is fixed at 0.5. The convolution kernel size is 128 x 1 and the number of first and second layers of filters is 64 and 72, respectively. The step size of the convolutional layer is 1, and no padding is used for the convolutional layer. GRU first layer neurons 64, second layer neurons 128. There are 128 neurons in the hidden state vector of note. Then two fully connected layers, 64 and 32 neurons respectively. Finally, a softmax layer with 10 neurons.
A training module 403, configured to input a signal of training data in a noise-free training data set in a random sampling manner, set a hyper-parameter, a weight initialization manner, and a loss function of a neural network model, train the neural network model by using a stochastic gradient descent method and a back propagation algorithm, and obtain a trained neural network model;
training all neural network models on noiseless original data, testing data added with different noise levels, inputting original signals of training data in the noiseless training data set in a random sampling mode, setting hyper-parameters such as learning rate, attenuation rate and training times of the neural network models, and using an Adam optimization algorithm as an optimizer. And (3) adopting a power attenuation learning rate, setting an initial value of 0.0001 and a weight of 0.9, setting a weight initialization mode, setting a loss function of the network model, training the model by using a back propagation algorithm, and storing the trained model weight.
And the diagnosis module 404 is configured to add additive white gaussian noise to the test data in the test data set, and input the test data to the trained neural network model to obtain a bearing fault diagnosis result.
Preparing a test data set, adding additive white Gaussian noise to the test data in the test data set, inputting the test data into a trained neural network model, simulating different noise conditions, and finally measuring by a signal-to-noise ratio (SNR), wherein the SNR is defined as the ratio of the power of a meaningful signal to the power of background noise [63 ]. The decibel form (dB) of the signal-to-noise ratio is expressed as:
Figure BDA0002226762390000161
wherein, P is the average power of the periodic signal c (t) in the period t, and is defined as:
Figure BDA0002226762390000171
when the signal is in discrete form, the average power can be calculated as:
Figure BDA0002226762390000172
for a signal with zero mean and known variance, the power can be measured by the variance
Figure BDA0002226762390000173
Indicating that, therefore, for standard normally distributed noise, the power is 1. Thus, the power of the original signal is first calculated, and then the noise signal P generated at the desired signal-to-noise ratio is calculatednoiseOf the power of (c). Finally, an additive white gaussian noise is generated by the following formula and then added to the original signal to make the signal have a desired signal-to-noise ratio. The standard normally distributed noise can be generated by:
Figure BDA0002226762390000174
wherein randn represents a function of a random number or matrix that produces a standard normal distribution;
and finally, obtaining a bearing fault diagnosis result.
In summary, the neural network architecture of the present invention mainly comprises four parts: random sample data input, CNN-based enhanced feature extractor, GRU-based feature classifier, attention-based mechanism and DNN-based feature post-processing. The input signal adopts a simple random sampling strategy, so that the adaptability of the network to noise is improved. In the feature extraction part, the convolutional layer adopts an exponential linear unit activation function and a dropout to improve the adaptability of the neural network under the noise condition. The function of the feature extractor is similar to that of a time-domain frequency-domain converter, a noise reduction filter and a feature selector, the feature extractor is similar to that of the time-domain frequency-domain converter, the noise reduction filter, the feature selector and the like, features can be automatically extracted from noise signals without manual feature selection and denoising processes, and the feature extractor has strong adaptability to noise. The features conveyed by the CNN part are further learned and classified using the famous GRU as a decoder or classifier when processing time-varying sequences. Finally, using the attention mechanism and DNN component as a post processor, appropriate features are selected from the GRU output as the final classification result. The neural network can distinguish not only different fault types, but also severity of corresponding faults; the method has stronger processing adaptability to signals with large noise and better universality to noise signals with different grades.
The embodiments in the present description are described in a progressive manner, each embodiment focuses on differences from other embodiments, and the same and similar parts among the embodiments are referred to each other. The device disclosed by the embodiment corresponds to the method disclosed by the embodiment, so that the description is simple, and the relevant points can be referred to the method part for description.
Those of skill would further appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, computer software, or combinations of both, and that the various illustrative components and steps have been described above generally in terms of their functionality in order to clearly illustrate this interchangeability of hardware and software. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
The steps of a method or algorithm described in connection with the embodiments disclosed herein may be embodied directly in hardware, in a software module executed by a processor, or in a combination of the two. A software module may reside in Random Access Memory (RAM), memory, Read Only Memory (ROM), electrically programmable ROM, electrically erasable programmable ROM, registers, hard disk, a removable disk, a CD-ROM, or any other form of storage medium known in the art.
The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present invention. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the invention. Thus, the present invention is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (10)

1. A bearing fault diagnosis method based on an adaptive anti-noise neural network is characterized by comprising the following steps:
acquiring a data set, and dividing the data set into a training data set and a testing data set according to a preset proportion;
constructing a neural network model, wherein the neural network model comprises: the system comprises an input layer, two convolutional layers, two gating recursion unit layers, an attention mechanism layer and a deep neural network layer; the input layer realizes random sampling of input data by using dropout;
training the constructed neural network model based on training data in a noise-free training data set to obtain a trained neural network model;
adding additive white Gaussian noise into the test data in the test data set, and inputting the test data into the trained neural network model to obtain a bearing fault diagnosis result;
wherein, the attention mechanism obtains the final result by adopting a weighting method for each time step, and pays attention to the weight vector atThe calculation formula of (a) is as follows:
Figure FDA0003517071060000011
h istExpressed as a current target hidden state, said
Figure FDA0003517071060000012
Represented as a hidden source state; the above-mentioned
Figure FDA0003517071060000013
The function score compares the current target hidden state htAnd each hidden source state
Figure FDA0003517071060000014
Calculating the contribution of each time step to the final output; the W isSExpressed as a trainable weight matrix;
weighted time step vector ctThe calculation formula of (a) is as follows:
Figure FDA0003517071060000015
attention to hidden states
Figure FDA0003517071060000016
The calculation formula of (a) is as follows:
Figure FDA0003517071060000021
vector c of weighted time stepstWith the target hidden state htAre connected to obtain outputHidden state of attention
Figure FDA0003517071060000022
The W iscRepresented as a trainable weight matrix.
2. The method of claim 1, wherein the obtaining the data set and the dividing the data set into a training data set and a testing data set according to a preset ratio comprises:
acquiring a data set, and enabling the data set to be as follows: the scale of 2 is divided into a training data set and a test data set.
3. The method of claim 1, wherein two of the convolutional layers act as feature extractors using ELUs as activation functions.
4. The method of claim 1, wherein training the constructed neural network model based on the training data in the noise-free training data set to obtain a trained neural network model comprises:
inputting the signals of the training data in the noise-free training data set in a random sampling mode;
setting a hyper-parameter, a weight initialization mode and a loss function of a neural network model;
and training the neural network model by using a random gradient descent method and a back propagation algorithm to obtain the trained neural network model.
5. The method of claim 4, wherein the hyper-parameters comprise: learning rate, attenuation rate and training times of the neural network model.
6. A bearing fault diagnostic system based on an adaptive noise immune neural network, comprising:
the acquisition module is used for acquiring a data set and dividing the data set into a training data set and a testing data set according to a preset proportion;
a building module configured to build a neural network model, wherein the neural network model comprises: the system comprises an input layer, two convolutional layers, two gating recursion unit layers, an attention mechanism layer and a deep neural network layer; the input layer realizes random sampling of input data by using dropout;
the training module is used for training the built neural network model based on training data in the noiseless training data set to obtain a trained neural network model;
the diagnosis module is used for adding additive white Gaussian noise into the test data in the test data set and inputting the test data into the trained neural network model to obtain a bearing fault diagnosis result;
wherein, the attention mechanism obtains the final result by adopting a weighting method for each time step, and pays attention to the weight vector atThe calculation formula of (a) is as follows:
Figure FDA0003517071060000031
h istExpressed as a current target hidden state, said
Figure FDA0003517071060000032
Represented as a hidden source state; the above-mentioned
Figure FDA0003517071060000033
The function score compares the current target hidden state htAnd each hidden source state
Figure FDA0003517071060000034
Calculating the contribution of each time step to the final output; the W isSExpressed as a trainable weight matrix;
weighted time step vector ctThe calculation formula of (a) is as follows:
Figure FDA0003517071060000035
attention to hidden states
Figure FDA0003517071060000036
The calculation formula of (a) is as follows:
Figure FDA0003517071060000037
vector c of weighted time stepstWith the target hidden state htConnected to obtain an output attention-hiding state
Figure FDA0003517071060000038
The W iscRepresented as a trainable weight matrix.
7. The system of claim 6, wherein the obtaining module, when executing obtaining the data set and dividing the data set into the training data set and the testing data set according to a preset ratio, is specifically configured to:
acquiring a data set, and enabling the data set to be as follows: the scale of 2 is divided into a training data set and a test data set.
8. The system of claim 6, wherein the two convolutional layers act as feature extractors using ELUs as activation functions.
9. The system according to claim 6, wherein the training module, when executing the training data in the noise-free training data set to train the constructed neural network model, is specifically configured to:
inputting the signals of the training data in the noise-free training data set in a random sampling mode;
setting a hyper-parameter, a weight initialization mode and a loss function of a neural network model;
and training the neural network model by using a random gradient descent method and a back propagation algorithm to obtain the trained neural network model.
10. The system of claim 9, wherein the hyper-parameters comprise: learning rate, attenuation rate and training times of the neural network model.
CN201910957332.1A 2019-10-09 2019-10-09 Bearing fault diagnosis method and system based on self-adaptive anti-noise neural network Active CN110647867B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910957332.1A CN110647867B (en) 2019-10-09 2019-10-09 Bearing fault diagnosis method and system based on self-adaptive anti-noise neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910957332.1A CN110647867B (en) 2019-10-09 2019-10-09 Bearing fault diagnosis method and system based on self-adaptive anti-noise neural network

Publications (2)

Publication Number Publication Date
CN110647867A CN110647867A (en) 2020-01-03
CN110647867B true CN110647867B (en) 2022-05-13

Family

ID=69012501

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910957332.1A Active CN110647867B (en) 2019-10-09 2019-10-09 Bearing fault diagnosis method and system based on self-adaptive anti-noise neural network

Country Status (1)

Country Link
CN (1) CN110647867B (en)

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111579243B (en) * 2020-06-17 2021-04-20 大连理工大学 Rolling bearing intelligent diagnosis system based on deep migration learning
CN112149730B (en) * 2020-09-22 2022-06-10 中南大学 High-speed train bearing fault prediction method based on oil sample vision
CN112629854B (en) * 2020-11-25 2022-08-05 西安交通大学 Bearing fault classification method based on neural network attention mechanism
CN112839349B (en) * 2020-12-25 2022-10-25 朗坤智慧科技股份有限公司 Angular domain stationary signal denoising method and device based on 5G technology and network server
CN113203566B (en) * 2021-04-06 2022-09-16 上海吞山智能科技有限公司 Motor bearing fault diagnosis method based on one-dimensional data enhancement and CNN
CN113343855B (en) * 2021-06-09 2022-09-16 西南交通大学 Rolling bearing fault diagnosis system and method based on guide type sub-field self-adaption
CN113473514B (en) * 2021-09-06 2021-11-16 国网江苏省电力有限公司信息通信分公司 Electric power wireless private network fault diagnosis model training method, diagnosis method and apparatus
CN113761225A (en) * 2021-09-09 2021-12-07 昆明理工大学 Automobile engine fault prediction method fusing knowledge graph and multivariate neural network model
CN114383844B (en) * 2021-12-15 2023-06-23 太原理工大学 Rolling bearing fault diagnosis method based on distributed deep neural network
CN114459760B (en) * 2021-12-31 2024-05-24 南京理工大学 Rolling bearing fault diagnosis method and system in strong noise environment
CN115929495B (en) * 2022-11-30 2024-05-14 昆明理工大学 Engine valve fault diagnosis method based on Markov transition field and improved Gaussian prototype network
CN115984801A (en) * 2023-03-07 2023-04-18 安徽蔚来智驾科技有限公司 Point cloud target detection method, computer equipment, storage medium and vehicle

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109446187A (en) * 2018-10-16 2019-03-08 浙江大学 Complex equipment health status monitoring method based on attention mechanism and neural network
CN109522600A (en) * 2018-10-16 2019-03-26 浙江大学 Complex equipment remaining life prediction technique based on combined depth neural network
CN109902399A (en) * 2019-03-01 2019-06-18 哈尔滨理工大学 Rolling bearing fault recognition methods under a kind of variable working condition based on ATT-CNN

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11188065B2 (en) * 2017-09-23 2021-11-30 Nanoprecise Sci Corp. System and method for automated fault diagnosis and prognosis for rotating equipment

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109446187A (en) * 2018-10-16 2019-03-08 浙江大学 Complex equipment health status monitoring method based on attention mechanism and neural network
CN109522600A (en) * 2018-10-16 2019-03-26 浙江大学 Complex equipment remaining life prediction technique based on combined depth neural network
CN109902399A (en) * 2019-03-01 2019-06-18 哈尔滨理工大学 Rolling bearing fault recognition methods under a kind of variable working condition based on ATT-CNN

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Bearing Fault Diagnosis Using Structure Optimized Deep Convolutional Neural Network under Noisy Environment;Guoqiang Jin等;《IOP Conference Series: Materials Science and Engineering》;20190807;1-5 *
深度学习在设备故障预测与健康管理中的应用;陈志强等;《仪器仪表学报》;20190930;206-226 *

Also Published As

Publication number Publication date
CN110647867A (en) 2020-01-03

Similar Documents

Publication Publication Date Title
CN110647867B (en) Bearing fault diagnosis method and system based on self-adaptive anti-noise neural network
CN110516305B (en) Intelligent fault diagnosis method under small sample based on attention mechanism meta-learning model
CN106682688B (en) Particle swarm optimization-based stacked noise reduction self-coding network bearing fault diagnosis method
Xia et al. Multi-stage fault diagnosis framework for rolling bearing based on OHF Elman AdaBoost-Bagging algorithm
CN110617966A (en) Bearing fault diagnosis method based on semi-supervised generation countermeasure network
CN110657985B (en) Gearbox fault diagnosis method and system based on singular value spectrum manifold analysis
CN110991424A (en) Fault diagnosis method based on minimum entropy deconvolution and stacking sparse self-encoder
CN110795843A (en) Method and device for identifying faults of rolling bearing
CN111595583B (en) GAU neural network-based bearing residual life prediction method
CN115017945A (en) Mechanical fault diagnosis method and system based on enhanced convolutional neural network
CN114169110B (en) Motor bearing fault diagnosis method based on feature optimization and GWAA-XGboost
CN111581892A (en) Method for predicting residual life of bearing based on GDAU neural network
CN113532829A (en) Reciprocating compressor fault diagnosis method based on improved RCMDE
CN115901249B (en) Rolling bearing performance degradation evaluation method combining feature optimization and multi-strategy optimization SVDD
CN115290326A (en) Rolling bearing fault intelligent diagnosis method
CN113657244B (en) Fan gearbox fault diagnosis method and system based on improved EEMD and speech spectrum analysis
Lee et al. An enhanced binary particle swarm optimization for optimal feature selection in bearing fault diagnosis of electrical machines
CN117150882A (en) Engine oil consumption prediction method, system, electronic equipment and storage medium
CN115931359B (en) Turbine pump bearing fault diagnosis method and device
Li et al. Research on rolling bearing fault diagnosis based on DRS frequency spectrum image and deep learning
CN117851872B (en) High-frequency vibration treatment diagnosis method and system for aero-engine
CN117974736B (en) Underwater sensor output signal noise reduction method and system based on machine learning
CN117708574B (en) CNN variable-speed rolling bearing fault diagnosis method embedded with physical information
CN114739673B (en) Bearing vibration signal characteristic interpretability dimension reduction and fault diagnosis method
Wu et al. Sound Visualization and Convolutional Neural Network in Fault Diagnosis of Electric Motorbike.

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant