CN110991471B - Fault diagnosis method for high-speed train traction system - Google Patents

Fault diagnosis method for high-speed train traction system Download PDF

Info

Publication number
CN110991471B
CN110991471B CN201910708383.0A CN201910708383A CN110991471B CN 110991471 B CN110991471 B CN 110991471B CN 201910708383 A CN201910708383 A CN 201910708383A CN 110991471 B CN110991471 B CN 110991471B
Authority
CN
China
Prior art keywords
encoder
lstm
feature vector
fault
self
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910708383.0A
Other languages
Chinese (zh)
Other versions
CN110991471A (en
Inventor
冒泽慧
闫宇
姜斌
严星刚
吕迅竑
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Aeronautics and Astronautics
Original Assignee
Nanjing University of Aeronautics and Astronautics
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Aeronautics and Astronautics filed Critical Nanjing University of Aeronautics and Astronautics
Priority to CN201910708383.0A priority Critical patent/CN110991471B/en
Publication of CN110991471A publication Critical patent/CN110991471A/en
Application granted granted Critical
Publication of CN110991471B publication Critical patent/CN110991471B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • G06F18/232Non-hierarchical techniques
    • G06F18/2321Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01MTESTING STATIC OR DYNAMIC BALANCE OF MACHINES OR STRUCTURES; TESTING OF STRUCTURES OR APPARATUS, NOT OTHERWISE PROVIDED FOR
    • G01M17/00Testing of vehicles
    • G01M17/08Railway vehicles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y04INFORMATION OR COMMUNICATION TECHNOLOGIES HAVING AN IMPACT ON OTHER TECHNOLOGY AREAS
    • Y04SSYSTEMS INTEGRATING TECHNOLOGIES RELATED TO POWER NETWORK OPERATION, COMMUNICATION OR INFORMATION TECHNOLOGIES FOR IMPROVING THE ELECTRICAL POWER GENERATION, TRANSMISSION, DISTRIBUTION, MANAGEMENT OR USAGE, i.e. SMART GRIDS
    • Y04S10/00Systems supporting electrical power generation, transmission or distribution
    • Y04S10/50Systems or methods supporting the power network operation or management, involving a certain degree of interaction with the load-side end user applications
    • Y04S10/52Outage or fault management, e.g. fault detection or location

Abstract

The invention discloses a fault diagnosis method for a traction system of a high-speed train, which relates to the field of fault diagnosis of the high-speed train and comprises the following steps: collecting sequence data from a semi-physical simulation platform and preprocessing the sequence data to obtain a data set, wherein the data set comprises a training set and a testing set; improving an LSTM self-encoder by utilizing state differential feedback control to obtain an LSTM self-encoder I, wherein the LSTM self-encoder I consists of L LSTM units; training the LSTM self-encoder I by utilizing the data set to obtain an LSTM self-encoder II; extracting original feature vectors from the second pair of test sets by using the LSTM self-encoder; performing feature dimension reduction on the original feature vector by using a t-SNE algorithm; performing fault diagnosis on the original feature vector after dimension reduction by a DBSCAN clustering method to obtain a diagnosis result. The invention can effectively solve the problems of difficult diagnosis and low diagnosis accuracy of the tiny gradual change fault of the high-speed train traction system.

Description

Fault diagnosis method for high-speed train traction system
Technical Field
The invention relates to the field of fault diagnosis of high-speed trains, in particular to a diagnosis method of tiny gradual change faults of a high-speed train traction system based on data learning.
Background
At present, high-speed trains play an increasingly important role in China passenger transport and freight transport, and a traction system is used as a power core system of the high-speed trains, so that accidents such as train stopping and delay are caused by faults of the traction system, and huge losses are caused. The micro gradual change fault occurs in the initial stage of the significant fault, and has the characteristics of unobvious fault characteristics and easy unknown disturbance and noise masking. Therefore, the detection and diagnosis of the minor gradual failure of the traction system are difficult. The small gradual change faults of the traction system are effectively detected and diagnosed, effective measures are timely taken, the safety of the system can be improved, and the maintenance cost and loss are reduced.
The effective fault diagnosis of the micro gradual change fault can improve the safety of the system, reduce the maintenance cost and loss, and has extremely important significance in the aspects of safety maintenance, health management of equipment and the like. Because the complexity of the traction system and the fault characteristics of the tiny gradual change faults are not obvious, the change characteristics are not obvious, and the detection and diagnosis of the tiny gradual change faults are difficult to effectively realize based on a data driving algorithm at present.
Disclosure of Invention
The invention aims to provide a fault diagnosis method for a high-speed train traction system, which aims to solve the technical problems of difficult diagnosis and low diagnosis accuracy of tiny gradual change faults in the prior art.
The invention provides a fault diagnosis method for a traction system of a high-speed train, which comprises the following steps:
(a) Collecting sequence data from a semi-physical simulation platform and preprocessing the sequence data to obtain a data set, wherein the data set comprises a training set and a testing set;
(b) Improving an LSTM self-encoder by utilizing state differential feedback control to obtain an LSTM self-encoder I, wherein the LSTM self-encoder I consists of L LSTM units;
(c) Training the LSTM self-encoder I by utilizing the data set to obtain an LSTM self-encoder II;
(d) Extracting original feature vectors from the second pair of test sets by using the LSTM self-encoder;
(e) Performing feature dimension reduction on the original feature vector by using a t-SNE algorithm;
(f) Performing fault diagnosis on the original feature vector after dimension reduction by a DBSCAN clustering method to obtain a diagnosis result.
Compared with the prior art, the diagnosis method for the micro gradual change fault of the high-speed train traction system based on data learning has the following beneficial effects:
the method can better extract the original characteristics of the micro gradual change fault by carrying out characteristic extraction on the basis of the sequence data of the micro gradual change fault by the improved LSTM unit self-encoder;
the original characteristics are subjected to dimension reduction processing through a t-SNE algorithm, redundancy of the original characteristic information is reduced, diagnosis performance is improved, and calculation complexity is reduced;
the unsupervised diagnosis of the micro gradual change faults is realized by the DBSCAN clustering method, the marking in advance is not needed, the diagnosis accuracy is very high, the unknown faults have certain diagnosis capability, and the micro gradual change fault diagnosis requirement is met.
Aiming at a traction motor system of a high-speed train, the method is used for diagnosing tiny gradual faults such as a middle capacitor, a middle resistor, a speed sensor degradation fault and the like of the traction system, has a wider application range and has engineering application value.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings that are needed in the description of the embodiments or the prior art will be briefly described, and it is obvious that the drawings in the description below are some embodiments of the present invention, and other drawings can be obtained according to the drawings without inventive effort for a person skilled in the art.
FIG. 1 shows an improved LSTM cell structure provided by an embodiment of the present invention;
FIG. 2 (a) is a block diagram of an overall framework and training process for a codec model according to an embodiment of the present invention;
FIG. 2 (b) is a general framework and flow diagram providing a fault diagnosis model for an embodiment of the present invention;
FIG. 3 (a) is an effect diagram of different dimension reduction methods for providing known fault types according to an embodiment of the present invention;
FIG. 3 (b) is an effect diagram of a method for dimension reduction of t-SNE for unknown fault types according to an embodiment of the present invention;
FIG. 4 (a) provides a graph of known fault class clustering effects for an embodiment of the present invention;
FIG. 4 (b) provides a graph of unknown fault class clustering effects for an embodiment of the present invention;
FIG. 5 is a schematic diagram of an architecture of a high speed train traction system according to an embodiment of the present invention;
Detailed Description
The following description of the embodiments of the present invention will be made apparent and fully in view of the accompanying drawings, in which some, but not all embodiments of the invention are shown. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
In the description of the present invention, it should be noted that the directions or positional relationships indicated by the terms "center", "upper", "lower", "left", "right", "vertical", "horizontal", "inner", "outer", etc. are based on the directions or positional relationships shown in the drawings, are merely for convenience of describing the present invention and simplifying the description, and do not indicate or imply that the devices or elements referred to must have a specific orientation, be configured and operated in a specific orientation, and thus should not be construed as limiting the present invention. Furthermore, the terms "first," "second," and "third" are used for descriptive purposes only and are not to be construed as indicating or implying relative importance.
In the description of the present invention, it should be noted that, unless explicitly specified and limited otherwise, the terms "mounted," "connected," and "connected" are to be construed broadly, and may be either fixedly connected, detachably connected, or integrally connected, for example; can be mechanically or electrically connected; can be directly connected or indirectly connected through an intermediate medium, and can be communication between two elements. The specific meaning of the above terms in the present invention will be understood in specific cases by those of ordinary skill in the art.
The following describes the embodiments of the present invention in further detail with reference to the drawings.
As shown in fig. 2 (b), the diagnosis method of the micro gradual change fault of the high-speed train traction system based on data learning mainly comprises the following steps:
1. collecting sequence data from a semi-physical simulation platform and preprocessing to obtain a data set, wherein the data set comprises a training set and a testing set
Preprocessing the acquired data: supplementing missing values, modifying abnormal values, smoothing filtering, reducing noise and normalizing.
Specifically, filling missing values in data by adopting a 2-nearest neighbor method for missing values of the training set and the testing set; analyzing abnormal values of the sequence data through a box graph, and modifying the numerical value which is judged to be the abnormal value in the data by adopting a 2-nearest neighbor method; the sequence data is subjected to smooth filtering in a wavelet filtering mode, so that noise in the sequence data is reduced; and normalizing the sequence data by adopting a Z-normalization method.
2. Improving an LSTM self-encoder by using state differential feedback control to obtain an LSTM self-encoder I, wherein the LSTM self-encoder I consists of L LSTM units
The basic LSTM unit is improved, and the state differential feedback control is applied to the LSTM unit: state differential feedback control is applied in the modified LSTM cell.
Specifically, the state differential signal p of the state vector of the dynamic system at the time t-1 and the time t-2 t-1 Updated at time t-1 and fed back to the system at time t to participate in controlling LSTM hidden unit state vector c t Forget and update of (a). The modified LSTM cell description is shown in the following formula:
Figure GDA0004211394970000051
wherein f t Representing forget gate layer output, r t Representing updated gate layer output, c t The state of the hidden unit is indicated,
Figure GDA0004211394970000052
representing the new state quantity to be selected, p t Differential vector, o, representing a unit hidden state vector t Represents the output gate layer output, h t Representing the LSTM cell output.
If LSTM control vector
Figure GDA0004211394970000056
And if so, the LSTM unit hidden unit state c is stabilized under the action of stabilizing the vector to be updated, so that the LSTM network is stabilized. Control vector->
Figure GDA0004211394970000057
Related to the state differential vector p of LSTM cell state c. When the sequence sample of the tiny gradual change fault sequence data is short, the differential vector of the fault information variable changes linearly when the Taylor expansion is first-order approximation. When the LSTM unit learns that the hidden unit state c has higher correlation with the tiny gradual change fault information variable, the differential vector p of the LSTM unit hidden unit state c is stable along with the iterative input of the sequence data. But->
Figure GDA0004211394970000053
Has a strong correlation with the differential vector p, thus generating a stable control vector +.>
Figure GDA0004211394970000054
At this time, the LSTM unit hidden unit state variable is not easy to forget and update. When the correlation between the hidden unit state c learned by the LSTM unit and the tiny gradual change fault information variable is weak, the differential vector p is unstable, and the hidden unit state c of the LSTM unit is controlled by the unstable control vector +>
Figure GDA0004211394970000055
Can be forgotten and updated frequently. Secondly, the LSTM self-encoder after training can acquire differential trend signals representing fault information variables, and can diagnose micro gradual change faults earlier.
3. Training the LSTM self-encoder I with the data set to obtain the LSTM self-encoder II
And acquiring the LSTM self-encoder by using the feature vector of the training set to unsupervised train the codec model. Improved LSTM based self-encoderAfter the system sequence data is processed, fault information variables are embedded into the original characteristic vectors with fixed lengths. During the training of the codec model, the LSTM reads from the encoder the sequence samples X of the traction system, each sample
Figure GDA0004211394970000061
Is vector type sequence data, wherein X=x 1 ,x 2 ...,x L-1 ,x L ,N v Representing the vector dimension, L represents the sequence data sample length. After the reading process of the sequence data sample X is completed, LSTM is embedded in the encoder from the last layer hidden unit state +.>
Figure GDA0004211394970000062
Will be retained and serve as the original feature vector. The GRU decoder initializes the hidden unit state of the first layer to the original feature vector +.>
Figure GDA0004211394970000063
And outputs the target sequence +.>
Figure GDA0004211394970000064
Wherein->
Figure GDA0004211394970000065
Training is targeted to codec output sequences
Figure GDA0004211394970000066
Remains similar under the metric to the input sequence X, but in reverse order. In the present invention, by minimizing +.>
Figure GDA0004211394970000067
And square error J between X ed Training:
Figure GDA0004211394970000068
wherein D is N Representing the entire training set, in the present invention, the plaitedThe decoder model trained optimizer is Adam optimizer.
4. Extracting original feature vectors from the test set by using the LSTM self-encoder and using a trained LSTM self-encoder model, inputting a sequence sample X of the test set at an input part, each sample
Figure GDA0004211394970000069
Is vector type sequence data, wherein X=x 1 ,x 2 ...,x L-1 ,x L ,N v Representing the vector dimension, L represents the sequence data sample length. After the reading process of the sequence data sample X is completed, LSTM is embedded in the encoder from the last layer hidden unit state +.>
Figure GDA00042113949700000610
Will be retained and used as the original feature vector to complete the task of the LSTM self-encoder to extract the original feature vector of the test set data.
The trained LSTM self-encoder extracts the original feature vector from the sequence data of the small gradual change fault of the traction system.
5. Performing feature dimension reduction on the original feature vector by using t-SNE algorithm
The original feature vector is subjected to t-SNE algorithm
Figure GDA0004211394970000071
The dimension reduction is carried out, and the problems of high dimension of the original feature vector and information redundancy are solved. In the t-SNE algorithm, the loss function J SNE The joint probability P of data distribution in the original feature vector space and the target dimension reduction space D And Q is equal to D Is a difference in (a) between the two. The probability distribution difference uses a Kullback-Leibler divergence. Loss function J of T-SNE algorithm SNE The expression is as follows:
Figure GDA0004211394970000072
wherein p is ij And q ij Is defined as:
Figure GDA0004211394970000073
subscripts i, j, k, l denote samples numbered i, j, k, l, respectively, of the samples,/->
Figure GDA0004211394970000074
Representing the original feature vector in the high-dimensional space, z representing the feature vector after dimension reduction, σ representing the sum of the sum and the original feature vector +.>
Figure GDA0004211394970000075
The bandwidth of the associated gaussian kernel function. In the optimization process, the KL divergence is minimized by a random gradient descent method. Loss function J SNE The gradient information of (2) is given by:
Figure GDA0004211394970000076
in the t-SNE algorithm, the similarity of feature vector distribution before and after dimension reduction is kept by minimizing the KL distance from the high-dimension original feature vector space to the low-dimension feature vector space.
6. Performing fault diagnosis on the original feature vector subjected to dimension reduction by using a DBSCAN clustering method to obtain a diagnosis result
And performing unsupervised training on the coder-decoder model by using the training set sequence data to obtain the LSTM self-encoder model. Raw eigenvector of sequence data to be diagnosed
Figure GDA0004211394970000081
The LSTM is extracted from the encoder by training. Original feature vector +.>
Figure GDA0004211394970000082
The dimension is high, information redundancy exists, in order to reduce the information redundancy and the complexity of algorithm operation, the original feature vector c is subjected to the t-SNE algorithm tL And obtaining a final feature vector z after dimension reduction. And finally, clustering the feature vector z by a DBSCAN clustering method to realize the diagnosis of the micro gradual change faults of the traction system.
The following simulation verification is carried out on the diagnosis method of the tiny gradual change fault of the high-speed train traction system based on data learning:
and 1, firstly, confirming the types of micro gradual change faults, wherein the micro gradual change fault types comprise an intermediate capacitance degradation fault (fault I), an intermediate resistance degradation fault (fault II) and a speed sensor degradation fault (fault III) in the verification. Also included in this experimental verification is a speed sensor bias fault (fault IV), which is a minor abrupt fault. The fault I and the fault II are used for unsupervised fault diagnosis model training and final fault diagnosis model testing processes. The untrained fault class (fault III and fault IV in the experimental verification) is only used for testing and verifying the trained fault diagnosis model, and the fault III and the fault IV are used for testing the generalization capability of the model to unknown faults. The detailed descriptions of the four faults in the experimental verification are shown in table 1.
TABLE 1 description of fault types
Figure GDA0004211394970000083
Figure GDA0004211394970000091
The data of the faults I and II are divided into a training set and a testing set, and the data form of each sample is a vector of sequence data. There are a total of about 14000 samples in the form of sequence data vectors in the overall data. 60% of the data in the whole dataset was randomly selected to make up the training set (8500 samples), with the remaining samples making up the test set (5455 samples). Wherein, the healthy sample number is 1180, the fault I sample number is 990, and the fault II sample number is 935. Likewise, the number of samples for failure III is 1175, the number of samples for failure IV is 1175, and the length of each sequence data sample ranges from 20 to 40.
And 2, extracting original feature vectors of the fault I and fault II test sets by the improved LSTM self-encoder which is completed through training. And applying t-SNE to perform feature vector dimension reduction. Fig. 3 (a) shows the dimensionality reduction effect of different dimensionality reduction methods (including t-SNE, kernel PCA and Isomap) on feature vectors of known fault types. Where, + represents a healthy sample point, + represents a sample point with fault I, x represents a sample point with fault II. As can be seen from the PCA dimension reduction result and the Isomap dimension reduction result, the feature vectors of the health samples represent stable annular distribution in the low-dimensional space, and the two types of micro-gradual-change faults of the faults I and II represent conical-like distribution starting from the health annular in the low-dimensional space, which is consistent with the slow time-varying characteristic of the fault information variable in the micro-gradual-change fault, so that the LSTM self-encoder is proved to successfully extract the feature vectors related to the micro-gradual-change fault information variable. In the present invention, an experimental verification test is performed on the type of failure (unknown failure class) that is not used for codec training. The test sets of faults III and IV are used to test the generalization ability of the fault diagnosis algorithm to unknown fault types. Fault III is a minor gradual fault in the traction system and fault IV is a minor abrupt fault occurring in the traction system. Fig. 3 (b) shows the dimension reduction effect of the feature vector for an unknown fault using the t-SNE algorithm, with good inter-class and intra-class distances for the unknown fault class.
And step 3, clustering the feature vector samples subjected to dimension reduction by using a DBSCAN clustering algorithm to finish fault diagnosis of the test set. The DBSCAN clustering algorithm is a density-based clustering algorithm that generally assumes that the class can be determined by how tightly the sample is distributed. Samples of the same class are closely connected, that is, there must be samples of the same class around any sample of that class. By grouping closely connected samples into one class, a cluster class is thus obtained. By grouping all closely connected sets of samples into different categories we get the final all clusters category result. The method describes how tight the sample set is based on a set of neighbors. The DBSCAN method is simple, it selects a core object without category as seed, then finds out the sample set with reachable density of the core object, which is a cluster. Then, another core object without category is continuously selected to search a sample set with reachable density, and thus, another cluster is obtained. Run until all core objects have a class.
Fig. 4 (a) clusters data of known fault types, and the clustering result shows that the fault diagnosis algorithm based on the improved LSTM self-encoder can successfully realize degradation fault diagnosis of the intermediate capacitance and the intermediate resistance of the traction system, namely realize micro gradual change fault diagnosis of the traction system. Fig. 4 (b) clusters data of unknown fault types, and the clustering result shows that the algorithm provided by the invention has good diagnosis effect on some unknown faults.
As can be seen from FIG. 4, the method of the invention can effectively diagnose the tiny gradual change faults of the traction motor of the high-speed train. For some unlabeled faults, the method also has good diagnosis effect and relatively strong generalization capability.
As shown in fig. 5, the embodiment of the present invention further provides a schematic architecture of a traction system for a high-speed train, where the traction system 120 includes: at least one processor 121, such as a CPU, at least one network interface 124 or other user interface 123, memory 125, and at least one communication bus 122. The communication bus 122 is used to enable connected communication between these components. Optionally, a user interface 123 is also included, including a display, keyboard or pointing device (e.g., mouse, trackball, touch pad or touch sensitive display). The memory 125 may comprise a high-speed RAM memory or may further comprise a non-volatile memory (non-volatile memory), such as at least one disk memory. Memory 125 may optionally include at least one storage device located remotely from the aforementioned processor 121.
In some implementations, the memory 125 stores the following elements, executable modules or data structures, or a subset thereof, or an extended set thereof:
an operating system 1251 containing various system programs for implementing various basic services and handling hardware-based tasks;
the application 1252 includes various applications for implementing various application services.
Specifically, the processor 121 is configured to execute the processes corresponding to steps (1) - (4); in this process, the processor 121 needs to accept the signal processing sent by the sensor set 126 through the network interface 124.
Meanwhile, the processor 121 is further configured to: the fault diagnosis result is output to the visual interface through the user interface 123.
Finally, it should be noted that: the above embodiments are only for illustrating the technical solution of the present invention, and not for limiting the same; although the invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical scheme described in the foregoing embodiments can be modified or some or all of the technical features thereof can be replaced by equivalents; such modifications and substitutions do not depart from the spirit of the invention.

Claims (5)

1. A fault diagnosis method for a traction system of a high-speed train is characterized by comprising the following steps of:
(a) Collecting sequence data from a semi-physical simulation platform and preprocessing the sequence data to obtain a data set, wherein the data set comprises a training set and a testing set;
(b) Improving an LSTM self-encoder by utilizing state differential feedback control to obtain an LSTM self-encoder I, wherein the LSTM self-encoder I consists of L LSTM units;
(c) Training the LSTM self-encoder I by utilizing the data set to obtain an LSTM self-encoder II;
(d) Extracting original feature vectors from the second pair of test sets by using the LSTM self-encoder;
(e) Performing feature dimension reduction on the original feature vector by using a t-SNE algorithm;
(f) Performing fault diagnosis on the original feature vector subjected to dimension reduction by a DBSCAN clustering method to obtain a diagnosis result;
the (c) includes:
training the LSTM self-encoder I by utilizing the data set to obtain an LSTM self-encoder II;
the LSTM reads the data set from the encoder, the data set
Figure FDA0004211394950000011
For vector sequence data, x=x 1 ,x 2 ...,x L-1 ,x L ,N v Representing vector dimensions, L representing sequence data sample length;
the LSTM self-encoder has a last layer hidden unit state
Figure FDA0004211394950000012
As a feature vector, the hidden unit state of the first layer of the GRU decoder is +.>
Figure FDA0004211394950000013
The GRU decoder output sequence +.>
Figure FDA0004211394950000014
Wherein->
Figure FDA0004211394950000015
For the said
Figure FDA0004211394950000016
Minimizing the square error J with X ed The training is carried out by the device,
Figure FDA0004211394950000017
wherein D is N Representing the entire training set, the training remains similar under the metrics, in reverse order.
2. The high-speed train traction system fault diagnosis method as claimed in claim 1, wherein: the pretreatment in (a) comprises:
filling the missing value of the sequence data by using a k-nearest neighbor method;
analyzing abnormal values of the sequence data through a box graph, and modifying the sequence data judged to be the abnormal values by using a k-nearest neighbor method;
smoothing and filtering the sequence data by utilizing wavelet filtering, and normalizing the sequence data by utilizing a Z-score normalization method.
3. The high-speed train traction system fault diagnosis method as claimed in claim 1, wherein: in the (d), further comprising:
the LSTM extracts an original feature vector from encoder two, the original feature vector comprising fault information, and embeds the fault information variable into the original feature vector of fixed length.
4. The high-speed train traction system fault diagnosis method as claimed in claim 1, wherein: the (e) includes:
definition of the loss function J SNE The loss function J SNE The probability P is combined for the original feature vector and the data distribution in the target dimension reduction D And Q is equal to D The probability P D And Q is equal to D The difference probability distribution difference of (1) adopts a Kullback-Leibler divergence;
Kullback-Leibler divergence is minimized by a random gradient descent method.
5. The high-speed train traction system fault diagnosis method as claimed in claim 1, wherein: in the above-mentioned (f),
and performing fault diagnosis on the feature vector subjected to dimension reduction by a DBSCAN clustering method to obtain a diagnosis result.
CN201910708383.0A 2019-08-01 2019-08-01 Fault diagnosis method for high-speed train traction system Active CN110991471B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910708383.0A CN110991471B (en) 2019-08-01 2019-08-01 Fault diagnosis method for high-speed train traction system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910708383.0A CN110991471B (en) 2019-08-01 2019-08-01 Fault diagnosis method for high-speed train traction system

Publications (2)

Publication Number Publication Date
CN110991471A CN110991471A (en) 2020-04-10
CN110991471B true CN110991471B (en) 2023-07-11

Family

ID=70081714

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910708383.0A Active CN110991471B (en) 2019-08-01 2019-08-01 Fault diagnosis method for high-speed train traction system

Country Status (1)

Country Link
CN (1) CN110991471B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111860686B (en) * 2020-07-31 2022-08-19 北京航空航天大学 Causal relationship adjacency matrix characteristic extraction method for fault diagnosis
CN112861071B (en) * 2021-02-05 2022-09-02 哈尔滨工程大学 High-speed rail traction system anomaly detection method based on depth self-coding
CN113139590B (en) * 2021-04-13 2022-03-22 索信达(北京)数据技术有限公司 Dimension reduction method and device for time series data, computer equipment and storage medium
CN113507331A (en) * 2021-07-01 2021-10-15 西北工业大学 Sight distance and non-sight distance transmission identification method and device based on channel state information

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108197648A (en) * 2017-12-28 2018-06-22 华中科技大学 A kind of Fault Diagnosis Method of Hydro-generating Unit and system based on LSTM deep learning models

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102055481B1 (en) * 2017-08-24 2019-12-12 한국과학기술원 Method and apparatus for quantitative evaluation assessment of vr content perceptual quality using deep running analysis of vr sickness factors
CN108536123B (en) * 2018-03-26 2019-08-30 北京交通大学 Train control on board equipment method for diagnosing faults based on long Memory Neural Networks in short-term
CN109561084B (en) * 2018-11-20 2020-03-17 四川长虹电器股份有限公司 URL parameter abnormal value detection method based on LSTM self-coding network
CN109919364A (en) * 2019-02-18 2019-06-21 华南理工大学 Multivariate Time Series prediction technique based on adaptive noise reduction and integrated LSTM
CN110021165A (en) * 2019-03-18 2019-07-16 浙江工业大学 A kind of traffic flow forecasting method based on Autoencoder-LSTM Fusion Model

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108197648A (en) * 2017-12-28 2018-06-22 华中科技大学 A kind of Fault Diagnosis Method of Hydro-generating Unit and system based on LSTM deep learning models

Also Published As

Publication number Publication date
CN110991471A (en) 2020-04-10

Similar Documents

Publication Publication Date Title
CN110991471B (en) Fault diagnosis method for high-speed train traction system
Yang et al. ART–KOHONEN neural network for fault diagnosis of rotating machinery
CN111666982B (en) Electromechanical equipment fault diagnosis method based on deep neural network
CN113935406B (en) Mechanical equipment unsupervised fault diagnosis method based on countermeasure flow model
CN111562108A (en) Rolling bearing intelligent fault diagnosis method based on CNN and FCMC
CN109116834B (en) Intermittent process fault detection method based on deep learning
CN111207926B (en) Fault diagnosis method based on rolling bearing, electronic device and storage medium
CN114722952B (en) Bearing fault detection and classification integrated method based on representation learning
CN112257263B (en) Equipment residual life prediction system based on self-attention mechanism
CN113673346A (en) Motor vibration data processing and state recognition method based on multi-scale SE-Resnet
CN105425150A (en) Motor fault diagnosis method based on RBF and PCA-SVDD
CN115204302A (en) Unmanned aerial vehicle small sample fault diagnosis system and method
CN115730637A (en) Multi-mode vehicle trajectory prediction model training method and device and trajectory prediction method
Zhang et al. Rotating machinery remaining useful life prediction scheme using deep-learning-based health indicator and a new RVM
CN113792758A (en) Rolling bearing fault diagnosis method based on self-supervision learning and clustering
CN113268833A (en) Migration fault diagnosis method based on deep joint distribution alignment
CN115169430A (en) Cloud network end resource multidimensional time sequence anomaly detection method based on multi-scale decoding
CN112596016A (en) Transformer fault diagnosis method based on integration of multiple one-dimensional convolutional neural networks
CN113850161A (en) Flywheel fault identification method based on LSTM deep noise reduction self-encoder
CN110717602B (en) Noise data-based machine learning model robustness assessment method
CN112257862A (en) Semi-supervised identification method based on relational network marker sample expansion
CN112115922A (en) Rotary machine fault diagnosis method for enhancing deep feature learning
CN109145764B (en) Method and device for identifying unaligned sections of multiple groups of detection waveforms of comprehensive detection vehicle
CN114626625A (en) Bearing life prediction method based on multi-module U-BiLSTM network
CN115290326A (en) Rolling bearing fault intelligent diagnosis method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant