CN113743534B - Transformer oil gas composite imaging identification method based on depth residual error network - Google Patents

Transformer oil gas composite imaging identification method based on depth residual error network Download PDF

Info

Publication number
CN113743534B
CN113743534B CN202111095213.3A CN202111095213A CN113743534B CN 113743534 B CN113743534 B CN 113743534B CN 202111095213 A CN202111095213 A CN 202111095213A CN 113743534 B CN113743534 B CN 113743534B
Authority
CN
China
Prior art keywords
gas
network
fault
depth residual
threshold
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111095213.3A
Other languages
Chinese (zh)
Other versions
CN113743534A (en
Inventor
胡昊
马鑫
郑野
尚毅梓
张长辉
李新凯
张红涛
李擎
钟凌
职保平
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Yellow River Conservancy Technical Institute
Original Assignee
Yellow River Conservancy Technical Institute
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Yellow River Conservancy Technical Institute filed Critical Yellow River Conservancy Technical Institute
Priority to CN202111095213.3A priority Critical patent/CN113743534B/en
Publication of CN113743534A publication Critical patent/CN113743534A/en
Application granted granted Critical
Publication of CN113743534B publication Critical patent/CN113743534B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/06Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
    • G06N3/061Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using biological neurons, e.g. biological neurons connected to an integrated circuit
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y04INFORMATION OR COMMUNICATION TECHNOLOGIES HAVING AN IMPACT ON OTHER TECHNOLOGY AREAS
    • Y04SSYSTEMS INTEGRATING TECHNOLOGIES RELATED TO POWER NETWORK OPERATION, COMMUNICATION OR INFORMATION TECHNOLOGIES FOR IMPROVING THE ELECTRICAL POWER GENERATION, TRANSMISSION, DISTRIBUTION, MANAGEMENT OR USAGE, i.e. SMART GRIDS
    • Y04S10/00Systems supporting electrical power generation, transmission or distribution
    • Y04S10/50Systems or methods supporting the power network operation or management, involving a certain degree of interaction with the load-side end user applications

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • General Physics & Mathematics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Computational Linguistics (AREA)
  • Neurology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Microelectronics & Electronic Packaging (AREA)
  • Housings And Mounting Of Transformers (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention provides a method for identifying gas in transformer oil by composite imaging based on a deep residual error network, which comprises the following steps: firstly, combining gas concentration, sampling time and sampling point temperature into a new gas characteristic vector, and performing data enhancement and fault evaluation on the gas characteristic vector to obtain a training set and a test set; secondly, optimizing a soft threshold optimization target function in the common channel depth residual shrinkage network by using a cross direction multiplier and a FISTA algorithm to obtain a sub-channel threshold depth residual shrinkage network; and finally, training and testing the training set and the test set respectively by using a subchannel threshold depth residual shrinkage network, and performing visual dimensionality reduction on an output result by using a UMAP algorithm to obtain a fault type. The invention optimizes and improves the network, solves the problem of partial solution loss caused by sparse matrix transposition in the network by utilizing the alternative direction multiplier, and accelerates the output speed of the network by utilizing the FISTA and UMAP algorithms.

Description

Method for recognizing gas composite imaging in transformer oil based on depth residual error network
Technical Field
The invention relates to the technical field of transformer fault identification, in particular to a transformer oil gas composite imaging identification method based on a deep residual error network.
Background
Most of domestic transformers are of an oil-paper combined insulation type, and various faults can occur in daily operation, such as abnormal operation temperature, arc discharge, partial discharge, insulation reduction and the like. When internal failures occur, different types of electrical discharges or temperatures can cause the insulating oil to crack, producing a variety of alkane gases and other gases. Due to the fact that the encapsulation degree of the transformer is high, a maintainer cannot visually observe the internal condition of the transformer, different types of faults are deteriorated and different in damage degree, and casualties can be caused if maintenance measures are taken according to an accident diagnosis result with low accuracy, and therefore the fault identification of the transformer is efficient and accurate.
The general point location and the severity of an accident can be judged by analyzing the components and the concentration of fault gas, and currently, the traditional methods such as a three-ratio method, a Rogers ratio method, a Dornenburg diagnostic method, a DuvalTriangle method and the like based on a DGA method are limited in precision, and some methods are not sensitive to DGA gas data and cannot make accurate diagnosis. In order to effectively improve the accuracy of diagnosis and identification of gas faults in transformer oil, researchers have developed transformer fault identification studies by using statistical analysis methods, machine learning methods and the like. Documents [ Benhamed K, Mooman A, Younnes A, et al. feature Selection for efficient Health indexes of Power transformations [ J].IEEE Transactions onPowerDelivery,2017:1-1.]Through the combination of simulation analysis and a mold, a GRNN method is utilized on the basis of a subsystem, a plurality of key factors influencing the health state of the transformer are obtained, and the fault point position and the accident type of the transformer can be accurately judged. The literature [ G.K.Irngu, A.O.Akumu and J.L.Munda, "Angle Fault Diagnostic Technique in Oil-Filled electric Equipment; the Dual of Duval Triangle, "IEEE Transactions on Dielectrics and electric Insulation, vol.23, No.6, pp.3405-3410, Dec 2016.]Then the pairs of the Deval triangles are adoptedOccasionally, the fault condition of the oil-immersed transformer is diagnosed, but the method can cause the accident identification results to conflict with each other. Literature [ Bacha, Khmais; souahlia, Seifeddine; gossa, Moncef, "Power transducer fault diagnosis based on dispersed gas analysis by supported vector machine," Electric Power Systems Research, vol.83, No.1, pp.73-79, Feb 2012.]The method comprises the steps of classifying dissolved gas by a support vector machine to analyze transformer faults, combining a ratio with an image to serve as an input quantity, and selecting a proper gas index to train, so that the transformer faults can be effectively identified, but the scheme has good effect under the simulation training of small sample data, and identification conflict and ambiguity can be caused if the data size is large; literature [ Li, Jinzhong; zhang, Qiaogen; wang, Ke, et al, "Optimal dispersed Gas Ratios selected genetic Algorithm for Power transducer Fault Diagnosis Based on Support Vector Machine," IEEE Transactions on diagnostics and Electrical Insulation, vol.23, No.2, pp.1198-1206, Apr 2016.]Then, the optimal gas dissolution proportion is obtained by utilizing a genetic algorithm, and the DGA ratio and the optimized parameters are considered, so that a good diagnosis rate can be obtained; therefore, the algorithms can play a certain role in identifying the gas fault in the transformer oil, the problem that the traditional DGA-based method is conflicted can be solved to a certain extent, but the setting of certain key threshold parameters needs to depend on a large amount of data statistics and expert experience, so that the method has certain limitations, and the identification accuracy of part of machine learning and classifier diagnosis methods is greatly reduced under the condition that the parameter threshold is inaccurate. Meanwhile, the DGA method only depends on obtaining the gas concentration to make a judgment, which brings about a large error because the concentration and the growth rate of different gases are different under different conditions (temperature and fault type). For example, when the temperature is below 150 ℃, if partial discharge occurs, CH4The concentration change of (2) is increased and gradually increased along with the temperature rise; when the temperature exceeds 500 ℃, partial discharge is deteriorated to cause arc fault, and then a certain amount of C is generated by cracking2H2And is then CH4The concentration continues to increase while the temperature continuesIncreasing further, above a certain value, CH4A drop in concentration occurs. In addition, insulation reduction and other conditions can occur along with the long-time running and aging of the transformer, fault gas with certain concentration can exist in the oil tank continuously, the preset threshold value does not have complete reliability any more, and for sub-health faults such as insulation reduction or partial discharge and mixed faults, if the sub-health faults and the mixed faults can be identified and found in time in the early period, accident deterioration can be prevented quickly, and influences caused by self difference of the transformers are ignored to a certain extent by the methods.
The deep residual shrinkage network is used as an improved version of the deep learning network, can quickly identify the characteristic information of a sample, and can also avoid the problems of difficulty in training, gradient disappearance and the like caused by large network layer number. The document S.Ma, F.Chu, and Q.Han, "Deep residual leading with modulated time-frequency defects for fault diagnosis of planar gear generating conditions," Mech.Syst.Signal Process, vol.127, pp.190-201,2019 ] makes use of the Deep residual network to diagnose gear operating faults. The scholars verify the advantages of the deep residual error network, the network can identify key information from a large amount of noise interference through training, when gas in transformer oil is subjected to data acquisition and is interfered by external information along with the input of various gas components, the traditional deep residual error network can confuse target characteristics with characteristics of other interference items, and accurate identification cannot be achieved. Therefore, it is important for the present invention that the concentration characteristic information of the target gas can be accurately extracted from a large amount of disturbance information.
Disclosure of Invention
Aiming at the defects in the background technology, the invention provides a transformer oil gas composite imaging identification method based on a deep residual error network, and solves the technical problem that the target characteristic cannot be accurately identified from other interference characteristics in the prior art.
The technical scheme of the invention is realized as follows:
a transformer oil gas composite imaging identification method based on a depth residual error network comprises the following steps:
the method comprises the following steps: acquiring the concentration of dissolved gas in the transformer oil and the temperature of a sampling point according to the sampling time, and recombining the concentration of the dissolved gas according to the sampling time and the temperature of the sampling point to obtain a gas characteristic vector;
step two: performing data enhancement on the gas characteristic vector to be used as an input vector, and performing fault assessment on the input vector to obtain a sample set, wherein the sample set comprises a training set and a testing set;
step three: constructing a soft threshold optimization objective function based on a soft threshold in the common channel depth residual shrinkage network, optimizing the soft threshold optimization objective function by using a cross direction multiplier and a FISTA algorithm, and taking the common channel depth residual shrinkage network corresponding to the optimized soft threshold as a sub-channel threshold depth residual shrinkage network;
step four: inputting the training set into a subchannel threshold depth residual shrinkage network for training to obtain a subchannel threshold depth residual shrinkage network model;
step five: and inputting the test set into a subchannel threshold depth residual shrinkage network model for identification, and performing visual dimensionality reduction on an output result by using a UMAP algorithm to obtain a fault type.
Preferably, the dissolved gas comprises CO, CO2、H2、CH4、C2H2、C2H4And C2H6Seven kinds of gases.
Preferably, the method for obtaining the gas characteristic vector by recombining the concentration of the dissolved gas according to the sampling time and the temperature of the sampling point comprises the following steps: mixing CO and CO2、H2、CH4、C2H2、C2H4And C2H6The concentration of the seven gases is used as a characteristic channel of an image, the temperature and the sampling time of a sampling point are respectively used as the height and the width of the image, and a gas characteristic vector with the dimension of C multiplied by W multiplied by I is obtained, wherein C represents the type of the gas, I represents the temperature interval, namely, a group of gas concentration data is taken when the temperature changes by I degrees, and W represents the corresponding gas concentration data in each sampling timeThe number of sampling points.
Preferably, the method for enhancing the data of the gas feature vector comprises the following steps: the concentration of dissolved gas of the transformer with the same model under a certain working condition and the average value of the concentrations of various gases are used as supplementary data to be randomly added into the gas characteristic vector, and meanwhile, noise information is added into the gas characteristic vector to obtain an input vector.
Preferably, the fault evaluation is performed on the input vector, that is, the fault type is labeled on the input vector, and the fault type includes a low-temperature fault, a medium-temperature overheat fault, a low-energy discharge, a high-energy discharge, a partial discharge, a high-temperature overheat fault, a mixture of an overheat fault and a discharge fault, a moisture fault and a normal state.
Preferably, the expression of the soft threshold in the common-channel depth residual shrinkage network is:
Figure BDA0003268850500000031
where X represents the input characteristic, Y represents the output characteristic, and λ is a positive parameter threshold.
Preferably, the soft threshold optimization objective function is:
Figure BDA0003268850500000032
wherein A, B each represent a constant matrix.
Preferably, the method for optimizing the soft threshold optimization objective function by using the cross direction multiplier and the FISTA algorithm comprises:
let AX-B be Z, rewrite the soft threshold optimization objective function to target I:
Figure BDA0003268850500000033
wherein f (X) | | X | | non-woven phosphor1
Figure BDA0003268850500000041
The augmented Lagrangian function for target I is:
Figure BDA0003268850500000042
wherein, beta is a penalty function, x represents the optimal solution of the local variable, and z represents the optimal solution of the global consistency variable;
the form of the ADMM iterative update is as follows:
Figure BDA0003268850500000043
Figure BDA0003268850500000044
Figure BDA0003268850500000045
wherein, AiAnd BiEach representing the ith element, X, in the corresponding constant matrixiIt is indicated that the (i) th variable,
Figure BDA0003268850500000046
represents the minimization variable XiAt the optimal point after k +1 dual iterations,
Figure BDA0003268850500000047
representing the transpose of the Lagrange multiplier over k dual iterations, i.e. the iteration step, ykLagrange multiplier, y, representing the k-th dual iterationkt1Representing a bivariate update value, ZkRepresenting the variable Z, Z after k dual iterationsiRepresents the ith global consistency optimal solution variable,
Figure BDA0003268850500000048
representing the ith globally consistent optimal solution after k +1 dual iterationsAn optimal point;
using a near-end gradient descent method or a pair of x-derivatives
Figure BDA0003268850500000049
Updating to obtain:
Figure BDA00032688505000000410
wherein, I is an identity matrix,
using soft threshold method pair
Figure BDA00032688505000000411
Updating to obtain:
Figure BDA00032688505000000412
wherein S isλ/βN(. represents a pair
Figure BDA00032688505000000413
A function solved in a simple closed manner by using sub-differential;
wherein the content of the first and second substances,
Figure BDA00032688505000000414
Figure BDA00032688505000000415
the quadratic approximation function at point Z is:
Figure BDA00032688505000000416
wherein the content of the first and second substances,
Figure BDA00032688505000000417
a gradient factor representing a globally consistent optimal solution;
the minimum of the quadratic approximation function is abbreviated as:
Figure BDA00032688505000000418
in conjunction with the abbreviated minimum expression, the minimum of the quadratic approximation function is written as:
Figure BDA00032688505000000419
the operation of FISTA algorithm starts, and L is0Greater than 0, eta > 1, and X is epsilon of RnSimultaneously order Z1=X0,t1=1;
Finding the smallest non-negative integer ikThen, then
Figure BDA0003268850500000051
Figure BDA0003268850500000055
Will be provided with
Figure BDA0003268850500000052
Substituting until the calculation is completed to obtain:
Figure BDA0003268850500000053
Figure BDA0003268850500000054
wherein, tkIndicating the appropriate step size.
Compared with the prior art, the invention has the following beneficial effects: the method combines the gas concentration, the sampling time and the sampling point temperature into a new characteristic gas high-dimensional vector, the characteristic is equivalent to forming an image, and a characteristic channel in the image is the gas concentration; training and identifying the characteristic gas high-dimensional vector by adopting an improved sub-channel threshold depth residual shrinkage network so as to ensure the running speed of the network and the integrity of an output result; the invention optimizes and improves the network, solves the problem of partial solution loss caused by sparse matrix transposition in the network by utilizing the alternative direction multiplier, and accelerates the output speed of the network by utilizing the FISTA and UMAP algorithms.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
Fig. 1 shows basic residual units.
Fig. 2 is a data structure of an input characteristic of the present invention.
FIG. 3 is a structure of a common channel depth residual shrinking network of the present invention; wherein, (a) is a basic module of the depth residual shrinkage network, and (b) is a schematic diagram of the whole structure.
Fig. 4 is an enhancement mode of the training set of the present invention.
FIG. 5 is a parameter setting model of the sub-channel threshold depth residual shrinkage network of the present invention.
FIG. 6 is a flow chart of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be obtained by a person skilled in the art without inventive effort based on the embodiments of the present invention, are within the scope of the present invention.
The initial deep neural network is often faced with a plurality of problems during learning, when the training parameters are too much, the network degradation is caused due to the increase of the number of network layers, especially when the training parameters are more complicated and have too high dimension, the objective function comprises a plurality of non-convex optimization solving problems, and the like, and the network training reduction effect is more obvious. In order to solve the problem, a deep residual error network is proposed in the documents [ HEKM, ZHANGXY, RENSQ, et al.Depressedusalllarningforimage Recognition [ C ]//2016IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 27-30,2016.Las Vegas, NV, USA.IEEE,2016: 770-. The basic residual unit is shown in fig. 1.
The deep residual error network is used as one of convolutional neural networks, can train strong noise data to further achieve the purpose of noise reduction, can be used as a basic network for deep learning, and meanwhile, a soft threshold value of the deep residual error network can be used as a self-adaptive threshold value for filtering noise to achieve the purpose of active noise reduction, so that the defects of stiffness and non-universal applicability of manual threshold value setting can be well overcome. The soft threshold method of the network is the key point for removing interference, and the expression is as follows:
Figure BDA0003268850500000061
where x represents the input characteristic, y represents the output characteristic, and τ is the positive parameter threshold. The operation idea is to set the eigenvalue close to zero in the activation function to zero so as to achieve the purpose of not deleting the negative eigenvalue. The processed derivative inputs are as follows:
Figure BDA0003268850500000062
it may zero out those features whose absolute values are below the threshold and run other features toward zero as well. As a non-linear variation, similar to the ReLU activation function, may also be used as the activation function. The network acquires useful information through global scanning, and the useful information is enhanced while redundant information is suppressed. Compared with the existing traditional neural network, the network also has the structures of convolution layers, activation functions, normalization, cross entropy error functions and the like. The convolution layer is used for replacing matrix multiplication, so that the number of training parameters can be reduced to a great extent, and the overfitting condition can be avoided, so that higher testing precision is obtained. This convolution is represented as:
Figure BDA0003268850500000063
wherein x isiIs the ith channel, y of the input element mapjIs the channel of the output profile, kijIs a convolution kernel, bjIs a deviation, MjIs the set of channels, which is the fifth channel used to compute the output feature map. The convolution may be repeated multiple times to obtain an output signature. In order to reduce the occurrence of covariance in the network, a batch normalization function (BN) is added into the model, when new parameters are continuously input, the network updates the feature information by continuously training the new parameters, the original feature distribution can be continuously changed, and after the BN is added, the model can perform normalization processing on the features captured before and after the model is added so as to ensure that the model can adapt to the continuous change of the features. To address this, it is necessary to keep the parameters in the convolutional layer constantly changing. The primary function of BN is to change each feature into a normalized distribution and then adjust the features to an ideal distribution during ongoing training. The process is as follows:
Figure BDA0003268850500000071
Figure BDA0003268850500000072
Figure BDA0003268850500000073
Figure BDA0003268850500000074
wherein x isnAnd ynRepresenting the input and output characteristics of BN, γ and β are two trainable parameters to scale and move the distribution, and e is a constant close to zero.
Activation functions are not available as common non-linear changes in neural networks, preventing the gradient from disappearing to some extent. Its function is represented as:
y=max(x,0)
where x and y represent the input and output of the activation function, respectively.
For each channel of the profile, the network may calculate a gas mean using GAP. And the objective function uses a cross entropy function, and aims to reduce different types of recognition tasks, thereby bringing higher training success rate. In order to calculate the cross entropy error, it is necessary to bring the eigenvalues within the (0,1) range with the softmax function. The expression is as follows:
Figure BDA0003268850500000075
wherein x isjAnd yjIs the input and output characteristic diagram of the softmax function: i and j are neurons of the output layer: n is a radical of hydrogenctassIs a class number. Here, yjIt can be seen as observing the prediction probability belonging to the j-th class. The error of the cross entropy function is as follows:
Figure BDA0003268850500000076
wherein, tjAs target output, tjIs the jth class output.
As shown in fig. 6, a method for identifying composite imaging of gas in transformer oil based on a depth residual error network specifically comprises the following steps:
the method comprises the following steps: acquiring the concentration of dissolved gas in the transformer oil and the temperature of a sampling point according to the sampling time, and recombining the concentration of the dissolved gas according to the sampling time and the temperature of the sampling point to obtain a gas characteristic vector; the dissolved gas comprises CO and CO2、H2、CH4、C2H2、C2H4And C2H6Seven kinds of gases. The method for recombining the concentration of the dissolved gas according to the sampling time and the temperature of the sampling point to obtain the gas characteristic vector comprises the following steps: mixing CO and CO2、H2、CH4、C2H2、C2H4And C2H6The concentrations of the seven gases are used as a characteristic channel of an image, the temperature and the sampling time of a sampling point are respectively used as the height and the width of the image, and a gas characteristic vector with the dimension of C multiplied by W multiplied by I is obtained, wherein C represents the type of the gas, I represents the temperature interval, namely, a group of gas concentration data is taken when the temperature changes by I degree, and W represents the corresponding number of sampling points in each sampling time.
Due to the fact that the characteristics of the transformers are different from the operating environment, any transformer has a unique working condition when in operation. These factors include voltage level, structure, materials, connection to temperature, humidity, age, load, etc. According to the DGA gas analysis method widely applied at present, a definite factor aiming at transformer fault diagnosis is fault characteristic gas concentration, wherein the improved characteristic gas three-ratio method is used for calculating the characteristic gas concentration ratio to analyze whether the transformer has faults or not. Taking into account most of the characteristics of each transformer, a cumbersome process of statistical data classification is required. In order to accurately analyze the operation condition of each transformer and avoid the problem that big data covers individual differences, the temperature and the operation time which are directly related to the gas concentration are selected as key characteristic factors, so that the gas concentration change of each transformer is generated according to the conditions and the influence of the temperature, and a plurality of factors do not need to be classified and analyzed. The basis for determining the temperature as a key attribute is that the molecular diffusion coefficient of a substance represents the diffusion capacity of the substance, and according to Fick's law, the diffusion coefficient passes through the mass of a certain unit area of the substance in a unit time along the diffusion direction, while the inside of the transformer is a mixture of multiple gases, namely, mixed gas diffusion, and the diffusion coefficient is expressed as:
Figure BDA0003268850500000081
wherein T represents a thermodynamic temperature, P represents a pressure, μA,μBRepresents the molecular weight of the gas and V represents the molar volume of the gas under normal conditions.
It can be seen that without a change in pressure, the temperature increases and the diffusion rate of the gas increases. At the same time, the concentration of the gas changes with time. The method selects temperature and time as the associated characteristic attributes of the gas concentration, can also ensure that the problem of overhigh dimensionality cannot be brought to the network when the associated characteristic attributes are used as the input characteristics of the deep learning network, can continuously adjust and adapt the obtained result according to the self characteristics of the transformer when the operation data of each transformer is used for carrying out deep learning network training aiming at the transformer, can solve the problem of neglecting the difference between the transformers due to large data coverage, and is equivalent to the action of controlling variables.
And selecting proper fault gas as an input characteristic vector, representing the change situation of each gas concentration of the transformer on a high-dimensional time sequence along with the temperature change during operation at a certain voltage level, and similar to the parameter characteristic structure of an image. As shown in fig. 2: the variable axis of the invention is equivalent to the concentration of several characteristic Gases during the operation of the transformer, and the invention is based on the IEEE Guide for the Interpretation of gas Generated in Oil-impregnated transducers (IEEE Std C57.104)TM-2008)issued by IEEE Power&Energy Society, selecting fault characteristic gases CO and CO2、H2、CH4、C2H2、C2H4、C2H6The characteristic that the characteristic gas is equivalent to a characteristic channel of image properties, and the temperature and time axes are similar to the height and width axes of an image enables the invention to learn the characteristics of fault gas in transformer oil by utilizing a deep residual error network in the field of image classification, and fault classification and prediction results of each transformer under different temperatures and operation times can be obtained by processing a gas concentration characteristic sequence. The initial input characteristics are shown in table 1.
TABLE 1 input characteristics
Figure BDA0003268850500000091
The input data is formed into a format of C multiplied by W multiplied by I, wherein C represents the fault gas quantity of the transformer and comprises CO and CO2、H2、CH4、C2H2、C2H4、C2H6The total of 7 gases are used as characteristic channels, and 7 gases are in a group, so the number of the channels is 7. I represents a temperature interval which is set to be 1 ℃, namely, a group of gas concentration data is taken when the temperature changes once, the value can be modified according to different monitoring equipment specifications, W represents the number of data acquisition points per minute, the frequency of the data acquisition is 10s/min, the number of sampling points is 6, and the time interval is 10 s. This configuration ensures that the present invention is able to identify and predict characteristic gas signals hierarchically on a time scale as temperature changes. The selection of the variable can also be the phase velocity and the absolute velocity of gas diffusion, and the invention adopts the gas concentration as an observation attribute to research, thereby avoiding the interference of gases with each other and the situation that the gases are dissolved in oil. In addition, under different work condition, the concentration that takes place its corresponding gas of different types trouble also can change to characteristic gas concentration is as the acquisition signal, and irrelevant gas concentration forms the condition of interference as the noise when also can handling the trouble, makes the network focus on the characteristics of trouble target gas more.
Step two: performing data enhancement on the gas characteristic vector to be used as an input vector, and performing fault assessment on the input vector to obtain a sample set, wherein the sample set comprises a training set and a testing set; the method for enhancing the data of the gas characteristic vector comprises the following steps: the concentration of dissolved gas of the transformer with the same model under a certain working condition and the average value of the concentrations of various gases are used as supplementary data to be randomly added into the gas characteristic vector, and meanwhile, noise information is added into the gas characteristic vector to obtain an input vector. And performing fault evaluation on the input vector, namely labeling fault types of the input vector, wherein the fault types comprise low-temperature faults, medium-temperature overheat faults, low-energy discharge, high-energy discharge, partial discharge, high-temperature overheat faults, overheat faults and discharge faults mixed, insulation moisture faults (mixed electric heating faults) and normal states.
In the field of image recognition, proper rotation of the image does not change the characteristics of the image, but can be used as new data to input a rich data set, so that the aims of making up the prior information loss and considering the anti-noise interference are fulfilled, and the method can prevent the occurrence of overfitting to a certain extent. In order to enhance the robustness of the model, the invention ensures the sufficient data of the input network and the reliability of the model through data enhancement, so that the data of the transformer with the same model under a certain working condition is added into a data set as supplement randomly, and the input of the random variable can be the average value of the related variable. Meanwhile, if the added noise data is small in assignment, the network identification and classification result cannot be changed, and when the data set is small, the noise information is added to play a role in enriching the data set. The training set after the initial parameter enhancement is shown in fig. 3.
For the initial set of training parameters, gaussian white noise H (0, ∈) is added to the input variable to obtain the latest training parameters. Then the new training parameter input variable X is:
X*=X×(1tδ),δ~H(0,∈)
wherein, epsilon is randomly generated from 0 to H, and the value of H can be artificially set according to the past statistical data. By enhancing the input characteristic data, the training set under the condition of information loss can be expanded by e times. However, it should be noted that the value of e is selected according to the training effect, and should not be too large, otherwise the calculation efficiency is affected. And obtaining an enhanced training feature sample set through processing the input feature parameters.
And correspondingly dividing the sample into a training set and a testing set according to the model of the transformer and the voltage grade. Considering routine maintenance and incomplete data of partial new transformers, the reference data is enhanced to supplement samples, and the experiment is carried out by adopting a cross validation method, wherein each data set is divided into 10 subsets, 1 subset is used as a test set, and 9 subsets are used as training sets. And the transformer with faults in the experimental process is subjected to fault assessment and maintenance by workers in the station. In the experiment, a network model is constructed by using a python language, and hardware is a computer carrying an i7-9750 central processing unit and an NVIDIA Geforce GTX processor.
For different types of fault problems, concentration change conditions of seven gases at different temperatures are considered in the experiment, and fault classifications are shown in table 2:
TABLE 2 nine experimental states of transformer DGA failure
Figure BDA0003268850500000101
Step three: constructing a soft threshold optimization objective function based on a soft threshold in the common channel depth residual shrinkage network, optimizing the soft threshold optimization objective function by using a cross direction multiplier and a FISTA algorithm, and taking the common channel depth residual shrinkage network corresponding to the optimized soft threshold as a sub-channel threshold depth residual shrinkage network;
the common channel depth residual error network adopted by the invention is provided by the documents [ ZHAO M, ZHONG S, FUX, et al. deep residual reducing transformer networks for fault diagnosis [ J ]. IEEE Transactions on Industrial information, 2020,16(7):4681 and 4690 ] when the bearing fault diagnosis is carried out, the network can effectively diagnose the fault, and the structure of the common channel depth residual error shrinkage network is shown in figure 3.
The common-channel depth residual shrinkage network comprises an input layer, a convolution layer, a depth residual shrinkage network basic module, a global mean pooling layer and a fully-connected output layer, and the construction process is as follows:
firstly, constructing a basic module of a depth residual shrinkage network, embedding a sub-network in the basic module for automatically setting a threshold value required by soft thresholding, adopting a residual shrinkage module for sharing the threshold value among channels, and firstly carrying out batch standardization, ReLu activation function and convolution layer operation on an input feature map twice; then, calculating absolute values of all the characteristics, and taking the average value as the characteristic; in the other path, after the absolute value of the feature is subjected to global mean pooling, the absolute value is input into a two-layer full-connection network, then the output is normalized to be between 0 and 1 by using a Sigmoid function, a scale parameter alpha is obtained, and the final threshold value is represented as alpha multiplied by F; finally, adding the original input characteristic graph and the threshold value by using the identity in Tensorhow and returning;
the structure input layer receives external input of the neural network model and transmits the external input to the convolutional layer, wherein the external input is a gas characteristic vector;
constructing a convolutional layer, receiving the output of an input layer by the convolutional layer, repeating the convolution operation to obtain a feature map, and transmitting the feature map to a depth residual shrinkage network basic module;
stacking a depth residual shrinkage network basic module, wherein the characteristic graph output by the convolution layer is processed by the depth residual shrinkage network module and then transmitted to a batch standardization layer;
constructing a batch standardization layer, and activating a function ReLu and a global averaging layer;
constructing a fully connected output layer, receiving the output from the global averaging layer;
and the fully-connected output layer corresponds to all classes contained in the sample data, the output value is the probability value of the sample belonging to each class, and the class corresponding to the maximum output value is taken as the sample class of model prediction.
The obvious difference between the common-channel depth residual shrinking network and the traditional depth residual network is that a soft threshold is introduced, and the soft threshold is used as an effective noise signal removing method and is also a key step of the network. The principle is to set the converted signal in the domain close to 0, and it can adjust adaptively along with the change of the input characteristic parameter. The traditional common signal noise removing method comprises wavelet denoising, hard threshold denoising and soft threshold denoising, but the hard threshold denoising can generate the conditions of jitter, unsmooth and the like, and more importantly, the hard threshold denoising needs to determine a threshold by means of expert experience or statistics, the threshold cannot have objectivity to a certain extent and is stiff, generally speaking, along with the increase of the running time of a transformer, the reduction of internal insulation and loss of the transformer can be aggravated, and the setting of the hard threshold can generate the defect of idealization. The conventional wavelet denoising requires a great deal of signal processing skills to be inherited, and is difficult. Therefore, denoising with a gradient-decreasing soft threshold is an optimal choice. The expression of the soft threshold in the common-channel depth residual shrinkage network is as follows:
Figure BDA0003268850500000121
where X represents the input characteristic, Y represents the output characteristic, and λ is a positive parameter threshold.
However, when used, the function of the soft threshold is fundamentally related to the transpose of a and x. Meanwhile, although the soft threshold has continuous smoothness, its constant deviation may cause partial data omission and may have a problem of lower signal-to-noise ratio than the hard threshold. The present invention optimizes this problem by using backtracking in the cross direction multipliers. Meanwhile, in order to accelerate the problem of the speed of solving the soft threshold value under a large amount of data, the FISTA algorithm is added in the network soft threshold value solving.
The soft threshold optimization objective function is:
Figure BDA0003268850500000122
wherein, A and B both represent any constant matrix and can be set according to actual conditions.
According to the definition of norm, decomposing the expression of the soft threshold optimization objective function, and comparing the relation between B and lambda to obtain the general form solution;
Figure BDA0003268850500000123
the result is a matrix of soft thresholds, which is solved for and optimized for a number of independent functions. Therefore, the method relates to the transposition problem of A and x in a sparse matrix in high-dimensional feature solution and the solution of a large-scale convex optimization problem. In the threshold determination process, the network is solved in a manner that seeks local optima. Because the function progressiveness can cause deviation to exist all the time, the method can not accurately and flexibly judge the accurate relation between the measured value and the fault threshold value all the time. For example, 1 is set as the failure threshold, and 0.9 and 0.99 are measured values, respectively, but under different standards, 0.9 may be determined as the early warning state, and 0.99 may be determined as the normal state. Therefore, this locally sought solution tends to result in edge solutions being lost and constant deviations being amplified during the rank-conversion process.
Aiming at the solving problem of the convex optimization problem, each subproblem is optimized and solved by utilizing a cross direction multiplier, and finally, the optimal solution of the subproblems is coordinated and a global coordinated optimal solution is finally obtained, so that incompleteness and feature omission of the soft threshold can be solved to a certain extent. The method comprises the following specific steps:
let AX-B be Z, rewrite the soft threshold optimization objective function to target I:
Figure BDA0003268850500000124
wherein f (X) | | X | | non-woven phosphor1
Figure BDA0003268850500000125
The augmented Lagrangian function for target I is:
Figure BDA0003268850500000126
wherein, beta is a penalty function, x represents a local variable, and z represents a global consistency variable; according to the constraint conditions, each local variable needs to be kept consistent, namely the whole data set is divided into each node, each node is trained independently, and finally a global consistent solution is converged. The determination mode of the rewritten fault threshold value is more accurate and flexible. For example: when the fault threshold is 1, the model captures a state difference value of 0.1 after identifying a large number of samples, and if the difference between the measured value and the fault threshold is greater than 0.1, the model is determined to be in a normal state. The weight (threshold) precision of each characteristic gas in the fault can be effectively improved by matching with the self-adaptive variable weight cross entropy function, and the threshold can be ensured to be self-updated along with the input of new data and the state change of equipment.
The form of the ADMM iterative update is as follows:
Figure BDA0003268850500000131
Figure BDA0003268850500000132
Figure BDA0003268850500000133
wherein A isiAnd BiEach representing the ith element, X, in the corresponding constant matrixiIt is indicated that the (i) th variable,
Figure BDA0003268850500000134
represents the minimization variable XiAt the optimal point after k +1 dual iterations,
Figure BDA0003268850500000135
representing the transpose of the Lagrangian multiplier over k dual iterations, i.e. the iteration step, ykLagrange representing the k-th dual iterationDaily multiplier, ykt1Representing a bivariate update value, ZkRepresenting the variable Z, Z after k dual iterationsiRepresents the ith global consistency optimal solution variable,
Figure BDA0003268850500000136
and (4) representing the optimal point of the ith global consistency optimal solution after k +1 dual iterations.
Using a near-end gradient descent method or a pair of x-derivatives
Figure BDA0003268850500000137
Updating to obtain:
Figure BDA0003268850500000138
wherein I is an identity matrix;
using soft threshold method pair
Figure BDA0003268850500000139
Updating to obtain:
Figure BDA00032688505000001310
wherein S isλp.beta.N (-) represents a pair
Figure BDA00032688505000001311
The function is solved using a simple closed form of sub-differentiation.
Wherein the content of the first and second substances,
Figure BDA00032688505000001312
it can be seen that in the conventional soft threshold calculation method, each iteration of the step needs a times its transpose, and then a soft threshold step is performed, and by continuous near-end gradient descent, a local minimum is obtained when the absolute value of the gradient is infinitely close to 0. In the operation, the Lipschitz constant of the fixed step length is not necessarily known to be calculated, the constant depends on the transposition of a noise matrix multiplied by the maximum value of a noise signal, the optimization problem of the L1 norm is also involved, when the input data is large, the calculation amount is also large, and in order to solve the problem, the soft threshold value is calculated through the FISTA algorithm with the backtracking function.
Figure BDA00032688505000001313
The quadratic approximation function at point Z is:
Figure BDA00032688505000001314
wherein the content of the first and second substances,
Figure BDA00032688505000001315
gradient factors representing a globally consistent optimal solution.
Since the constant term has no effect on the result, neglecting the constant term, the minimum of the quadratic approximation function is abbreviated as:
Figure BDA0003268850500000141
in conjunction with the abbreviated minimum expression, the minimum of the quadratic approximation function is written as:
Figure BDA0003268850500000142
the operation of FISTA algorithm is started, and L is0Greater than 0, eta > 1, and X ∈ RnSimultaneously order Z1=X0,t1=1;
Finding the smallest non-negative integer ikThen, then
Figure BDA0003268850500000143
Figure BDA0003268850500000148
Will be provided with
Figure BDA0003268850500000144
Substituting until the calculation is completed to obtain:
Figure BDA0003268850500000145
Figure BDA0003268850500000146
wherein, tkIndicating the appropriate step size.
The setting of the hyperparameters such as the number of network layers, the number of convolution kernels and the like does not have an optimal standard at present, so the method is set according to the conventional proposal at present. The parameter setting of the sub-channel threshold depth residual shrinkage network adopted by the transformer fault identification is based on a framework of ResNet34, as shown in FIG. 5, the parameter setting of each residual block in FIG. 5 is shown in Table 3, and the parameter setting of the sub-channel threshold depth residual shrinkage network is completed in a parameter debugging stage through the process of FIG. 5.
TABLE 3 setting of residual blocks
Figure BDA0003268850500000147
Step four: inputting the training set into a subchannel threshold depth residual shrinkage network for training to obtain a subchannel threshold depth residual shrinkage network model; model parameters (reference hyper-parameter settings) are initialized. The output of the network was determined to be 9 neurons containing 1 normal state and 8 fault states as shown in table 2. In the training process, the training rate changes with the change of the period, from the first 20 time periods to 0.1, the next 20 time periods to 0.03, the middle 20 time periods to 0.01 and 0.003, and the last 20 time periods to 0.001, in order to ensure that the parameters are updated in step size at the beginning of the training to obtain the optimal output. The penalty term is added into the objective function in order to avoid overfitting in the training process by matching with L2 regularization, and the penalty coefficient is set to be 0.0001 adopted by the traditional deep neural network. Finally, there are 9 neurons at the output level, i.e. 1 healthy state and 8 faulty states.
Step five: and inputting the test set into a subchannel threshold depth residual shrinkage network model for identification, and performing visual dimensionality reduction on an output result by using a UMAP algorithm to obtain a fault type.
Specific examples
In 29/2020 to 11/8/2020, the method of the present invention was tested in field operation on transformers from some branch of the national grid, and the test results are shown in table 4.
TABLE 4 Transformer characteristic gas concentration
Figure BDA0003268850500000151
The proposed identification network of the invention emits an alarm signal at a date 08:30AM, which indicates the possible presence of arcing therein. 30 minutes in the afternoon, the power plant monitoring equipment only captures this signal due to heavy gas protection, transformer tripping. The staff then analyzes the oil chromatogram data inside the transformer and concludes that the concentration of the fault gas conforms to the three-ratio code combination 102, which means that an arc fault has occurred inside the transformer. The cause of this failure is the cracking of the insulating mat due to its own problems. With the operation of the transformer and the change of the load (the main transformer No.1 returns to the factory for maintenance, the load of the transformer No.2 increases), the insulation damage is accelerated, then a short circuit phenomenon occurs, the fault current may reach several hundred amperes, and high temperature is generated to burn out the faulty insulation pad, and finally, arc discharge is caused. When an alarm signal is sent out, the internal fault already enters the hatching stage, the characteristic gas in the transformer slightly changes, but the concentration does not meet the conditions of the three-ratio method and the accuracy requirement of the monitoring system.
In 6 noon, 7/2020, the method monitors that the concentrations of hydrogen and alkane gas in the transformer are both increased greatly, and according to the operation regulation and equipment observation results of the transformer, the method estimates that water enters the transformer, continuous spark discharge (short circuit) exists in the transformer, and the concentration change of the fault potential gas is obvious. In the evening, the transformer explodes. The power-off maintenance discovers that the internal insulation of the transformer is damaged, and a large amount of rainwater permeates into the transformer to cause the accident.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents, improvements and the like that fall within the spirit and principle of the present invention are intended to be included therein.

Claims (7)

1. A transformer oil gas composite imaging identification method based on a depth residual error network is characterized by comprising the following steps:
the method comprises the following steps: acquiring the concentration of dissolved gas in the transformer oil and the temperature of a sampling point according to the sampling time, and recombining the concentration of the dissolved gas according to the sampling time and the temperature of the sampling point to obtain a gas characteristic vector;
step two: performing data enhancement on the gas characteristic vector to be used as an input vector, and performing fault assessment on the input vector to obtain a sample set, wherein the sample set comprises a training set and a testing set;
step three: constructing a soft threshold optimization objective function based on a soft threshold in the common channel depth residual shrinkage network, optimizing the soft threshold optimization objective function by using a cross direction multiplier and a FISTA algorithm, and taking the common channel depth residual shrinkage network corresponding to the optimized soft threshold as a sub-channel threshold depth residual shrinkage network;
step four: inputting the training set into a subchannel threshold depth residual shrinkage network for training to obtain a subchannel threshold depth residual shrinkage network model;
step five: inputting the test set into a subchannel threshold depth residual shrinkage network model for identification, and performing visual dimensionality reduction on an output result by using a UMAP algorithm to obtain a fault type;
the method for optimizing the soft threshold optimization objective function by using the cross direction multiplier and the FISTA algorithm comprises the following steps:
let AX-B be Z, rewrite the soft threshold optimization objective function to target I:
Figure FDA0003537567510000011
s.t AX-Z=B;
wherein f (X) | | X | | non-woven phosphor1
Figure FDA0003537567510000012
The augmented Lagrangian function for target I is:
Figure FDA0003537567510000013
wherein, beta is a penalty function, x represents the optimal solution of the local variable, and z represents the optimal solution of the global consistency variable;
the form of iterative update of ADMM is as follows:
Figure FDA0003537567510000014
Figure FDA0003537567510000015
Figure FDA0003537567510000016
wherein A isiAnd BiEach representing the ith element, X, in the corresponding constant matrixiIt is indicated that the (i) th variable,
Figure FDA0003537567510000017
represents the minimization variable XiThroughThe optimum point after k +1 dual iterations,
Figure FDA0003537567510000018
representing the transpose of the Lagrange multiplier over k dual iterations, i.e. the iteration step, ykLagrange multiplier, y, representing the k-th dual iterationk+1Representing a bivariate update value, ZkRepresenting the variable Z, Z after k dual iterationsiRepresents the ith global consistency optimal solution variable,
Figure FDA0003537567510000019
the optimal point of the ith global consistency optimal solution after k +1 dual iterations is represented;
using a near-end gradient descent method or a pair of x-derivatives
Figure FDA00035375675100000110
Updating to obtain:
Figure FDA00035375675100000111
wherein, I is an identity matrix,
using soft threshold method pair
Figure FDA0003537567510000021
Updating to obtain:
Figure FDA0003537567510000022
wherein S isλ/βN(. represents a pair
Figure FDA0003537567510000023
A function of simple closed solution by using sub-differential;
wherein the content of the first and second substances,
Figure FDA0003537567510000024
Figure FDA0003537567510000025
the quadratic approximation function at point Z is:
Figure FDA0003537567510000026
wherein the content of the first and second substances,
Figure FDA0003537567510000027
a gradient factor representing a globally consistent optimal solution;
the minimum of the quadratic approximation function is abbreviated as:
Figure FDA0003537567510000028
in conjunction with the abbreviated minimum expression, the minimum of the quadratic approximation function is written as:
Figure FDA0003537567510000029
the operation of FISTA algorithm starts, and L is0Greater than 0, eta > 1, and X ∈ RnSimultaneously order Z1=X0,t1=1;
Finding the smallest non-negative integer ikThen, then
Figure FDA00035375675100000210
Figure FDA00035375675100000211
Will be provided with
Figure FDA00035375675100000212
Substituting until the calculation is completed to obtain:
Figure FDA00035375675100000213
Figure FDA00035375675100000214
wherein, tkIndicating the appropriate step size.
2. The method for identifying the gas in transformer oil based on the deep residual error network as claimed in claim 1, wherein the dissolved gas comprises CO and CO2、H2、CH4、C2H2、C2H4And C2H6Seven kinds of gases.
3. The transformer oil gas composite imaging identification method based on the depth residual error network according to claim 2, wherein the method for recombining the concentration of dissolved gas according to the sampling time and the sampling point temperature to obtain the gas characteristic vector comprises the following steps: mixing CO and CO2、H2、CH4、C2H2、C2H4And C2H6The concentrations of the seven gases are used as a characteristic channel of an image, the temperature and the sampling time of a sampling point are respectively used as the height and the width of the image, and a gas characteristic vector with the dimension of C multiplied by W multiplied by I is obtained, wherein C represents the type of the gas, I represents the temperature interval, namely, a group of gas concentration data is taken when the temperature changes by I degree, and W represents the corresponding number of sampling points in each sampling time.
4. The method for identifying the gas in transformer oil composite imaging based on the depth residual error network as claimed in claim 2 or 3, wherein the method for enhancing the gas feature vector comprises: the concentration of dissolved gas of the transformer with the same model under a certain working condition and the average value of the concentrations of various gases are used as supplementary data to be randomly added into the gas characteristic vector, and meanwhile, noise information is added into the gas characteristic vector to obtain an input vector.
5. The method for identifying the gas in the transformer oil based on the deep residual error network as claimed in claim 1, wherein the fault evaluation is performed on the input vector, that is, the fault type is labeled on the input vector, and the fault type includes a low temperature fault, a medium temperature overheating fault, a low energy discharge, a high energy discharge, a partial discharge, a high temperature overheating fault, a mixture of the overheating fault and the discharge fault, a moisture fault and a normal state.
6. The method for identifying the gas in transformer oil composite imaging based on the depth residual error network as claimed in claim 1, wherein the expression of the soft threshold in the common channel depth residual error shrinkage network is as follows:
Figure FDA0003537567510000031
where X represents the input characteristic, Y represents the output characteristic, and λ is a positive parameter threshold.
7. The method for identifying the gas in transformer oil composite imaging based on the depth residual error network as claimed in claim 6, wherein the soft threshold optimization objective function is:
Figure FDA0003537567510000032
wherein A, B each represent a constant matrix.
CN202111095213.3A 2021-09-17 2021-09-17 Transformer oil gas composite imaging identification method based on depth residual error network Active CN113743534B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111095213.3A CN113743534B (en) 2021-09-17 2021-09-17 Transformer oil gas composite imaging identification method based on depth residual error network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111095213.3A CN113743534B (en) 2021-09-17 2021-09-17 Transformer oil gas composite imaging identification method based on depth residual error network

Publications (2)

Publication Number Publication Date
CN113743534A CN113743534A (en) 2021-12-03
CN113743534B true CN113743534B (en) 2022-06-07

Family

ID=78739843

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111095213.3A Active CN113743534B (en) 2021-09-17 2021-09-17 Transformer oil gas composite imaging identification method based on depth residual error network

Country Status (1)

Country Link
CN (1) CN113743534B (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111797567A (en) * 2020-06-09 2020-10-20 合肥工业大学 Deep learning network-based bearing fault classification method and system
CN112115638A (en) * 2020-08-28 2020-12-22 合肥工业大学 Transformer fault diagnosis method based on improved Adam algorithm optimization neural network

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10657446B2 (en) * 2017-06-02 2020-05-19 Mitsubishi Electric Research Laboratories, Inc. Sparsity enforcing neural network
CN111832637B (en) * 2020-06-30 2022-08-30 南京邮电大学 Distributed deep learning classification method based on alternating direction multiplier method ADMM
CN112163574A (en) * 2020-11-23 2021-01-01 南京航天工业科技有限公司 ETC interference signal transmitter identification method and system based on deep residual error network

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111797567A (en) * 2020-06-09 2020-10-20 合肥工业大学 Deep learning network-based bearing fault classification method and system
CN112115638A (en) * 2020-08-28 2020-12-22 合肥工业大学 Transformer fault diagnosis method based on improved Adam algorithm optimization neural network

Also Published As

Publication number Publication date
CN113743534A (en) 2021-12-03

Similar Documents

Publication Publication Date Title
Phillips et al. Classifying machinery condition using oil samples and binary logistic regression
Mao et al. Investigation of polymer electrolyte membrane fuel cell internal behaviour during long term operation and its use in prognostics
Long et al. Improved diagnostics for the incipient faults in analog circuits using LSSVM based on PSO algorithm with Mahalanobis distance
CN111210024A (en) Model training method and device, computer equipment and storage medium
CN109460618A (en) A kind of rolling bearing remaining life on-line prediction method and system
CN110929847A (en) Converter transformer fault diagnosis method based on deep convolutional neural network
CN109000930A (en) A kind of turbogenerator performance degradation assessment method based on stacking denoising self-encoding encoder
CN106446317A (en) Mathematic model-based sealed relay storage life prediction method
CN113782113B (en) Method for identifying gas fault in transformer oil based on deep residual error network
CN115225516B (en) LSSVM network flow prediction method based on improved ABC-VMD
Jha et al. Application of artificial intelligence techniques for dissolved gas analysis of transformers-A review
CN111273125A (en) RST-CNN-based power cable channel fault diagnosis method
Kaur et al. Fault detection in power transformers using random neural networks
CN104077493A (en) Method for constructing state evaluation index system of electric relaying protection system
CN114462508A (en) Power transformer health state assessment method based on multi-mode neural network
CN116610998A (en) Switch cabinet fault diagnosis method and system based on multi-mode data fusion
Lu et al. Grey relational analysis using Gaussian process regression method for dissolved gas concentration prediction
CN117371207A (en) Extra-high voltage converter valve state evaluation method, medium and system
Xing et al. Health evaluation of power transformer using deep learning neural network
CN111586728A (en) Small sample characteristic-oriented heterogeneous wireless network fault detection and diagnosis method
Sadoughi et al. A deep learning approach for failure prognostics of rolling element bearings
Leauprasert et al. Intelligent machine learning techniques for condition assessment of power transformers
CN114595883A (en) Oil-immersed transformer residual life personalized dynamic prediction method based on meta-learning
US11693041B2 (en) Method for monitoring the electric insulation status of a piece of equipment for MV or HV electric systems
CN113743534B (en) Transformer oil gas composite imaging identification method based on depth residual error network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant