CN112699244A - Deep learning-based method and system for classifying defect texts of power transmission and transformation equipment - Google Patents

Deep learning-based method and system for classifying defect texts of power transmission and transformation equipment Download PDF

Info

Publication number
CN112699244A
CN112699244A CN202110279537.6A CN202110279537A CN112699244A CN 112699244 A CN112699244 A CN 112699244A CN 202110279537 A CN202110279537 A CN 202110279537A CN 112699244 A CN112699244 A CN 112699244A
Authority
CN
China
Prior art keywords
vector
transformation equipment
text
power transmission
word
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110279537.6A
Other languages
Chinese (zh)
Inventor
张葛祥
朱明�
王茜
杨强
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chengdu University of Information Technology
Original Assignee
Chengdu University of Information Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chengdu University of Information Technology filed Critical Chengdu University of Information Technology
Priority to CN202110279537.6A priority Critical patent/CN112699244A/en
Publication of CN112699244A publication Critical patent/CN112699244A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Biophysics (AREA)
  • Evolutionary Computation (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Databases & Information Systems (AREA)
  • Machine Translation (AREA)

Abstract

The invention provides a deep learning-based method and a deep learning-based system for classifying defect texts of power transmission and transformation equipment, wherein the method comprises the following steps: s1: preprocessing the acquired defective text of the power transmission and transformation equipment, and then embedding words to obtain a first word vector with electric power semantic features; s2: acquiring forward and backward characteristic information of a defect text of the power transmission and transformation equipment through a bidirectional long-time and short-time memory network, and outputting a hidden layer state vector; s3: carrying out weighted transformation on the hidden layer state vector by using a self-attention mechanism to obtain deep semantic features and obtain a final sentence vector to be classified; s4: and outputting the vectors to be classified to a Softmax classifier through a full connection layer to obtain a classification result of the defect texts of the power transmission and transformation equipment. The method can solve the technical problems that the labor cost of the existing electric power field defect text classification is high, the classification result is easily influenced by the experience of classification technicians, and the traditional text classification method is not suitable for the electric power field.

Description

Deep learning-based method and system for classifying defect texts of power transmission and transformation equipment
Technical Field
The invention belongs to the technical field of natural language processing, and particularly relates to a deep learning-based method and system for classifying a defect text of power transmission and transformation equipment.
Background
With the continuous development of the smart power grid, a large amount of defect text data can be generated in the daily operation and maintenance process of the power grid; and the analysis of the defect text data is always the original basis for processing and analyzing the defect fault of the power grid equipment. At present, the defect text analysis of the power grid power transmission and transformation equipment is mainly completed manually, the cost is high, the efficiency is low, and the classification result is prone to being influenced by manual experience difference, so that the classification result has deviation. The development of artificial intelligence and natural language processing technology provides possibility for text mining of defects of electric power equipment. The existing text classification technology has naive Bayes, a support vector machine, a decision tree and the like, but a traditional text classifier based on a machine learning related algorithm is difficult to dig out deep features of a text and is not beneficial to further analysis, research and application of text data, meanwhile, the text in the electric power field contains a large amount of professional terms and special symbols, the specialization is strong, a general classification model in deep learning is difficult to obtain direct migration and application, and the existing electric power text mining is still in a starting stage.
Disclosure of Invention
In view of the above, an object of the present invention is to provide a method for classifying a defect text of a power transmission and transformation device based on deep learning, which is applicable to classification of defect texts in the power field.
In order to achieve the purpose, the technical scheme of the invention is as follows: a deep learning-based power transmission and transformation equipment defect text classification method comprises the following steps:
s1: preprocessing the acquired power transmission and transformation equipment defect text, and then performing word embedding on the preprocessed power transmission and transformation equipment defect text to obtain a first word vector with electric power semantic features;
s2: acquiring forward and backward characteristic information of a defect text of the power transmission and transformation equipment through a bidirectional long-time and short-time memory network, and outputting a hidden layer state vector;
s3: carrying out weighted transformation on the hidden layer state vector by using a self-attention mechanism to obtain deep semantic features and obtain a final sentence vector to be classified;
s4: and outputting the sentence vectors to be classified to a Softmax classifier through a full connection layer to obtain a classification result of the defect texts of the power transmission and transformation equipment.
Further, the preprocessing comprises word segmentation, stop word removal and unified expression processing on the power transmission and transformation equipment defect text.
Further, in step S1, the step of performing word embedding on the preprocessed power transmission and transformation equipment defect text to obtain a first word vector with electric semantic features specifically includes:
reading the preprocessed power transmission and transformation equipment defect text, and counting word frequency information;
constructing a dictionary, and initializing a Huffman tree and a random initialization word vector;
training a model by using a row unit to obtain an input sample in a current row;
accumulating the value of each dimension in the context word vector and averaging to obtain a projection layer vector;
traversing each intermediate node from the current word to the root node;
calculating the corresponding gradient g of the intermediate node, refreshing the error vector from the projection layer to the intermediate node, refreshing the vector of the intermediate node, and refreshing the vector of the context word.
Further, the step S2 specifically includes the following steps:
defining a forward LSTM structure and a backward LSTM structure, splicing results output by a network by adopting a dynamic RNN unit, inputting the results into a next layer of bidirectional long-time memory network, and dividing the results output by the last layer of Bi-LSTM into forward and backward outputs by a split method;
the outputs in the forward and backward directions are added to obtain the final hidden layer state.
Further, each time state in the LSTM structure is updated by the following method:
Figure 100002_DEST_PATH_IMAGE001
wherein,
Figure 756420DEST_PATH_IMAGE002
setting hyperbolic tangent function tanh as excitation functions of an LSTM state and an LSTM internal state, setting b as a bias constant, and respectively representing an input gate, a forgetting gate and an output gate by i, f and o in subscripts; g is a control gate unit updated along with time steps, namely a feedforward neural network taking a sigmiod function as an excitation function,
Figure 876823DEST_PATH_IMAGE003
for the current state of time t to be,
Figure 125402DEST_PATH_IMAGE004
the state of the previous time is the state of the previous time,
Figure 979000DEST_PATH_IMAGE005
is an input for the current time of day,
Figure 262214DEST_PATH_IMAGE006
is a weight value of the weight value,
Figure 553518DEST_PATH_IMAGE007
in order to enter the gate weight value,
Figure 289393DEST_PATH_IMAGE008
in order to output the weight value of the gate,
Figure 206402DEST_PATH_IMAGE009
in order to forget the weight value of the door,
Figure 609702DEST_PATH_IMAGE010
for the purpose of abstracting the information at the current moment,
Figure 71907DEST_PATH_IMAGE011
is the abstracted information of the previous time step,
Figure 295078DEST_PATH_IMAGE012
are weight coefficients.
Further, the deep semantic features in step S3 are obtained by:
Figure 750199DEST_PATH_IMAGE013
wherein Q = WqH,K= WkH,V=WvH,Wq、Wk、WvAll are initialized weight matrix, H is hidden layer state matrix, Self-attention value, Q is square value, K is key value, V is value,
Figure 742426DEST_PATH_IMAGE014
is the word vector dimension and T is the matrix transpose.
The invention also aims to provide a power transmission and transformation equipment defect text classification system based on deep learning, which can be applied to defect text classification in the electric power field.
In order to achieve the purpose, the technical scheme of the invention is as follows: a deep learning-based electric transmission and transformation equipment defect text classification system comprises:
the text processing module is used for preprocessing the acquired electric transmission and transformation equipment defect text and then performing word embedding on the preprocessed electric transmission and transformation equipment defect text to obtain a first word vector with electric power semantic features;
the semantic feature extraction module is connected with the text processing module and used for acquiring forward and backward feature information of the defective text of the power transmission and transformation equipment through a bidirectional long-time and short-time memory network, outputting a hidden layer state vector, performing weighted transformation on the hidden layer state vector by using a self-attention mechanism, acquiring deep semantic features and obtaining a final sentence vector to be classified;
and the text classification module is used for inputting the received sentence vector to be classified into the full connection layer and outputting the sentence vector to be classified into the Softmax classifier to obtain a classification result of the defect text of the power transmission and transformation equipment.
Further, the text processing module is used for preprocessing the electric transmission and transformation equipment defect text and comprises the steps of performing word segmentation, stop word removal and unified expression processing on the electric transmission and transformation equipment defect text.
Further, the deep semantic features are obtained by:
Figure 641112DEST_PATH_IMAGE015
wherein Q = WqH,K= WkH,V=WvH,Wq、Wk、WvAll are initialized weight matrix, H is hidden layer state matrix, Self-attention value, Q is square value, K is key value, V is value,
Figure 351579DEST_PATH_IMAGE016
is the word vector dimension and T is the matrix transpose.
Furthermore, the semantic feature extraction module is further configured to define a forward LSTM structure and a backward LSTM structure, splice results output by the network by using a dynamic RNN unit, input the results into a next layer of bidirectional long-and-short-term memory network, segment results output by the last layer of Bi-LSTM into forward and backward outputs by a split method, and add the forward and backward outputs to obtain a final hidden layer state.
Compared with the prior art, the invention has the following advantages:
1. the labor cost of actual production of the power grid is reduced, the fault defect text classification result is prevented from being influenced by experience of different personnel, the fault text classification efficiency of the power transmission and transformation equipment is greatly improved, and the fault text classification accuracy of the power transmission and transformation equipment is improved.
2. In the actual production of the power industry, the automatic classification of the defect texts of the power transmission and transformation equipment can provide objective and efficient fault division reference, and meanwhile, the comprehensive state estimation can be performed on the health state of the power transmission and transformation equipment by combining with structured data;
3. the classification of the defect texts of the power transmission and transformation equipment is part of research content of electric power text mining, and lays a foundation for mining and analyzing the defect text data of the electric power equipment and other electric power text data.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below. It is obvious that the drawings in the following description are some embodiments of the invention, and that for a person skilled in the art, other drawings can be derived from them without inventive exercise.
FIG. 1 is a structural diagram of a deep learning-based electric transmission and transformation equipment defect text classification system according to the present invention;
FIG. 2 is a flowchart of a deep learning-based method for classifying defect texts of electric transmission and transformation equipment according to the present invention;
FIG. 3 is a diagram illustrating a 100-dimensional word vector after training according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention. It is to be understood that the embodiments described are only a few embodiments of the present invention, and not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The examples are given for the purpose of better illustration of the invention, but the invention is not limited to the examples. Therefore, those skilled in the art should make insubstantial modifications and adaptations to the embodiments of the present invention in light of the above teachings and remain within the scope of the invention.
Example 1
Referring to fig. 1, a diagram of a text classification system for defects of power transmission and transformation equipment based on deep learning according to the present invention is shown, and specifically, the system includes:
the text processing module 1 is used for preprocessing the acquired electric transmission and transformation equipment defect text and then performing word embedding on the preprocessed electric transmission and transformation equipment defect text to obtain a first word vector with electric power semantic features;
in this embodiment, the preprocessing the electric transmission and transformation equipment defect text by the text processing module includes performing word segmentation, stop word removal and unified phrase processing on the electric transmission and transformation equipment defect text.
The semantic feature extraction module 2 is connected with the text processing module 1 and used for acquiring forward and backward feature information of the defective text of the power transmission and transformation equipment through a bidirectional long-time and short-time memory network, outputting a hidden layer state vector, performing weighted transformation on the hidden layer state vector by using a self-attention mechanism, acquiring deep semantic features and obtaining a final sentence vector to be classified;
in the semantic feature extraction module 2, a forward LSTM structure and a backward LSTM structure are defined, a dynamic RNN unit is adopted to splice the results output by the network, then the results are input into a next layer of bidirectional long-time memory network, the results output by the last layer of Bi-LSTM are divided into forward and backward outputs by a split method, and the forward and backward outputs are added to obtain a final hidden layer state.
Preferably, the deep semantic features are obtained by:
Figure DEST_PATH_IMAGE017
wherein Q = WqH,K= WkH,V=WvH,Wq、Wk、WvAll are initialized weight matrix, H is hidden layer state matrix, Self-attention value, Q is square value, K is key value, V is value,
Figure 813653DEST_PATH_IMAGE018
is the word vector dimension and T is the matrix transpose.
And the text classification module 3 is connected with the semantic feature extraction module 2 and is used for inputting the received sentence vectors to be classified into the full connection layer and outputting the sentence vectors to the Softmax classifier to obtain a classification result of the defect texts of the power transmission and transformation equipment.
Example 2
Based on the system of embodiment 1, the present embodiment provides a deep learning-based method for classifying a defect text of an electric transmission and transformation device, and referring to fig. 2, the method includes the following steps:
s1: preprocessing the acquired power transmission and transformation equipment defect text, and then performing word embedding on the preprocessed power transmission and transformation equipment defect text to obtain a first word vector with electric power semantic features;
in the step, the preprocessing of the electric transmission and transformation equipment defect text comprises the steps of performing word segmentation, stop word removal and unified phrase processing on the electric transmission and transformation equipment defect text;
further, the step of performing word embedding on the preprocessed electric transmission and transformation equipment defect text to obtain a first word vector with electric power semantic features specifically includes:
reading the preprocessed power transmission and transformation equipment defect text, and counting word frequency information;
constructing a dictionary, and initializing a Huffman tree and a random initialization word vector, wherein the dimension of the word vector is 100;
training a model by using a row unit to obtain an input sample in a current row;
accumulating the value of each dimension in the context word vector and averaging to obtain a projection layer vector;
traversing each intermediate node from the current word to the root node;
calculating the corresponding gradient g of the intermediate node, refreshing the error vector from the projection layer to the intermediate node, refreshing the vector of the intermediate node, and refreshing the vector of the context word.
In a specific embodiment, the defect record data of the power grid equipment of 2018 and 2019 of a certain power saving company is selected, and the defect record of the power transmission and transformation equipment is selected as the original data because the power transmission and transformation equipment is the equipment type with the largest defect record number, and the other equipment types are fewer and do not have the condition of forming a deep learning data set. Preprocessing the defect text of the power transmission and transformation equipment, including word segmentation, and removing stop words and unified expressions; adopting a 5-fold verification method, randomly arranging defect texts of all power transmission and transformation equipment, then averagely dividing the defect texts into 5 parts, selecting 4 parts as a training set in turn, and verifying 1 part as a test set, wherein one part of the test data set is assumed to be that the oil level of a main transformer body oil conservator of a No. 1 main transformer of a 35kv kali-slope transformer substation is too low, and the oil seal of a main transformer tap changer breather of a 35kv hot water power plant is too much; after the data set is preprocessed, the data set is respectively ' 35kv ', ' main transformer ', ' body ', ' oil storage cabinet ', ' oil level ', ' too low ', ' 35kv ', ' main transformer ', ' tap switch ', ' respirator ', ' oil seal ' and too much '.
For example, in one embodiment, assuming the word segmentation result of one of the test data sets is "insulator", the 100-dimensional word vector after training is shown in fig. 3;
s2: acquiring forward and backward characteristic information of a defect text of the power transmission and transformation equipment through a bidirectional long-time and short-time memory network, and outputting a hidden layer state vector;
in the embodiment, a forward long-short time memory network (LSTM) structure and a backward long-short time memory network (LSTM) structure are defined, the results (fw and bw, fw is a forward LSTM output result, and bw is a backward LSTM output result) output by the network are spliced by using a dynamic RNN unit, then the results are input to the next layer of bidirectional long-short time memory network, and the result output by the last layer of Bi-LSTM is divided into forward and backward outputs by a split method (split function) in python;
the outputs in the forward and backward directions are added to obtain the final hidden layer state.
Further, each time (step) state in the LSTM structure is updated by the following method:
Figure DEST_PATH_IMAGE019
wherein,
Figure 660386DEST_PATH_IMAGE020
setting hyperbolic tangent function tanh as excitation functions of a system state and an internal state, setting b as a bias constant, and respectively representing an input gate, a forgetting gate and an output gate by i, f and o in subscripts; g is a control gate unit updated along with time steps, namely a feedforward neural network taking a sigmiod function as an excitation function,
Figure 464394DEST_PATH_IMAGE021
in the state of the present time t,
Figure 911425DEST_PATH_IMAGE022
the layer state is hidden for the previous time,
Figure 459081DEST_PATH_IMAGE023
is an input for the current time of day,
Figure 425900DEST_PATH_IMAGE024
is a weight value of the weight value,
Figure 666388DEST_PATH_IMAGE025
in order to enter the gate weight value,
Figure 335136DEST_PATH_IMAGE026
in order to output the weight value of the gate,
Figure 952062DEST_PATH_IMAGE009
in order to forget the weight value of the door,
Figure DEST_PATH_IMAGE027
for the purpose of abstracting the information at the current moment,
Figure 242229DEST_PATH_IMAGE028
is the abstracted information of the previous time step,
Figure DEST_PATH_IMAGE029
are weight coefficients.
S3: carrying out weighted transformation on the hidden layer state vector by using a self-attention mechanism to obtain deep semantic features and obtain a final sentence vector to be classified;
in this step, the hidden layer state is introduced into a self-attention mechanism; and calculating attention weights assigned to the various inputs, and applying the weights obtained by calculation to the feature vector to obtain a final output feature vector. The specific calculation formula is as follows:
Figure 371728DEST_PATH_IMAGE030
wherein Q = WqH,K= WkH,V=WvH,Wq、Wk、WvAll are initialized weight matrix, H is hidden layer state matrix, Self-attention value, Q is square value, K is key value, V is value,
Figure DEST_PATH_IMAGE031
is the word vector dimension and T is the matrix transpose.
S4: and (4) outputting the sentence vectors to be classified to a Softmax classifier through a full connection layer to obtain a classification result of the defect texts of the power transmission and transformation equipment.
In this embodiment, the sentence vectors to be classified are output to the softmax classifier through the full connection layer, so as to obtain the probability of the corresponding class, and the class with the highest probability is selected as the final class.
Preferably, in the present embodiment, the construction process of the softmax classifier is as follows: the method comprises the steps of obtaining a defect text of the power transmission and transformation equipment with a known defect type, preprocessing the data through a defect text preprocessing unit, obtaining word vectors with text features through a defect text feature representation unit, obtaining semantic feature sentence vectors to be classified through a defect text feature extraction unit, obtaining training data, training model parameters of a softmax classifier by adopting a random gradient descent method, finishing training when a loss function is minimized, and obtaining the softmax classifier.
While the present invention has been described with reference to the embodiments shown in the drawings, the present invention is not limited to the embodiments, which are illustrative and not restrictive, and it will be apparent to those skilled in the art that various changes and modifications can be made therein without departing from the spirit and scope of the invention as defined in the appended claims.

Claims (10)

1. A deep learning-based power transmission and transformation equipment defect text classification method is characterized by comprising the following steps:
s1: preprocessing the acquired power transmission and transformation equipment defect text, and then performing word embedding on the preprocessed power transmission and transformation equipment defect text to obtain a first word vector with electric power semantic features;
s2: acquiring forward and backward characteristic information of a defect text of the power transmission and transformation equipment through a bidirectional long-time and short-time memory network, and outputting a hidden layer state vector;
s3: carrying out weighted transformation on the hidden layer state vector by using a self-attention mechanism to obtain deep semantic features and obtain a final sentence vector to be classified;
s4: and outputting the sentence vectors to be classified to a Softmax classifier through a full connection layer to obtain a classification result of the defect texts of the power transmission and transformation equipment.
2. The method of claim 1, wherein the pre-processing comprises performing word segmentation, stop word removal, and unionization wording on the electric transmission and transformation equipment defect text.
3. The method according to claim 1, wherein the step of performing word embedding on the preprocessed transmission and transformation equipment defect text in step S1 to obtain the first word vector with the electric power semantic features specifically comprises:
reading the preprocessed power transmission and transformation equipment defect text, and counting word frequency information;
constructing a dictionary, and initializing a Huffman tree and a random initialization word vector;
training a model by using a row unit to obtain an input sample in a current row;
accumulating the value of each dimension in the context word vector and averaging to obtain a projection layer vector;
traversing each intermediate node from the current word to the root node;
calculating the corresponding gradient g of the intermediate node, refreshing the error vector from the projection layer to the intermediate node, refreshing the vector of the intermediate node, and refreshing the vector of the context word.
4. The method according to claim 1, wherein the step S2 specifically comprises the steps of:
defining a forward LSTM structure and a backward LSTM structure, splicing results output by a network by adopting a dynamic RNN unit, inputting the results into a next layer of bidirectional long-time memory network, and dividing the results output by the last layer of Bi-LSTM into forward and backward outputs by a split method;
the outputs in the forward and backward directions are added to obtain the final hidden layer state.
5. The method of claim 4, wherein each time state in the LSTM structure is updated by:
Figure DEST_PATH_IMAGE001
wherein,
Figure 339188DEST_PATH_IMAGE002
setting hyperbolic tangent function tanh as excitation functions of an LSTM state and an LSTM internal state, setting b as a bias constant, and respectively representing an input gate, a forgetting gate and an output gate by i, f and o in subscripts; g is a control gate unit updated with time steps,
Figure 477915DEST_PATH_IMAGE003
in the state of the current time t,
Figure 419326DEST_PATH_IMAGE004
the state of the previous time is the state of the previous time,
Figure 1617DEST_PATH_IMAGE005
is an input for the current time of day,
Figure 395689DEST_PATH_IMAGE006
is a weight value of the weight value,
Figure 66668DEST_PATH_IMAGE007
in order to enter the gate weight value,
Figure 597006DEST_PATH_IMAGE008
in order to output the weight value of the gate,
Figure 350198DEST_PATH_IMAGE009
in order to forget the weight value of the door,
Figure 497146DEST_PATH_IMAGE010
for the purpose of abstracting the information at the current moment,
Figure 243254DEST_PATH_IMAGE011
is the abstracted information of the previous time step,
Figure 893678DEST_PATH_IMAGE012
are weight coefficients.
6. The method according to claim 1, wherein the deep semantic features in step S3 are obtained by:
Figure 552193DEST_PATH_IMAGE013
wherein Q = WqH,K= WkH,V=WvH,Wq、Wk、WvAll are initialized weight matrix, H is hidden layer state matrix, Self-attention value, Q is square value, K is key value, V is value,
Figure 920857DEST_PATH_IMAGE014
is the word vector dimension and T is the matrix transpose.
7. The utility model provides a power transmission and transformation equipment defect text classification system based on deep learning which characterized in that includes:
the text processing module is used for preprocessing the acquired electric transmission and transformation equipment defect text and then performing word embedding on the preprocessed electric transmission and transformation equipment defect text to obtain a first word vector with electric power semantic features;
the semantic feature extraction module is connected with the text processing module and used for acquiring forward and backward feature information of the defective text of the power transmission and transformation equipment through a bidirectional long-time and short-time memory network, outputting a hidden layer state vector, performing weighted transformation on the hidden layer state vector by using a self-attention mechanism, acquiring deep semantic features and obtaining a final sentence vector to be classified;
and the text classification module is used for inputting the received sentence vector to be classified into the full connection layer and outputting the sentence vector to be classified into the Softmax classifier to obtain a classification result of the defect text of the power transmission and transformation equipment.
8. The system of claim 7, wherein the text processing module pre-processes the electrical transmission and transformation equipment defect text including word segmentation, stop word removal, and unionization.
9. The system of claim 7, wherein the deep semantic features are obtained by:
Figure 205077DEST_PATH_IMAGE015
wherein Q = WqH,K= WkH,V=WvH,Wq、Wk、WvAll are initialized weight matrix, H is hidden layer state matrix, Self-attention value, Q is square value, K is key value, V is value,
Figure 975587DEST_PATH_IMAGE016
is the word vector dimension and T is the matrix transpose.
10. The system of claim 7, wherein the semantic feature extraction module is further configured to define a forward LSTM structure and a backward LSTM structure, splice the results output by the network using a dynamic RNN unit, input the results into a next layer of bidirectional long-and-short-term memory network, divide the results output by the last layer of Bi-LSTM into forward and backward outputs by a split method, and add the forward and backward outputs to obtain a final hidden layer state.
CN202110279537.6A 2021-03-16 2021-03-16 Deep learning-based method and system for classifying defect texts of power transmission and transformation equipment Pending CN112699244A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110279537.6A CN112699244A (en) 2021-03-16 2021-03-16 Deep learning-based method and system for classifying defect texts of power transmission and transformation equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110279537.6A CN112699244A (en) 2021-03-16 2021-03-16 Deep learning-based method and system for classifying defect texts of power transmission and transformation equipment

Publications (1)

Publication Number Publication Date
CN112699244A true CN112699244A (en) 2021-04-23

Family

ID=75515268

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110279537.6A Pending CN112699244A (en) 2021-03-16 2021-03-16 Deep learning-based method and system for classifying defect texts of power transmission and transformation equipment

Country Status (1)

Country Link
CN (1) CN112699244A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113935338A (en) * 2021-10-13 2022-01-14 北京有竹居网络技术有限公司 Method, apparatus, device and medium for language translation based on layer prediction
CN114328797A (en) * 2021-11-09 2022-04-12 腾讯科技(深圳)有限公司 Content search method, content search device, electronic device, storage medium, and program product
CN115563528A (en) * 2022-11-25 2023-01-03 四川飞宇电力股份有限公司 Transformer maintenance equipment control method, electronic equipment and medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108363753A (en) * 2018-01-30 2018-08-03 南京邮电大学 Comment text sentiment classification model is trained and sensibility classification method, device and equipment
CN110135551A (en) * 2019-05-15 2019-08-16 西南交通大学 A kind of robot chat method of word-based vector sum Recognition with Recurrent Neural Network
CN110457562A (en) * 2019-08-15 2019-11-15 中国农业大学 A kind of food safety affair classification method and device based on neural network model
US20200380213A1 (en) * 2018-02-09 2020-12-03 Salesforce.Com, Inc. Multitask Learning As Question Answering

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108363753A (en) * 2018-01-30 2018-08-03 南京邮电大学 Comment text sentiment classification model is trained and sensibility classification method, device and equipment
US20200380213A1 (en) * 2018-02-09 2020-12-03 Salesforce.Com, Inc. Multitask Learning As Question Answering
CN110135551A (en) * 2019-05-15 2019-08-16 西南交通大学 A kind of robot chat method of word-based vector sum Recognition with Recurrent Neural Network
CN110457562A (en) * 2019-08-15 2019-11-15 中国农业大学 A kind of food safety affair classification method and device based on neural network model

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
冯斌 等: "基于BiLSTM-Attention神经网络的电力设备缺陷文本挖掘", 《中国电机工程学报》 *
廖伟涵: "基于机器学习与自然语言处理的变压器故障诊断与状态评价研究", 《中国优秀硕士学位论文全文数据库 工程科技Ⅱ辑》 *
张振: "基于车辆通行大数据的高速公路路况预测的研究与应用", 《中国优秀硕士学位论文全文数据库 工程科技Ⅱ辑》 *
杨鹏: "基于领域词典与机器学习的中文评论情感分析", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *
类库: ""双向LSTM模型的tensorflow实现"", 《HTTPS://WWW.136.LA/SHIDA/SHOW-97712.HTML》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113935338A (en) * 2021-10-13 2022-01-14 北京有竹居网络技术有限公司 Method, apparatus, device and medium for language translation based on layer prediction
CN114328797A (en) * 2021-11-09 2022-04-12 腾讯科技(深圳)有限公司 Content search method, content search device, electronic device, storage medium, and program product
CN114328797B (en) * 2021-11-09 2024-03-19 腾讯科技(深圳)有限公司 Content search method, device, electronic apparatus, storage medium, and program product
CN115563528A (en) * 2022-11-25 2023-01-03 四川飞宇电力股份有限公司 Transformer maintenance equipment control method, electronic equipment and medium

Similar Documents

Publication Publication Date Title
CN112699244A (en) Deep learning-based method and system for classifying defect texts of power transmission and transformation equipment
WO2018218705A1 (en) Method for recognizing network text named entity based on neural network probability disambiguation
CN112905804B (en) Dynamic updating method and device for power grid dispatching knowledge graph
CN109933792B (en) Viewpoint type problem reading and understanding method based on multilayer bidirectional LSTM and verification model
CN111767397A (en) Electric power system secondary equipment fault short text data classification method
CN112395417A (en) Network public opinion evolution simulation method and system based on deep learning
CN114661905B (en) BERT-based power grid fault diagnosis method
CN112837175B (en) Frequent power failure work order information extraction method and system based on information extraction technology
CN114926150A (en) Digital intelligent auditing method and device for transformer technology conformance assessment
CN112419096A (en) Automatic user power demand worksheet transferring method based on NLP information extraction and few-sample self-learning
CN111274814A (en) Novel semi-supervised text entity information extraction method
CN116842337A (en) Transformer fault diagnosis method based on LightGBM (gallium nitride based) optimal characteristics and COA-CNN (chip on board) model
CN113723083A (en) Weighted negative supervision text emotion analysis method based on BERT model
CN112559741B (en) Nuclear power equipment defect record text classification method, system, medium and electronic equipment
CN115292490A (en) Analysis algorithm for policy interpretation semantics
CN113051886B (en) Test question duplicate checking method, device, storage medium and equipment
CN117932295A (en) Multi-source data fusion power grid monitoring operation characteristic information extraction method and system
Lin et al. The study on classification and prediction for data mining
CN117033961A (en) Multi-mode image-text classification method for context awareness
CN116029295A (en) Electric power text entity extraction method, defect positioning method and fault diagnosis method
CN114936667A (en) Power grid day-ahead scheduling plan optimization decision method based on knowledge graph
Qu et al. Improvement of power equipment defect text quality based on improved BI-LSTM
CN114153975A (en) Text classification method and device, electronic equipment and storage medium
CN114818651A (en) Text similarity determination method and device, storage medium and electronic device
Tan et al. Application of self-supervised learning in non-intrusive load monitoring

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20210423

RJ01 Rejection of invention patent application after publication