CN115545361B - Method, system and medium for predicting climate environment of power grid transmission line - Google Patents

Method, system and medium for predicting climate environment of power grid transmission line Download PDF

Info

Publication number
CN115545361B
CN115545361B CN202211534275.4A CN202211534275A CN115545361B CN 115545361 B CN115545361 B CN 115545361B CN 202211534275 A CN202211534275 A CN 202211534275A CN 115545361 B CN115545361 B CN 115545361B
Authority
CN
China
Prior art keywords
layer
sequence
decoding
attention
prediction
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202211534275.4A
Other languages
Chinese (zh)
Other versions
CN115545361A (en
Inventor
李磊
周正
胡钰林
廖荣涛
王逸兮
叶宇轩
王晟玮
胡欢君
张剑
宁昊
董亮
刘芬
郭岳
罗弦
张岱
李想
陈家璘
冯浩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan University WHU
Information and Telecommunication Branch of State Grid Hubei Electric Power Co Ltd
Original Assignee
Wuhan University WHU
Information and Telecommunication Branch of State Grid Hubei Electric Power Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan University WHU, Information and Telecommunication Branch of State Grid Hubei Electric Power Co Ltd filed Critical Wuhan University WHU
Priority to CN202211534275.4A priority Critical patent/CN115545361B/en
Publication of CN115545361A publication Critical patent/CN115545361A/en
Application granted granted Critical
Publication of CN115545361B publication Critical patent/CN115545361B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/049Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/06Energy or water supply
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/10Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation

Landscapes

  • Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Economics (AREA)
  • General Physics & Mathematics (AREA)
  • Strategic Management (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Resources & Organizations (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • General Business, Economics & Management (AREA)
  • Marketing (AREA)
  • Tourism & Hospitality (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Public Health (AREA)
  • Water Supply & Treatment (AREA)
  • Primary Health Care (AREA)
  • Development Economics (AREA)
  • Game Theory and Decision Science (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The application relates to a method, a system and a medium for predicting climate environment of a power grid transmission line, wherein the method comprises the steps of obtaining historical climate environment data and constructing an input sample; constructing a transducer prediction model of a sparse attention mechanism, and predicting an input sample to obtain an original prediction result; comparing the original prediction result with real data to obtain an error sequence, and inputting the error sequence into an ARIMA model for error prediction to obtain an error correction sequence; and adding the original prediction result and the error correction sequence to obtain a final prediction result. The method and the device effectively improve the stability and accuracy of overall prediction, effectively reduce the operand of self-attention operation in the traditional transducer model, accelerate the model training and prediction speed, and reduce the hardware requirements for deploying the model.

Description

Method, system and medium for predicting climate environment of power grid transmission line
Technical Field
The application relates to the field of power systems, in particular to a method, a system and a medium for predicting a climate environment of a power grid transmission line based on ARIMA correction transducer.
Background
In the power system, the climate environment information of the power grid transmission line has important reference value for planning, scheduling and maintaining the power network, and the climate environment information of the power grid transmission line can be predicted in advance to assist scheduling decision, so that the construction of the intelligent power grid is assisted. Therefore, the climate environment prediction method of the power grid transmission line has important significance.
Existing predictive models mainly include traditional linear models and emerging recurrent neural network models. The traditional linear model performs poorly for nonlinear parts in the environmental data, and the predictive performance depends on the choice of parameters. However, a series of models based on the recurrent neural network have difficulty in capturing long-term dependency relationships between historical data, and the required calculation amount is large. The use of hybrid models has received a great deal of attention due to the limitations of a single predictive model. Prediction in combination with different models is considered as an effective way to improve the prediction results.
Disclosure of Invention
The embodiment of the application aims to provide a method, a system and a medium for predicting the climate environment of a power grid transmission line, which effectively reduce the operation amount of self-attention operation in a traditional transducer model, accelerate the model training and prediction speed and reduce the hardware requirement for deploying the model.
In order to achieve the above purpose, the present application provides the following technical solutions:
in a first aspect, an embodiment of the present application provides a method for predicting a climate environment of a transmission line of a power grid, including the following specific steps:
acquiring historical climate environment data to construct an input sample;
constructing a transducer prediction model of a sparse attention mechanism, and predicting an input sample to obtain an original prediction result;
comparing the original prediction result with real data to obtain an error sequence, inputting an ARIMA model for error prediction to obtain an error correction sequence, wherein the error correction sequence gives possible deviation between the original prediction result of the transducer and the actual data;
and adding the original prediction result and the error correction sequence, and improving the defect of lack of linear characteristics of a transformation former model through the error correction sequence of the ARIMA model to obtain a final predicted future environmental data sequence with higher accuracy and stability.
The obtained historical climate environment data construction input sample is specifically that the historical climate environment data comprises a time sequence composed of historical values of temperature, humidity and wind speed information around a transmission line, the historical values are normalized and stored as vectors
Figure 27862DEST_PATH_IMAGE001
As a model input.
The transducer prediction model of the sparse attention mechanism comprises an encoder and a decoder, wherein the encoder performs feature extraction on input data; the decoder performs multi-head sparse self-attention operation by utilizing the features extracted by the encoder, and finally realizes the climate environment prediction of the transmission line, wherein the input of the encoding layer is as follows
Figure 401075DEST_PATH_IMAGE001
The input of the decoding layer is +.>
Figure 530705DEST_PATH_IMAGE001
Output of coding layer ∈>
Figure 423706DEST_PATH_IMAGE002
The encoder consists of several identical encoding layers, the decoder consists of several identical decoding layers except that the first encoding layer is input as +.>
Figure 980589DEST_PATH_IMAGE001
All coding layers have the output of the last coding layer as its input, the output of the last coding layer as the output of the encoder>
Figure 106677DEST_PATH_IMAGE002
Except that the first decoding layer input is +.>
Figure 508839DEST_PATH_IMAGE001
And->
Figure 912139DEST_PATH_IMAGE002
All the decoding layers take the output of the last decoding layer as input, and the output of the last decoding layer is the predicted sequence of the transducer prediction model for future climate data.
The coding layer of the transform prediction model comprises a one-dimensional convolution layer and sparse self-attentionThe encoding layer performs characteristic extraction, specifically, an input vector
Figure 279404DEST_PATH_IMAGE001
After one-dimensional convolution layer and adding position coding, the +.>
Figure 502575DEST_PATH_IMAGE003
Three weight matrices are used respectively +.>
Figure 974007DEST_PATH_IMAGE004
,/>
Figure 356447DEST_PATH_IMAGE005
And->
Figure 255133DEST_PATH_IMAGE006
And->
Figure 44229DEST_PATH_IMAGE003
Matrix multiplication is performed to obtain a multi-channel query matrix respectively>
Figure 381669DEST_PATH_IMAGE007
Key matrix->
Figure 602304DEST_PATH_IMAGE008
And value matrix->
Figure 937470DEST_PATH_IMAGE009
In order to reduce the calculation amount of the model and prevent overfitting, the self-attention is obtained by calculating three matrixes through sparse self-attention, and the calculated self-attention is mapped into the original input dimension through a linear layer and then is combined with +.>
Figure 135233DEST_PATH_IMAGE003
Adding, and then obtaining the output +.A result of the adding is obtained through the feedforward layer>
Figure 338681DEST_PATH_IMAGE002
The sparse self-attention operation operates as follows: selecting a matrix
Figure 305500DEST_PATH_IMAGE007
Maximum value +.>
Figure 545989DEST_PATH_IMAGE010
The matrix after thinning is obtained by setting other elements to 0 by each element>
Figure 371994DEST_PATH_IMAGE011
Wherein->
Figure 723341DEST_PATH_IMAGE010
For a number of appropriate sizes set in advance, then matrix +.>
Figure 810245DEST_PATH_IMAGE011
,/>
Figure 346269DEST_PATH_IMAGE008
,/>
Figure 518624DEST_PATH_IMAGE009
The self-attention operation is carried out as input, and the formula is
Figure 673662DEST_PATH_IMAGE012
Wherein the method comprises the steps of
Figure 723395DEST_PATH_IMAGE013
Activating a function for softmax +.>
Figure 571266DEST_PATH_IMAGE014
For matrix->
Figure 89972DEST_PATH_IMAGE007
Dimension of->
Figure 48700DEST_PATH_IMAGE015
To calculate the self-attention.
The decoding layer of the transform prediction model comprises a decoding one-dimensional convolution layer, a decoding sparse self-attention layer, a decoding feedforward network layer and a decoding linear layer, and the decoding layer performs specific prediction operation to carry out the first input data
Figure 657667DEST_PATH_IMAGE001
After decoding one-dimensional convolution layer and one layer of decoding sparse self-attention layer, mapping into query matrix by decoding linear layer
Figure 801073DEST_PATH_IMAGE016
And thinning to obtain->
Figure 682441DEST_PATH_IMAGE017
Then use the second input data +.>
Figure 444861DEST_PATH_IMAGE002
Mapping into key matrix by decoding linear layer
Figure 469186DEST_PATH_IMAGE018
And value matrix->
Figure 658859DEST_PATH_IMAGE019
Attention is then calculated, given by
Figure 293103DEST_PATH_IMAGE020
Wherein the method comprises the steps of
Figure 452689DEST_PATH_IMAGE021
Is->
Figure 223199DEST_PATH_IMAGE016
Is used for obtaining a transducer prediction model for future environment after the calculated attention is decoded by a linear layer and a feedforward layerPredicted sequence of data.
In a second aspect, embodiments of the present application provide a system for predicting climate conditions of a power grid transmission line, comprising,
the input sample construction module is used for acquiring historical climate environment data and constructing an input sample;
the transducer prediction model construction module is used for constructing a transducer prediction model of a sparse attention mechanism and predicting an input sample to obtain an original prediction result;
the error correction sequence acquisition module is used for comparing the original prediction result with real data to obtain an error sequence, and inputting the error sequence into the ARIMA model for error prediction to obtain an error correction sequence;
and the prediction result acquisition module is used for adding the original prediction result and the error correction sequence to obtain a final prediction result.
In a third aspect, embodiments of the present application provide a computer readable storage medium storing program code which, when executed by a processor, implements the steps of a grid transmission line climate environment prediction method as described above.
Compared with the prior art, the invention has the beneficial effects that:
(1) According to the method, the prediction model based on the transducer is used for effectively extracting the time domain characteristics and the long-term dependence information of the input long-time sequence by utilizing the one-dimensional convolutional neural network and the sparse self-attention mechanism, so that the problem that the long-time sequence data are difficult to effectively process in the prior art is solved;
(2) According to the method, the prediction model based on the transducer and the ARIMA is combined, and the nonlinear feature extraction capacity and the linear feature extraction capacity of the transducer model are combined, so that the ARIMA is utilized to carry out error correction on the prediction model based on the transducer, and the stability and the accuracy of overall prediction are effectively improved;
(3) According to the method and the device, sparse self-attention operation is introduced, so that the operation amount of self-attention operation in a traditional transducer model is effectively reduced, the model training and predicting speed is accelerated, and the hardware requirement for deploying the model is reduced.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings that are needed in the embodiments of the present application will be briefly described below, it should be understood that the following drawings only illustrate some embodiments of the present application and should not be considered as limiting the scope, and other related drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a flow chart of a method according to an embodiment of the present application;
FIG. 2 is a schematic flow chart of a method implementation in an embodiment of the present application;
fig. 3 is a schematic diagram of an adaptive dynamic convolution AdaConv module according to an embodiment of the present application;
FIG. 4 is a system block diagram of an embodiment of the present application;
fig. 5 is a graph showing the comparison of the prediction effect in the embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be described below with reference to the drawings in the embodiments of the present application. It should be noted that: like reference numerals and letters denote like items in the following figures, and thus once an item is defined in one figure, no further definition or explanation thereof is necessary in the following figures.
The terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
Referring to fig. 1, fig. 1 provides a method for predicting climate environment of a transmission line of a power grid according to an embodiment of the present application, including the following specific steps:
acquiring historical climate environment data to construct an input sample;
constructing a transducer prediction model of a sparse attention mechanism, and predicting an input sample to obtain an original prediction result;
comparing the original prediction result with real data to obtain an error sequence, inputting an ARIMA model for error prediction to obtain an error correction sequence, wherein the error correction sequence gives possible deviation between the original prediction result of the transducer and the actual data;
and adding the original prediction result and the error correction sequence, and improving the defect of lack of linear characteristics of a transformation former model through the error correction sequence of the ARIMA model to obtain a final predicted future environmental data sequence with higher accuracy and stability.
The obtained historical climate environment data construction input sample is specifically that the historical climate environment data comprises a time sequence composed of historical values of temperature, humidity and wind speed information around a transmission line, the historical values are normalized and stored as vectors
Figure 318194DEST_PATH_IMAGE001
As a model input.
As shown in fig. 2 and fig. 3, the transform prediction model of the sparse attention mechanism comprises an encoder and a decoder, and the encoder performs feature extraction on input data; the decoder performs multi-head sparse self-attention operation by utilizing the features extracted by the encoder, and finally realizes the climate environment prediction of the transmission line, wherein the input of the encoding layer is as follows
Figure 49520DEST_PATH_IMAGE001
The input of the decoding layer is +.>
Figure 153743DEST_PATH_IMAGE001
Output of coding layer ∈>
Figure 778759DEST_PATH_IMAGE002
The encoder consists of several identical coding layer groupsThe decoder consists of several identical decoding layers except that the first coding layer is input +.>
Figure 434868DEST_PATH_IMAGE001
All coding layers have the output of the last coding layer as its input, the output of the last coding layer as the output of the encoder>
Figure 512546DEST_PATH_IMAGE002
Except that the first decoding layer input is +.>
Figure 154880DEST_PATH_IMAGE001
And->
Figure 8304DEST_PATH_IMAGE002
All the decoding layers take the output of the last decoding layer as input, and the output of the last decoding layer is the predicted sequence of the transducer prediction model for future climate data.
The coding layer of the transform prediction model comprises a one-dimensional convolution layer, a sparse self-attention layer and a feedforward network layer, and the characteristic extraction operation of the coding layer is specifically that an input vector is input
Figure 710681DEST_PATH_IMAGE001
After one-dimensional convolution layer and adding position coding, the +.>
Figure 541233DEST_PATH_IMAGE003
Three weight matrices are used respectively +.>
Figure 315154DEST_PATH_IMAGE004
,/>
Figure 55708DEST_PATH_IMAGE005
And->
Figure 663407DEST_PATH_IMAGE006
And->
Figure 715677DEST_PATH_IMAGE003
Matrix multiplication is performed to obtain a multi-channel query matrix respectively>
Figure 824447DEST_PATH_IMAGE007
Key matrix->
Figure 544141DEST_PATH_IMAGE008
And value matrix->
Figure 696643DEST_PATH_IMAGE009
In order to reduce the calculation amount of the model and prevent overfitting, the self-attention is obtained by calculating three matrixes through sparse self-attention, and the calculated self-attention is mapped into the original input dimension through a linear layer and then is combined with +.>
Figure 236209DEST_PATH_IMAGE003
Adding, and then obtaining the output +.A result of the adding is obtained through the feedforward layer>
Figure 758457DEST_PATH_IMAGE002
The sparse self-attention operation operates as follows: selecting a matrix
Figure 457292DEST_PATH_IMAGE007
Maximum value +.>
Figure 141214DEST_PATH_IMAGE010
The matrix after thinning is obtained by setting other elements to 0 by each element>
Figure 309021DEST_PATH_IMAGE011
Wherein->
Figure 369381DEST_PATH_IMAGE010
For a number of appropriate sizes set in advance, then matrix +.>
Figure 922722DEST_PATH_IMAGE011
,/>
Figure 43125DEST_PATH_IMAGE008
,/>
Figure 557283DEST_PATH_IMAGE009
The self-attention operation is carried out as input, and the formula is
Figure 806954DEST_PATH_IMAGE012
Wherein the method comprises the steps of
Figure 90167DEST_PATH_IMAGE013
Activating a function for softmax +.>
Figure 771685DEST_PATH_IMAGE014
For matrix->
Figure 507559DEST_PATH_IMAGE007
Dimension of->
Figure 440880DEST_PATH_IMAGE015
To calculate the self-attention.
The decoding layer of the transform prediction model comprises a decoding one-dimensional convolution layer, a decoding sparse self-attention layer, a decoding feedforward network layer and a decoding linear layer, and the decoding layer performs specific prediction operation to carry out the first input data
Figure 453967DEST_PATH_IMAGE001
After decoding one-dimensional convolution layer and one layer of decoding sparse self-attention layer, mapping into query matrix by decoding linear layer
Figure 181751DEST_PATH_IMAGE016
And thinning to obtain->
Figure 529556DEST_PATH_IMAGE017
Then use the second input data +.>
Figure 266568DEST_PATH_IMAGE002
Mapping into key matrix by decoding linear layer
Figure 258795DEST_PATH_IMAGE018
And value matrix->
Figure 531382DEST_PATH_IMAGE019
Attention is then calculated, given by
Figure 241849DEST_PATH_IMAGE020
/>
Wherein the method comprises the steps of
Figure 297399DEST_PATH_IMAGE021
Is->
Figure 3186DEST_PATH_IMAGE016
The calculated attention is decoded by the linear layer and the feedforward layer to obtain a predicted sequence of the transducer prediction model on future environment data.
The ARIMA model is established as follows:
drawing the historical data time sequence, observing whether the historical data time sequence is stable or not, and differentiating the historical data time sequence to obtain a stable time sequence if the historical data time sequence is a non-stable sequence;
for the obtained stationary time series, the autocorrelation coefficients ACF and the partial autocorrelation coefficients PACF are calculated, and the formula for calculating the autocorrelation coefficients ACF is as follows:
Figure 338353DEST_PATH_IMAGE022
where N is the length of the sequence,
Figure 50963DEST_PATH_IMAGE023
for the number of sequence lags, +.>
Figure 270723DEST_PATH_IMAGE024
For the sequence mean>
Figure 345864DEST_PATH_IMAGE025
For the sequence->
Figure 461718DEST_PATH_IMAGE026
A point. The partial autocorrelation coefficients PACF are typically calculated using the least squares method.
Determining appropriate model parameters from ACF and PACF
Figure 740253DEST_PATH_IMAGE027
. Wherein->
Figure 91600DEST_PATH_IMAGE028
For the hysteresis number of the time series data itself used in the predictive model,/->
Figure 286827DEST_PATH_IMAGE029
Differential times required for obtaining a smooth time sequence, < > are given>
Figure 698217DEST_PATH_IMAGE030
Is the hysteresis number of the prediction error employed in the prediction model.
As shown in fig. 4, an embodiment of the present application provides a climate environment prediction system for a power grid transmission line, comprising,
the input sample construction module 1 is used for acquiring historical climate environment data and constructing an input sample;
the transducer prediction model construction module 2 is used for constructing a transducer prediction model of a sparse attention mechanism, and predicting an input sample to obtain an original prediction result;
the error correction sequence acquisition module 3 is used for comparing the original prediction result with real data to obtain an error sequence, and inputting the error sequence into the ARIMA model for error prediction to obtain an error correction sequence;
the prediction result obtaining module 4 is configured to add the original prediction result to the error correction sequence to obtain a final prediction result.
As shown in fig. 5, by applying the method of the present application, the air temperature of the power transmission line is predicted, and according to the comparative analysis of the measured temperature and the predicted temperature, it can be known that the weather environment prediction method of the power grid transmission line of the present application has a small difference between the predicted temperature and the measured temperature, and the prediction accuracy is high.
Embodiments of the present application provide a computer readable storage medium storing program code which, when executed by a processor, implements the steps of a grid transmission line climate environment prediction method as described above.
It will be appreciated by those skilled in the art that embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
In one typical configuration, a computing device includes one or more processors (CPUs), input/output interfaces, network interfaces, and memory.
The memory may include volatile memory in a computer-readable medium, random Access Memory (RAM) and/or nonvolatile memory, etc., such as Read Only Memory (ROM) or flash RAM. Memory is an example of a computer-readable medium.
Computer readable media, including both non-transitory and non-transitory, removable and non-removable media, may implement information storage by any method or technology. The information may be computer readable instructions, data structures, modules of a program, or other data. Examples of storage media for a computer include, but are not limited to, phase change memory (PRAM), static Random Access Memory (SRAM), dynamic Random Access Memory (DRAM), other types of Random Access Memory (RAM), read Only Memory (ROM), electrically Erasable Programmable Read Only Memory (EEPROM), flash memory or other memory technology, compact disc read only memory (CD-ROM), digital Versatile Discs (DVD) or other optical storage, magnetic cassettes, magnetic tape magnetic disk storage or other magnetic storage devices, or any other non-transmission medium, which can be used to store information that can be accessed by a computing device. Computer-readable media, as defined herein, does not include transitory computer-readable media (transmission media), such as modulated data signals and carrier waves.
It should also be noted that the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article or apparatus that comprises an element.
The foregoing is merely exemplary embodiments of the present application and is not intended to limit the scope of the present application, and various modifications and variations may be suggested to one skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principles of the present application should be included in the protection scope of the present application.

Claims (6)

1. The method for predicting the climate environment of the power grid transmission line is characterized by comprising the following specific steps of:
acquiring historical climate environment data to construct an input sample;
constructing a transducer prediction model of a sparse attention mechanism, and predicting an input sample to obtain an original prediction result;
comparing the original prediction result with real data to obtain an error sequence, inputting an ARIMA model for error prediction to obtain an error correction sequence, wherein the error correction sequence gives possible deviation between the original prediction result of the transducer and the actual data;
adding the original prediction result and the error correction sequence, and improving the defect of lack of linear characteristics of a transformation former model through the error correction sequence of the ARIMA model to obtain a final predicted future environmental data sequence with higher accuracy and stability;
the obtained historical climate environment data construction input sample is specifically that the historical climate environment data comprises a time sequence composed of historical values of temperature, humidity and wind speed information around a transmission line, the historical values are normalized and stored as vectors
Figure QLYQS_1
As a model input;
the ARIMA model is established as follows:
drawing the historical data time sequence, observing whether the historical data time sequence is stable or not, and differentiating the historical data time sequence to obtain a stable time sequence if the historical data time sequence is a non-stable sequence;
for the obtained stationary time series, the autocorrelation coefficients ACF and the partial autocorrelation coefficients PACF are calculated, and the formula for calculating the autocorrelation coefficients ACF is as follows:
Figure QLYQS_2
where N is the length of the sequence,
Figure QLYQS_3
for the number of sequence lags, +.>
Figure QLYQS_4
For the sequence mean>
Figure QLYQS_5
For the sequence->
Figure QLYQS_6
The partial autocorrelation coefficients PACF are calculated using the least squares method,
determining appropriate model parameters from ACF and PACF
Figure QLYQS_7
Wherein->
Figure QLYQS_8
For the hysteresis number of the time series data itself used in the predictive model,/->
Figure QLYQS_9
Differential times required for obtaining a smooth time sequence, < > are given>
Figure QLYQS_10
Hysteresis number of prediction error adopted in the prediction model;
the transducer prediction model of the sparse attention mechanism comprises an encoder and a decoder, wherein the encoder performs feature extraction on input data; the decoder performs multi-head sparse self-attention operation by utilizing the features extracted by the encoder, and finally realizes the climate environment prediction of the transmission line, wherein the input of the encoding layer is as follows
Figure QLYQS_11
The input of the decoding layer is +.>
Figure QLYQS_12
Output of coding layer ∈>
Figure QLYQS_13
The encoder consists of several identical encoding layers, the decoder consists of several identical decoding layers except that the first encoding layer is input as +.>
Figure QLYQS_14
All coding layers have the output of the last coding layer as its input, the output of the last coding layer as the output of the encoder>
Figure QLYQS_15
Except that the first decoding layer input is +.>
Figure QLYQS_16
And->
Figure QLYQS_17
All the decoding layers take the output of the last decoding layer as input, and the output of the last decoding layer is the predicted sequence of the transducer prediction model for future climate data.
2. A method for predicting climate environment of power grid transmission line according to claim 1The method is characterized in that the coding layer of the transformation former prediction model comprises a one-dimensional convolution layer, a sparse self-attention layer and a feedforward network layer, and the characteristic extraction operation of the coding layer is that an input vector is specifically
Figure QLYQS_19
After one-dimensional convolution layer and adding position coding to obtain
Figure QLYQS_21
Three weight matrices are used respectively +.>
Figure QLYQS_24
,/>
Figure QLYQS_20
And->
Figure QLYQS_23
And->
Figure QLYQS_25
Matrix multiplication is performed to obtain multi-channel query matrix respectively
Figure QLYQS_27
Key matrix->
Figure QLYQS_18
And value matrix->
Figure QLYQS_22
In order to reduce the calculation amount of the model and prevent overfitting, the self-attention is obtained by calculating three matrixes through sparse self-attention, and the calculated self-attention is mapped into the original input dimension through a linear layer and then is combined with +.>
Figure QLYQS_26
Adding, and then obtaining the output +.A result of the adding is obtained through the feedforward layer>
Figure QLYQS_28
3. A method of predicting a climate environment in a power grid transmission line according to claim 2, wherein the sparse self-attention algorithm operates as: selecting a matrix
Figure QLYQS_29
Maximum value +.>
Figure QLYQS_30
The matrix after thinning is obtained by setting other elements to 0 by each element>
Figure QLYQS_31
Wherein->
Figure QLYQS_32
For a number of appropriate sizes set in advance, then matrix +.>
Figure QLYQS_33
,/>
Figure QLYQS_34
,/>
Figure QLYQS_35
The self-attention operation is carried out as input, and the formula is
Figure QLYQS_36
Wherein the method comprises the steps of
Figure QLYQS_37
Activating a function for softmax +.>
Figure QLYQS_38
For matrix->
Figure QLYQS_39
Dimension of->
Figure QLYQS_40
To calculate the self-attention.
4. The method of claim 2, wherein the decoding layer of the transform prediction model comprises a decoding one-dimensional convolution layer, a decoding sparse self-attention layer, a decoding feed-forward network layer, and a decoding linear layer, wherein the decoding layer performs prediction in such a way that the first input data is processed
Figure QLYQS_41
After decoding one-dimensional convolution layer and one layer of decoding sparse self-attention layer, mapping into query matrix by decoding linear layer>
Figure QLYQS_42
And thinning to obtain->
Figure QLYQS_43
Then use the second input data +.>
Figure QLYQS_44
Mapping into key matrix by decoding linear layer>
Figure QLYQS_45
And value matrix->
Figure QLYQS_46
Attention is then calculated, given by
Figure QLYQS_47
Wherein the method comprises the steps of
Figure QLYQS_48
Is->
Figure QLYQS_49
The calculated attention is decoded by the linear layer and the feedforward layer to obtain a predicted sequence of the transducer prediction model on future environment data.
5. A system for predicting the climate environment of a power grid transmission line, which is configured to implement the method for predicting the climate environment of a power grid transmission line according to any one of claims 1 to 4.
6. A computer readable storage medium, characterized in that the computer readable storage medium stores a program code which, when executed by a processor, implements the steps of the grid transmission line climate environment prediction method according to any of claims 1-4.
CN202211534275.4A 2022-12-02 2022-12-02 Method, system and medium for predicting climate environment of power grid transmission line Active CN115545361B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211534275.4A CN115545361B (en) 2022-12-02 2022-12-02 Method, system and medium for predicting climate environment of power grid transmission line

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211534275.4A CN115545361B (en) 2022-12-02 2022-12-02 Method, system and medium for predicting climate environment of power grid transmission line

Publications (2)

Publication Number Publication Date
CN115545361A CN115545361A (en) 2022-12-30
CN115545361B true CN115545361B (en) 2023-05-09

Family

ID=84722391

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211534275.4A Active CN115545361B (en) 2022-12-02 2022-12-02 Method, system and medium for predicting climate environment of power grid transmission line

Country Status (1)

Country Link
CN (1) CN115545361B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117490908B (en) * 2023-12-31 2024-04-09 武汉华康世纪医疗股份有限公司 Negative pressure detection method and system for negative pressure ward

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112819136A (en) * 2021-01-20 2021-05-18 南京邮电大学 Time sequence prediction method and system based on CNN-LSTM neural network model and ARIMA model
CN112990587A (en) * 2021-03-24 2021-06-18 北京市腾河智慧能源科技有限公司 Method, system, equipment and medium for accurately predicting power consumption of transformer area
CN113723669A (en) * 2021-08-09 2021-11-30 贵州电网有限责任公司 Power transmission line icing prediction method based on Informmer model
CN113919233A (en) * 2021-10-29 2022-01-11 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) Urban VOCs pollution total amount time sequence prediction method, system, storage medium and equipment
CN114239971A (en) * 2021-12-20 2022-03-25 浙江大学 Daily precipitation prediction method based on Transformer attention mechanism
CN114580710A (en) * 2022-01-28 2022-06-03 西安电子科技大学 Environment monitoring method based on Transformer time sequence prediction

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111985361A (en) * 2020-08-05 2020-11-24 武汉大学 Wavelet denoising and EMD-ARIMA power system load prediction method and system
CN114239718B (en) * 2021-12-15 2024-03-01 杭州电子科技大学 High-precision long-term time sequence prediction method based on multi-element time sequence data analysis
CN114943368A (en) * 2022-04-26 2022-08-26 天津大学 Sea surface wind speed prediction method based on Transformer
CN115049169B (en) * 2022-08-16 2022-10-28 国网湖北省电力有限公司信息通信公司 Regional power consumption prediction method, system and medium based on combination of frequency domain and spatial domain

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112819136A (en) * 2021-01-20 2021-05-18 南京邮电大学 Time sequence prediction method and system based on CNN-LSTM neural network model and ARIMA model
CN112990587A (en) * 2021-03-24 2021-06-18 北京市腾河智慧能源科技有限公司 Method, system, equipment and medium for accurately predicting power consumption of transformer area
CN113723669A (en) * 2021-08-09 2021-11-30 贵州电网有限责任公司 Power transmission line icing prediction method based on Informmer model
CN113919233A (en) * 2021-10-29 2022-01-11 合肥综合性国家科学中心人工智能研究院(安徽省人工智能实验室) Urban VOCs pollution total amount time sequence prediction method, system, storage medium and equipment
CN114239971A (en) * 2021-12-20 2022-03-25 浙江大学 Daily precipitation prediction method based on Transformer attention mechanism
CN114580710A (en) * 2022-01-28 2022-06-03 西安电子科技大学 Environment monitoring method based on Transformer time sequence prediction

Also Published As

Publication number Publication date
CN115545361A (en) 2022-12-30

Similar Documents

Publication Publication Date Title
Xiaoyun et al. Short-term prediction of wind power based on deep long short-term memory
CN111191783B (en) Self-adaptive quantization method and device, equipment and medium
CN113837370B (en) Method and apparatus for training a model based on contrast learning
CN115545361B (en) Method, system and medium for predicting climate environment of power grid transmission line
CN115049169B (en) Regional power consumption prediction method, system and medium based on combination of frequency domain and spatial domain
WO2023123930A1 (en) Image processing method, system, device and readable storage medium
Daş Forecasting the energy demand of Turkey with a NN based on an improved Particle Swarm Optimization
CN116504230A (en) Data closed-loop method, device, computer equipment and computer readable storage medium
CN117154680A (en) Wind power prediction method based on non-stationary transducer model
JP2024043504A (en) Acceleration method, device, electronic apparatus, and medium for neural network model inference
CN115600666A (en) Self-learning method and device for power transmission and distribution line defect detection model
CN114881344A (en) Training method, device and medium for building energy consumption prediction model
Zadorozhnyi Fractal queues simulation peculiarities
CN114819425A (en) Intelligent prediction method and system for regional power consumption and storage medium
CN113011597B (en) Deep learning method and device for regression task
CN111199153A (en) Word vector generation method and related equipment
CN110543549B (en) Semantic equivalence judgment method and device
CN108304513A (en) Increase the multifarious method and apparatus of production dialog model result
CN113962432A (en) Wind power prediction method and system integrating three-dimensional convolution and light-weight convolution threshold unit
CN110210518B (en) Method and device for extracting dimension reduction features
CN112732777A (en) Position prediction method, apparatus, device and medium based on time series
CN111856618A (en) Meteorological element prediction method and equipment
CN115563442A (en) XGboost type method for special initial value and related equipment
CN117059201B (en) Method, device, equipment and storage medium for predicting chemical oxygen demand of sewage
CN115496002B (en) Multi-dimensional feature interactive line dynamic capacity increasing method, system and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant