CN111222689A - LSTM load prediction method, medium, and electronic device based on multi-scale temporal features - Google Patents

LSTM load prediction method, medium, and electronic device based on multi-scale temporal features Download PDF

Info

Publication number
CN111222689A
CN111222689A CN201911244260.2A CN201911244260A CN111222689A CN 111222689 A CN111222689 A CN 111222689A CN 201911244260 A CN201911244260 A CN 201911244260A CN 111222689 A CN111222689 A CN 111222689A
Authority
CN
China
Prior art keywords
lstm
load
load prediction
data
prediction method
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911244260.2A
Other languages
Chinese (zh)
Inventor
杨梅
王仕发
牛晓伟
谢辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing Three Gorges University
Original Assignee
Chongqing Three Gorges University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing Three Gorges University filed Critical Chongqing Three Gorges University
Priority to CN201911244260.2A priority Critical patent/CN111222689A/en
Publication of CN111222689A publication Critical patent/CN111222689A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/06Energy or water supply

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Economics (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Human Resources & Organizations (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Strategic Management (AREA)
  • General Business, Economics & Management (AREA)
  • Marketing (AREA)
  • Tourism & Hospitality (AREA)
  • Primary Health Care (AREA)
  • Water Supply & Treatment (AREA)
  • Public Health (AREA)
  • Development Economics (AREA)
  • Game Theory and Decision Science (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a Long Short-Term Memory (LSTM) load prediction method (marked as W _ LSTM), a medium and an electronic device based on multi-scale time characteristics. The result of a comparison experiment of the method provided by the invention with a standard LSTM model, a Self-Organizing mapping network (SOM) and a Gaussian Process Regression (GPR) shows that the method provided by the invention has higher prediction precision. Meanwhile, experiments on the data of the noise load prove that the method also has certain anti-noise performance.

Description

LSTM load prediction method, medium, and electronic device based on multi-scale temporal features
Technical Field
The invention relates to a computer algorithm technology, in particular to an LSTM load prediction method, a medium and an electronic device based on multi-scale time characteristics.
Background
The power load data is the basis of reliable economic operation and scientific intelligent management of the power grid. Under the pressure of energy conservation and emission reduction and the promotion of an open power market, accurate load prediction has important significance on unit optimization, equipment maintenance, economic dispatching and the power market. However, the application of large-scale intermittent energy power generation systems and access and demand response mechanisms of electric vehicles causes high randomness and dynamic variability of power loads, and challenges are brought to load prediction. Meanwhile, the construction of the ubiquitous Internet of things and massive multi-source historical data provide a data basis for a new method for load prediction; therefore, load forecasting has new opportunities and challenges in the context of smart grids and artificial intelligence.
Because the traditional machine learning method cannot guarantee the dynamic load prediction precision under the background of high randomness and big data, a large number of researchers start turning to the deep learning-based method. Deep learning, as a branch of machine learning, has a deeper architecture (the number of network layers is 3 or more) and higher prediction accuracy. For example, a deep confidence network and a convolutional neural network are used for load prediction, and have higher prediction accuracy under the conditions that a training sample is large and load influence factors are complex, but have certain defects, such as easy occurrence of overfitting and no consideration of time correlation of power load. Therefore, a Recurrent Neural Network (RNN) model based on time series characteristics is used for load prediction, but it can only memorize short time series data, and as the data volume increases and the time interval increases, the RNN loses important information of a previous person, so that the gradient disappears and the prediction model fails. For this problem of RNN, LSTM is proposed, which has higher accuracy than the hyper-parametric optimized machine learning model. The gated round-robin unit model, after extrapolation, is also used for load prediction, with reduced model parameters compared to LSTM, but with somewhat poorer performance than LSTM.
Compared with single-factor load prediction, more and more researchers consider load prediction methods of composite influence factors. Besides historical load, other relevant factors such as date factors, meteorological factors, economic factors and the like are introduced into input data, so that more reliable and accurate prediction is achieved. This can improve accuracy to some extent, but the collection of such data requires additional effort and the introduction of additional variables may negatively impact predictive performance. For example, these variables are highly correlated, so the supervising algorithm may not be able to identify the true relationship of different time steps. Moreover, introducing too many input variables will greatly increase the complexity of the model, which will generally increase the risk of overfitting, increase the computational burden, and slow the convergence rate. The wavelet decomposition can realize the decomposition of the load data on different time scales, and the wavelet decomposition is carried out on the historical load data to express the load sequence components under the influence of different factors. LSTM has longer memory than RNN, but can only remember about 100s of long-term information. The literature decomposes photovoltaic sequences into different sequence components using wavelet transforms, reconstructs each sequence after prediction, which increases the accumulation of errors and is time consuming by reconstructing each sequence after prediction.
Disclosure of Invention
The technical problem to be solved by the invention is to provide an LSTM load prediction method, medium and electronic device based on multi-scale time characteristics, so as to solve the problems that the existing load prediction means for predicting and reconstructing each sequence increases error accumulation, the short-term load prediction precision is low and the prediction model is poor in robustness.
The invention is realized by the following technical scheme:
the LSTM load prediction method based on the multi-scale time characteristics comprises the following steps:
step 1: normalizing the original load sequence;
step 2: carrying out wavelet decomposition and reconstruction on the normalized original load sequence to obtain load sequences with different time characteristics;
and step 3: recombining the load sequences with different time characteristics to obtain a recombined sequence;
and 4, step 4: extracting data with greater relevance and diversity from the recombined sequences by a determinant point process;
and 5: and inputting the extracted data with larger correlation and diversity as input data into a trained LSTM load prediction model for load prediction.
Further, the wavelet is dbN wavelet.
Further, the vanishing moment N of the dbN wavelet is 15.
Further, the step 2 specifically includes performing 4-level wavelet decomposition and reconstruction on the normalized original load sequence to obtain 5 load sequences with different time characteristics.
Further, the LSTM load prediction model is trained through an adaptive moment estimation optimization algorithm.
A computer storage medium having stored thereon a computer program which, when executed by a processor, implements the LSTM load prediction method as described above.
An electronic device comprising a memory, a processor and a computer program stored in the memory and executable in the processor, the processor implementing the LSTM load prediction method as described above when executing the computer program.
Compared with the prior art, the LSTM load prediction method, the medium and the electronic device based on the multi-scale time characteristics, provided by the invention, have the advantages that the original load sequence is decomposed through wavelet decomposition to obtain the load sequences with different time characteristics, so that the signal change rule is subjected to deep analysis in the subsequent process, data with high relevance and diversity are extracted through a determinant point process to realize data dimension reduction so as to improve the training speed, and finally the data are input into an LSTM load prediction model to realize load prediction. Compared with a standard LSTM model, the method provided by the invention has higher prediction precision. Meanwhile, experiments on the data of the noise load prove that the method also has certain anti-noise performance.
Drawings
FIG. 1 is a schematic diagram of the structure of an LSTM cell;
FIG. 2 is a schematic diagram of a network structure of a layer 2 LSTM employed in an embodiment of the present invention;
FIG. 3 is a schematic diagram of the LSTM load prediction model structure and the working principle according to the embodiment of the present invention;
FIG. 4 is a schematic general flowchart of an LSTM load prediction method based on multi-scale temporal features according to an embodiment of the present invention;
FIG. 5 is a schematic diagram of a load sequence with different time characteristics according to an embodiment of the present invention;
FIG. 6 is a schematic diagram of a wavelet decomposition and sequence recombination implementation process;
FIG. 7 is a schematic diagram of data extraction by a routine punctuation process implemented in accordance with the present invention;
FIG. 8 is a schematic diagram of a training loss reduction curve comparison between the LSTM load prediction method based on multi-scale time characteristics and a standard LSTM;
FIG. 9a is a graph showing the results of Elia data set prediction comparison tests (load prediction before one hour) of the LSTM load prediction method based on multi-scale time characteristics and other methods provided by the embodiment of the present invention;
FIG. 9b is a diagram showing the results of Elia data set prediction comparison test (load prediction before day) based on LSTM load prediction method of multi-scale time characteristics and other methods provided by the embodiment of the present invention;
fig. 10 is a schematic diagram of evaluation indexes of the LSTM load prediction method based on multi-scale temporal features according to different proportions of noise according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention will be further described in detail with reference to the following embodiments and the accompanying drawings.
LSTM was proposed to solve the problem of gradient disappearance of RNN appearance, and is applied in many time series data scenarios, and the structure of LSTM unit is shown in fig. 1. Each LSTM unit contains a forgetting gate, an input gate, and an output gate, with three input data: output h of last moment of previous LSTM unitt-1State s of the last moment of the current LSTM cellt-1And current time input xt. The forgetting gate is used for calculating the forgotten information, the input gate is used for calculating the information stored in the state unit, the output gate calculates the information required to be output through a sigmoid function, and the information is multiplied by the value of the current unit state through a tanh function to obtain the output. The LSTM unit algorithm implements mathematical expressions of equations (1) - (6):
ft=σ(Wfx·xt+Wfh·ht-1+bf) (1)
it=σ(Wix·xt+Wih·ht-1+bi) (2)
gt=tanh(Wcx·xt+Wch·ht-1+bc) (3)
st=ft*st-1+it*gt(4)
ot=σ(Wox·xt+Woh·ht-1+bo) (5)
ht=ottanh(st) (6)
in formulae (1) to (6): f. oftIs a state of forgetting to leave the door it、gtIs the state of the input gate, otIs the state of the output gate, st、htThe state and output of the current LSTM cell, respectively. Wfx、Wfh、Wix、Wih、Wcx、 Wch、Wox、WohRespectively corresponding gate state and input xtAnd the output h of the last-in-time LSTM cellt-1The weights of the multiplications, σ, denote sigmoid activation functions, and σ denotes the bitwise multiplication of elements in the vector.
The inputs to the LSTM network output layer are:
Figure RE-GDA0002459589300000051
the LSTM network outputs are:
Figure RE-GDA0002459589300000052
the present invention employs a 2-layer LSTM network, the structure of which is shown in fig. 2.
The structure of the LSTM load prediction model based on the multi-scale time characteristics is shown in FIG. 3. The model mainly comprises two parts: data feature engineering and LSTM networks. The data characteristic engineering realizes the acquisition of load sequence components on different time scales; the LSTM network implements load prediction.
Based on the above basis, the embodiment of the invention provides an LSTM load prediction method based on multi-scale time characteristics. As shown in fig. 4, the LSTM load prediction method generally includes the steps of:
step S1: normalizing the original load sequence;
step S2: carrying out wavelet decomposition and reconstruction on the normalized original load sequence to obtain load sequences with different time characteristics;
step S3: recombining the load sequences with different time characteristics to obtain a recombined sequence;
step S4: extracting data with greater relevance and diversity from the recombined sequences by a determinant point process;
step S5: and inputting the extracted data with larger correlation and diversity as input data into a trained LSTM load prediction model for load prediction.
In step S1, in order to improve the convergence speed and accuracy of the model, the values of the load data cannot differ too much, so that the original load sequence needs to be normalized, and the normalization formula of the original load sequence in the embodiment of the present invention is:
Figure RE-GDA0002459589300000061
in the formula (9), x is a normalized value,
Figure RE-GDA0002459589300000062
is the average of the load data and,
Figure RE-GDA0002459589300000063
the number of data bits-1.
In step S2, each component with characteristic difference in the signal can be effectively separated by wavelet decomposition, so as to obtain a stationary sequence and non-stationary sequences on different scales. Commonly used wavelets are Haar wavelets, Daubeches (dbN) wavelets, Mexican Hat wavelets, Meyer wavelets, and the like. According to the embodiment of the invention, dbN wavelets are selected, the vanishing moment N of dbN wavelets is selected to be 15 through experiments, and the normalized original load sequence is subjected to 4-level wavelet decomposition and reconstruction. That is, in this embodiment, step S2 is specifically to perform 4-level wavelet decomposition and reconstruction on the normalized original load sequence, and obtain 5 load sequences with different time characteristics as shown in fig. 5. As can be seen from fig. 5, in the obtained 5 load sequences (components) with different time characteristics, the level 3 detail component and the level 4 detail component are relatively stable, and the level 4 approximate component reflects the load trend, such as the change of seasons and the load increase caused by the economic and industrial development of the area, and can reflect the change law of the load for a long time; the layer 4 detail component represents a fixed invariant portion of the load; the fluctuation of the detail component of the layer 3 can reflect the growing intensity of the trend load; the fluctuation of detail components of layer 1 and layer 2 is large, and represents the magnitude and duration of peak-valley value on the load curve to a certain extent, and can reflect the daily load time characteristic.
The wavelet decomposition and sequence reassembly implementation processes described in steps S2 and S3 are shown in fig. 6, where M is the number of load sequence data, 4-layer wavelet decomposition is employed to obtain 5 data sequences with different characteristics, and the data sequences are marked as H1={h11,h12,…,h1M}、H2={h21,h22,…,h2M}、H3={h31,h32,…,h3M}、H4.d={h41.d,h42.d,…,h4M.d}、 H4.a={h41.a,h42.a,…,h4M.a}. The input required for LSTM is sequence data, which requires the reassembly of 4 sequences in a time relationship, the reassembled sequence being Hr={h11,h21,h31,h41.d,h41.a;h12,h22,h32,h42.d,h42.a;…;h1M,h2M,h3M,h4M.d,h4M.a}. Increasing the number of layers of the LSTM may improve the predictive power of the model, but too many may also cause an overfitting phenomenon. According to the quantity of the training data, the number of layers of the model and the quantity of neurons in each layer are selected through an experimental method.
In step S4, the determinant point process defines a discrete set D ═ D1,d2,d3,…,dnProbability distribution of each subset Y of }: given the probability of occurrence of the null set, there is a real symmetric semi-positive definite matrix L ∈ R consisting of the elements of the set Dn×nFor each subset Y of the set D, the probability P (Y) that the subset Y appears, oc det (L)Y)。
Figure RE-GDA0002459589300000071
L in the formula (10)YA submatrix representing a matrix L composed of rows and columns having subscripts belonging to Y, and det () representing a value of a determinant. I is as large as Rn×nIs an identity matrix. Since L is a semi-positive definite matrix, there is a matrix V such that L ═ VVT,V∈Rn×r. Likewise, LYCan be decomposed into
Figure RE-GDA0002459589300000072
In practical application, each column vector V in the matrix V is usediThe decomposition is as follows:
Vi=rifi(11)
then there are:
Lij=〈Vi,Vj〉=〈rifi,rjfj〉=rirjsij(12)
in the formulae (11) and (12), riData d of not less than 0iCorrelation with the data set D. sij=〈fi,fjIs diAnd djThe more similar the similarity information is, the smaller the value is, and the | fi||2=1;<>Representing a vector cross product.
As can be seen from equation (12), the greater the correlation between the data and the data set, and the richer the data diversity, the larger the determinant of the matrix L. Therefore, by finding L with the largest determinant valueYThe optimal dimensionality reduction data subset Y can be obtainedmThe specific calculation formula is as follows:
Ym=argmax{det(LY)} (13)
data extracted by using a determinant point process algorithm is shown in fig. 7, wherein + represents original data, solid dots represent extracted data with large correlation and diversity, the extracted data are used as input data and input into a trained LSTM load prediction model to realize load prediction, and a load prediction result is output.
In terms of model training, the embodiment of the present invention adopts an ADAM (Adaptive motion Estimation) optimization algorithm to train the LSTM load prediction model based on the multi-scale time features. The method is an algorithm which combines the Momentum algorithm and the RMSProp algorithm, and has the characteristics of high convergence rate and small fluctuation amplitude. The loss function is minimized by iteratively updating the weights and biases of the network nodes. The loss function uses Mean Square Error (MSE):
Figure RE-GDA0002459589300000081
in the formula (14), n is the number of data, yiIs the actual value of the load and,
Figure RE-GDA0002459589300000082
is the corresponding predicted value.
The number of layers of the LSTM network is 2, the number of neurons in the first layer is 50, the number of neurons in the second layer is 100, and the number of iterations is 200. To avoid overfitting, the value of dropout is set to 0.2. The comparison between the LSTM load prediction method based on multi-scale time characteristics and the training loss decline curve of the standard LSTM is shown in FIG. 8: when the training times reach 150 times, the loss is basically stable and does not decrease, so the number of the training times is 150. As can be seen from FIG. 8, the training loss of the LSTM load prediction model based on multi-scale temporal features of the invention is smaller than that of the conventional standard LSTM.
The technical effect experiment analysis of the invention is as follows:
experimental hardware environment and experimental data
The hardware adopts a computer with a processor of Intel (R) core (TM) i3, a main frequency of 3.30GHZ and a memory of 12G. Under the Tensorflow framework, the programming is done using the python language. The invention performs experiments on load data (referred to as Elia data set for short) disclosed by Elia of belgium electric network operators. Elia load data was collected every 15 minutes from 2004 to 2014, and the present invention was uniformly selected as 24 points per day. Training samples are load data of one to two years, the first 95% are sample data, and the last 5% are test data. The processing time scheduled in each country differs, for example one hour ahead in the uk and one day ahead in china, so the experiments of the present invention are load prediction one hour before and one day before, respectively. In the one-hour pre-load prediction experiment, each training sample was constructed by: 24 load data of the first 24 hours are taken as input, and 25 th data are taken as output; in the day-ahead load prediction experiment, the data of each training sample is as follows: 168 load data on the first 7 days were input and 24 data on day 8 were output.
In order to prove the effectiveness of the technical scheme of the invention, the prediction results of the SOM, LSTM and GP models are compared with the method of the invention. The experiments are divided into two groups, wherein the first group is a comparison experiment of load prediction before one hour and load prediction before day on an Elia data set; in the second group, 1% -3% of noise is added into the Elia data set to perform an anti-noise performance test experiment.
Evaluation index of experiment
In order to evaluate the prediction performance of the technical scheme of the invention, the average absolute percentage error E is selectedMAPERoot mean square error ERMSEAnd determining three indexes of a coefficient (R Squared, RS), wherein the expression is as follows:
Figure RE-GDA0002459589300000101
Figure RE-GDA0002459589300000102
the coefficient is determined by representing the quality of fitting through the change of data, the value range is 0-1, and the closer to 1, the better the model fits the data. Is the ratio of the Sum of Squares of Regression (SSR) to the Sum of Squares of Total (SST).
Figure RE-GDA0002459589300000103
Formula (II)
Figure RE-GDA0002459589300000104
In, yiIs the actual value of the load,
Figure RE-GDA0002459589300000105
in order to predict the value of the load,
Figure RE-GDA0002459589300000106
is the average value of the actual load, and n is the number of data.
Analysis of Experimental results
Experiment on Elia data set:
fig. 9 shows predicted load and actual load for three days in 2013, 12, wherein fig. 9(a) shows the result of the one-hour load prediction experiment and fig. 9b shows the result of the one-day load prediction experiment. By combining fig. 9(a) and fig. 9(b), the W _ LSTM of the present invention is superior to other algorithms for load prediction one hour ago or day ago. With the increase of the prediction period, the relevance of data on time is reduced, so that the accuracy of load prediction is reduced, and the prediction precision is reduced compared with other methods because the LSTM load prediction method based on the multi-scale time characteristics utilizes different time scale characteristics of wavelets.
The following table is an evaluation index of prediction accuracy on Elia data set. It can be seen that the method of the present invention performs optimally under three evaluation indexes. Compared with Self-Organizing mapping (SOM), Gaussian Process Regression (GPR) and LSTM, in the one-hour pre-load prediction experiment, E is a measure of the loadMAPERespectively increased by 2.308, 1.474 and 1.425; eRMSE106.085, lower by 265.597, 169.165, 155.4, respectively; the RS is 0.980, which is improved by 0.227, 0.115 and 0.102 respectively. In the experiment for predicting the load before day, EMAPE3.741, compared with the other three methods, the yield is respectively improved by 3.550, 2.877 and 4.699; eRMSERespectively 332.322, 307.831, 450.424; RS is improved by 0.525, 0.389 and 0.69 respectively. Therefore, the prediction precision and the model performance of the method are greatly improved. Comparing the experimental results of load prediction before one hour and load prediction before the day, the prediction accuracy of the method is reduced with the increase of the prediction period, but the method is reduced less compared with other three methods; illustrating that the method of the present invention further exploits the temporal characteristics of the load sequence.
Figure RE-GDA0002459589300000111
And (3) interference resistance analysis:
in practice, noise or data loss is inevitably generated in data acquisition, and in order to test the anti-noise performance of the method, the robustness of the model is tested through a disturbance analysis experiment. 1% -3% of random noise is randomly added into training data, and the evaluation indexes of the LSTM load prediction method under different proportion noises are expressed by using average absolute error percentage and prediction precision. The results are shown in fig. 10, in which the horizontal axis represents the noise belgium and the vertical axis represents the percentage of the average absolute error and the change in prediction accuracy. It can be seen that when the noise is within 2%, the change of the metric value is small, which indicates that the W _ LSTM model has certain robustness and can handle the noise problem within a certain range; after the noise exceeds 2%, the fitting capability of the model is greatly reduced.
Based on the LSTM load prediction method based on the multi-scale time characteristics, the embodiment of the invention also provides a computer storage medium. The computer storage medium has stored thereon a computer program which, when executed by a processor, implements the LSTM load prediction method as described above.
Based on the LSTM load prediction method based on the multi-scale time characteristics, the embodiment of the invention also provides an electronic device. The electronic device comprises a memory, a processor and a computer program stored in the memory and capable of running in the processor, wherein the processor executes the computer program to realize the LSTM load prediction method.
The above embodiments are only preferred embodiments and are not intended to limit the scope of the present invention. Any modification, equivalent replacement, and improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (7)

1. The LSTM load prediction method based on the multi-scale time characteristics is characterized by comprising the following steps:
step 1: normalizing the original load sequence;
step 2: carrying out wavelet decomposition and reconstruction on the normalized original load sequence to obtain load sequences with different time characteristics;
and step 3: recombining the load sequences with different time characteristics to obtain a recombined sequence;
and 4, step 4: extracting data with greater relevance and diversity from the recombined sequences by a determinant point process;
and 5: and inputting the extracted data with larger correlation and diversity as input data into a trained LSTM load prediction model for load prediction.
2. The LSTM load prediction method based on multi-scale temporal features of claim 1, wherein the wavelet is dbN wavelet.
3. The LSTM load prediction method based on multi-scale temporal features of claim 2 wherein the vanishing moment N of the dbN wavelet is 15.
4. The LSTM load prediction method based on multi-scale temporal features as claimed in claim 3, wherein said step 2 is specifically to perform 4-level wavelet decomposition and reconstruction on the normalized original load sequence to obtain 5 load sequences with different temporal characteristics.
5. The LSTM load prediction method based on multi-scale temporal features of claim 1, wherein the LSTM load prediction model is trained by an adaptive moment estimation optimization algorithm.
6. A computer storage medium having a computer program stored thereon, wherein the computer program, when executed by a processor, implements the LSTM load prediction method of any of claims 1 to 5.
7. An electronic device comprising a memory, a processor, and a computer program stored in the memory and executable in the processor, wherein the processor implements the LSTM load prediction method of any of claims 1 to 5 when executing the computer program.
CN201911244260.2A 2019-12-06 2019-12-06 LSTM load prediction method, medium, and electronic device based on multi-scale temporal features Pending CN111222689A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911244260.2A CN111222689A (en) 2019-12-06 2019-12-06 LSTM load prediction method, medium, and electronic device based on multi-scale temporal features

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911244260.2A CN111222689A (en) 2019-12-06 2019-12-06 LSTM load prediction method, medium, and electronic device based on multi-scale temporal features

Publications (1)

Publication Number Publication Date
CN111222689A true CN111222689A (en) 2020-06-02

Family

ID=70829880

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911244260.2A Pending CN111222689A (en) 2019-12-06 2019-12-06 LSTM load prediction method, medium, and electronic device based on multi-scale temporal features

Country Status (1)

Country Link
CN (1) CN111222689A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112330078A (en) * 2021-01-04 2021-02-05 南方电网数字电网研究院有限公司 Power consumption prediction method and device, computer equipment and storage medium
CN113068211A (en) * 2021-04-21 2021-07-02 东南大学 Wireless access point throughput prediction method based on deep learning and Gaussian process regression
CN115760850A (en) * 2023-01-05 2023-03-07 长江勘测规划设计研究有限责任公司 Method for identifying water level without scale by using machine vision
CN117081082A (en) * 2023-10-17 2023-11-17 国网上海市电力公司 Active power distribution network operation situation sensing method and system based on Gaussian process regression

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107071423A (en) * 2017-04-24 2017-08-18 天津大学 Application process of the vision multi-channel model in stereoscopic video quality objective evaluation
CN108272463A (en) * 2017-08-07 2018-07-13 上海交通大学 A kind of flyer's fatigue detection method based on EEG signals
US20190004490A1 (en) * 2017-06-28 2019-01-03 Siemens Aktiengesellschaft Method for recognizing contingencies in a power supply network
CN109472404A (en) * 2018-10-31 2019-03-15 山东大学 A kind of Short-Term Load Forecasting of Electric Power System, model, apparatus and system
US20190228481A1 (en) * 2018-01-22 2019-07-25 Pason Power Inc. Intelligent energy management system for distributed energy resources and energy storage systems using machine learning
CN110070229A (en) * 2019-04-26 2019-07-30 中国计量大学 The short term prediction method of home electrical load
CN110263866A (en) * 2019-06-24 2019-09-20 苏州智睿新能信息科技有限公司 A kind of power consumer load setting prediction technique based on deep learning
US20190340392A1 (en) * 2018-05-04 2019-11-07 New York University Anomaly detection in real-time multi-threaded processes on embedded systems and devices using hardware performance counters and/or stack traces

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107071423A (en) * 2017-04-24 2017-08-18 天津大学 Application process of the vision multi-channel model in stereoscopic video quality objective evaluation
US20190004490A1 (en) * 2017-06-28 2019-01-03 Siemens Aktiengesellschaft Method for recognizing contingencies in a power supply network
CN108272463A (en) * 2017-08-07 2018-07-13 上海交通大学 A kind of flyer's fatigue detection method based on EEG signals
US20190228481A1 (en) * 2018-01-22 2019-07-25 Pason Power Inc. Intelligent energy management system for distributed energy resources and energy storage systems using machine learning
US20190340392A1 (en) * 2018-05-04 2019-11-07 New York University Anomaly detection in real-time multi-threaded processes on embedded systems and devices using hardware performance counters and/or stack traces
CN109472404A (en) * 2018-10-31 2019-03-15 山东大学 A kind of Short-Term Load Forecasting of Electric Power System, model, apparatus and system
CN110070229A (en) * 2019-04-26 2019-07-30 中国计量大学 The short term prediction method of home electrical load
CN110263866A (en) * 2019-06-24 2019-09-20 苏州智睿新能信息科技有限公司 A kind of power consumer load setting prediction technique based on deep learning

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
ZIHAN CHANG ET AL: "Electricity price prediction based on hybrid model of adam optimized LSTM neural network and wavelet transform", 《ENERGY》 *
杨梅等: "基于多尺度时间特征LSTM 的短期负荷预测", 《控制工程》 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112330078A (en) * 2021-01-04 2021-02-05 南方电网数字电网研究院有限公司 Power consumption prediction method and device, computer equipment and storage medium
CN113068211A (en) * 2021-04-21 2021-07-02 东南大学 Wireless access point throughput prediction method based on deep learning and Gaussian process regression
CN115760850A (en) * 2023-01-05 2023-03-07 长江勘测规划设计研究有限责任公司 Method for identifying water level without scale by using machine vision
CN117081082A (en) * 2023-10-17 2023-11-17 国网上海市电力公司 Active power distribution network operation situation sensing method and system based on Gaussian process regression
CN117081082B (en) * 2023-10-17 2024-01-23 国网上海市电力公司 Active power distribution network operation situation sensing method and system based on Gaussian process regression

Similar Documents

Publication Publication Date Title
CN109492822B (en) Air pollutant concentration time-space domain correlation prediction method
Ma et al. A hybrid attention-based deep learning approach for wind power prediction
CN110909926A (en) TCN-LSTM-based solar photovoltaic power generation prediction method
CN111222689A (en) LSTM load prediction method, medium, and electronic device based on multi-scale temporal features
CN113297801A (en) Marine environment element prediction method based on STEOF-LSTM
CN110309603B (en) Short-term wind speed prediction method and system based on wind speed characteristics
CN111860982A (en) Wind power plant short-term wind power prediction method based on VMD-FCM-GRU
CN112766078B (en) GRU-NN power load level prediction method based on EMD-SVR-MLR and attention mechanism
CN112733444A (en) Multistep long time sequence prediction method based on CycleGAN neural network
CN110751318A (en) IPSO-LSTM-based ultra-short-term power load prediction method
CN110987436B (en) Bearing fault diagnosis method based on excitation mechanism
CN110766060B (en) Time series similarity calculation method, system and medium based on deep learning
Gu et al. Bayesian Takagi–Sugeno–Kang fuzzy model and its joint learning of structure identification and parameter estimation
CN116562908A (en) Electric price prediction method based on double-layer VMD decomposition and SSA-LSTM
CN113095598A (en) Multi-energy load prediction method, system, device and medium
CN114022311A (en) Comprehensive energy system data compensation method for generating countermeasure network based on time sequence condition
CN111191823B (en) Deep learning-based production logistics prediction method
CN116169670A (en) Short-term non-resident load prediction method and system based on improved neural network
CN115659254A (en) Power quality disturbance analysis method for power distribution network with bimodal feature fusion
CN116665483A (en) Novel method for predicting residual parking space
CN114596726B (en) Parking berth prediction method based on interpretable space-time attention mechanism
CN116632834A (en) Short-term power load prediction method based on SSA-BiGRU-Attention
CN116720057A (en) River water quality prediction method and system based on feature screening and weight distribution
CN116542701A (en) Carbon price prediction method and system based on CNN-LSTM combination model
CN111090679A (en) Time sequence data representation learning method based on time sequence influence and graph embedding

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200602