CN112396234A - User side load probability prediction method based on time domain convolutional neural network - Google Patents

User side load probability prediction method based on time domain convolutional neural network Download PDF

Info

Publication number
CN112396234A
CN112396234A CN202011308724.4A CN202011308724A CN112396234A CN 112396234 A CN112396234 A CN 112396234A CN 202011308724 A CN202011308724 A CN 202011308724A CN 112396234 A CN112396234 A CN 112396234A
Authority
CN
China
Prior art keywords
load
data
time
neural network
quantile
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011308724.4A
Other languages
Chinese (zh)
Inventor
原凯
宋毅
陈浩然
张真源
孙充勃
潘尔生
郭铭群
吴志力
姜世公
李敬如
杨卫红
靳夏宁
胡丹蕾
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Electronic Science and Technology of China
State Grid Tianjin Electric Power Co Ltd
State Grid Economic and Technological Research Institute
Original Assignee
University of Electronic Science and Technology of China
State Grid Tianjin Electric Power Co Ltd
State Grid Economic and Technological Research Institute
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Electronic Science and Technology of China, State Grid Tianjin Electric Power Co Ltd, State Grid Economic and Technological Research Institute filed Critical University of Electronic Science and Technology of China
Priority to CN202011308724.4A priority Critical patent/CN112396234A/en
Publication of CN112396234A publication Critical patent/CN112396234A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/06Energy or water supply

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Economics (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Human Resources & Organizations (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Strategic Management (AREA)
  • General Business, Economics & Management (AREA)
  • Marketing (AREA)
  • Tourism & Hospitality (AREA)
  • Primary Health Care (AREA)
  • Water Supply & Treatment (AREA)
  • Public Health (AREA)
  • Development Economics (AREA)
  • Game Theory and Decision Science (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a user side load probability prediction method based on a time domain convolutional neural network, which comprises the steps of firstly collecting historical load data of a user and weather types and temperature values in a corresponding time range, and then preprocessing the historical load data to meet the input requirement of the time domain convolutional neural network; then, a time domain convolution neural network is constructed to serve as a load probability prediction model, and the time domain convolution neural network can accurately predict the load probability through model training; and finally, predicting the load data acquired by the user in real time.

Description

User side load probability prediction method based on time domain convolutional neural network
Technical Field
The invention belongs to the technical field of power load prediction, and particularly relates to a user side load probability prediction method based on a time domain convolutional neural network.
Background
With the widespread use of advanced metering infrastructure in modern power systems, fine load curves and individual power usage behaviors at the customer end can be easily obtained. However, the explosive growth of data exceeds the capabilities of conventional load forecasting systems, and in particular, the vast amount of data, if collected, cannot be efficiently analyzed and utilized. In recent years, with the development of Artificial Intelligence (AI) technology, abundant fine-grained electricity consumption data can be processed, analyzed and utilized. Load forecasting for individual customers provides the opportunity for power companies to refine the management of the grid, thereby enabling better decisions on the power retail market.
In order to better analyze mass household electrical load data and obtain an accurate load prediction result, a plurality of load prediction methods based on machine learning are available at present, and the methods can be divided into two types according to the development stage of the machine learning, wherein the first type is a traditional machine learning method based on a support vector machine, an autoregressive integrated moving average (ARIMA), a clustering method (clustering-based approach) and the like. The second category is based on Recurrent Neural Networks (RNN) and Convolutional Neural Networks (CNN), and deep learning methods derived on this basis.
However, in the first method, the model often requires complex feature engineering from human experience, which not only brings a lot of workload, but also results in poor generalization of the prediction model. The second method is different from the traditional machine learning method, and the deep learning method has a highly flexible framework, can directly learn the characteristics from the original data, and can improve the prediction precision. However, the existing personal user short-term load prediction method based on RNN and CNN has model structural defects, and is difficult to deal with serious practical challenges: in the face of high-frequency model training and online updating of mass load data, the most basic requirement of a prediction model is as reliable as possible, and meanwhile, efficient calculation is carried out.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provides a user side load probability prediction method based on a time domain convolutional neural network, and the probability prediction of user load data is realized based on a load probability prediction model of the time domain convolutional neural network.
In order to achieve the above object, the present invention provides a method for predicting a load probability on a user side based on a time domain convolutional neural network, which is characterized by comprising the following steps:
(1) data acquisition and preprocessing
(1.1) collecting historical load, temperature and weather data
Active power of a user is collected as historical load data X according to a fixed period within a certain time rangeloadSimultaneously collecting weather types X in corresponding time rangesweatherAnd a temperature value XtemperatureThe weather and temperature data is then processed into data samples X ═ X (X) at the same frequency as the historical load dataweather,Xtemperature,Xload);
(1.2) data sample cleaning
Performing data cleaning on the data sample X, replacing an abnormal value in the data sample with a missing value, and then performing proximity interpolation processing at the missing value to obtain a complete data sample X' (X)weather′,Xtemperature′,Xload′);
(1.3), Z-score normalization
And (5) carrying out standardization processing on the complete data sample to obtain a standard data sample X ″ (X ═ X)weather″,Xtemperature″,Xload"), the Z-score normalization formula is as follows:
Figure BDA0002789039670000021
wherein x is a standard value, x is a data sample value before processing, mu is an average value of the data sample, and sigma is a standard deviation of the data sample;
(1.4) addition time characteristics
Taking the time corresponding to the standard data sample, including the week and time characteristics, as a fourth characteristic X in the form of One-hot codingdateAdded to X ", a standard data sample with time signature X" (X)weather″,Xtemperature″,Xload″,Xdate);
(1.5) construction of Standard data set
On X', a standard dataset was constructed using a sliding window method.
(1.5.1) generating the t frame of the standard data set by a sliding window method
A segment of data { X '″ of length n preceding time t is taken at X't-n,x″′i-n+1,...,x″′t-1And taking load data x' at the time t as a training sample of the modelloadAs a training label of the model, the training sample and the training label are correspondingly matched and used as one frame of the standard data set;
(1.5.2) repeating the step (1.5.1), and sequentially sliding backwards for M steps on the X' to construct a standard data set with the length of M frames;
(2) and constructing a load probability prediction model
Constructing a time domain convolution-based neural network as a load probability prediction model, wherein the time domain convolution neural network is formed by serially connecting an input layer, a plurality of residual blocks and an output layer, the output layer is formed by connecting three full-connection layers in parallel, and the number of quantiles corresponds to the number of the full-connection layers;
the ith residual block is provided with two input branches, wherein the left input branch is the input of the output of the (i-1) th residual block to the ith residual block after the computation of the convolutional layer, the weight normalization, the ReLU activation function and the neuron immediate inactivation (dropout) respectively; the right input branch is to directly input the output of the i-1 th residual block to the i-th residual block after 1 × 1 convolution; wherein the convolution layer is a convolution kernel formed by a plurality of dilation causal convolutions;
(3) training load probability prediction model
(3.1) randomly selecting a frame of data from the standard data set, inputting the frame of data into a time domain convolutional neural network model, converting the frame of data into a tensor form through an input layer, and inputting the tensor form into the serially connected residual blocks;
(3.2) in the time domain convolution neural network model, the input tensor of the ith residual block is set as Z(i-1)I ═ 1,2, …, K denotes the number of residual blocks; in the left branch of the ith residual block, the tensor Z(i-1)Performing feature extraction through a convolution kernel formed by a plurality of dilation causal convolutions, and then sequentially performing batch normalization, layer normalization, ReLU activation function and neuron random inactivation to obtain the output tensor of the left branch
Figure BDA0002789039670000031
In the right branch of the i-th residual block, the tensor Z(i-1)Convolving by 1 × 1 to make its output tensor
Figure BDA0002789039670000032
Matching the dimensionality of the output tensors of the left branch, and then adding the output tensors of the two branches to obtain the output of the ith residual block
Figure BDA0002789039670000033
(3.3) repeating the step (3.2) until the last residual block output Z(K)Finally, Z is(K)The outputs of the three full-connection layers connected in parallel are respectively recorded as
Figure BDA0002789039670000034
And
Figure BDA0002789039670000035
(3.4) calculating the upper quantile q at the time t by a Pinball quantile formulathLower quantile (1-q)thAnd 0.5thLoss value L corresponding to quantilet,q(ut,q),Lt,(1-q)(ut,(1-q)) And Lt,0.5(ut,0.5)
(3.4.1) setting the preset probability of the Pinball quantile as q;
(3.4.2) calculating the upper quantile q at time tthLower quantile (1-q)thAnd 0.5thQuantile prediction error
Figure BDA0002789039670000036
And
Figure BDA0002789039670000037
ytis the real load value;
(3.4.3) calculating the upper quantile q at the time tthLower quantile (1-q)thAnd 0.5thLoss value L corresponding to quantilet,q(ut,q),Lt,(1-q)(ut,(1-q)) And Lt,0.5(ut,0.5);
Figure BDA0002789039670000041
Figure BDA0002789039670000042
Figure BDA0002789039670000043
(3.5) carrying out weighted average on the loss values;
Figure BDA0002789039670000044
wherein the content of the first and second substances,
Figure BDA0002789039670000045
respectively represent Lt,q(ut,q),Lt,(1-q)(ut,(1-q)) And Lt,0.5(ut,0.5) The weight of (c);
(3.6) repeating the steps (3.1) - (3.5), comparing the weighted average loss obtained by each round of training, executing a learning rate automatic attenuation strategy when the weighted average loss does not decrease in the P round of training, and executing an early stop strategy when the weighted average loss does not decrease in the 3 x P round of training, namely automatically stopping the training to obtain a trained load probability prediction model; otherwise, entering the step (3.7);
(3.7) updating the weight parameters in the time domain convolution neural network model in the negative gradient direction of the weighted average loss function by using a batch gradient descent algorithm according to the negative gradient direction, and then returning to the step (3.1);
(4) load probability real-time prediction
Real-time collection of load data of a user, and weather and temperature related to the load dataBuilding a frame of standard data according to the method in the step (1), and inputting the standard data into a trained load probability prediction model so as to output a load probability prediction interval
Figure BDA0002789039670000046
And deterministic prediction results
Figure BDA0002789039670000047
The invention aims to realize the following steps:
the invention relates to a user side load probability prediction method based on a time domain convolutional neural network, which comprises the steps of firstly collecting historical load data of a user and weather types and temperature values in a corresponding time range, and then preprocessing the historical load data to meet the input requirement of the time domain convolutional neural network; then, a time domain convolution neural network is constructed to serve as a load probability prediction model, and the time domain convolution neural network can accurately predict the load probability through model training; and finally, predicting the load data acquired by the user in real time.
Meanwhile, the user side load probability prediction method based on the time domain convolution neural network also has the following beneficial effects:
(1) the time domain convolution neural network model is used as the load probability prediction model, and the time domain convolution neural network model can give consideration to both calculation efficiency and prediction reliability;
(2) the method carries out weighted loss calculation on the output of the time domain convolution neural network model through a Pinball quantile function and a weighted average mode, thereby expanding a deterministic prediction model into a probabilistic prediction model;
(3) in the training process of the model, the early-stopping strategy and the learning rate automatic attenuation strategy can reduce the long-tail effect in the training process of the model.
Drawings
FIG. 1 is a flow chart of a user-side load probability prediction method based on a time domain convolutional neural network according to the present invention;
FIG. 2 is a schematic diagram of a residual block;
FIG. 3 is a schematic diagram of a Pinball quantile function;
FIG. 4 is a probability prediction curve for the method of the present invention along with several other methods;
fig. 5 shows probability prediction curves for different predetermined probability Prediction Intervals (PINC) according to the method of the invention.
Detailed Description
The following description of the embodiments of the present invention is provided in order to better understand the present invention for those skilled in the art with reference to the accompanying drawings. It is to be expressly noted that in the following description, a detailed description of known functions and designs will be omitted when it may obscure the subject matter of the present invention.
Examples
FIG. 1 is a flow chart of a user-side load probability prediction method based on a time domain convolutional neural network.
In this embodiment, as shown in fig. 1, the method for predicting the load probability on the user side based on the time domain convolutional neural network of the present invention includes the following steps:
s1, data acquisition and preprocessing
S1.1, collecting historical load, temperature and weather data
Active power of a user is collected as historical load data X according to a fixed period within a certain time rangeloadSimultaneously collecting weather types X in corresponding time rangesweatherAnd a temperature value XtemperatureIn the present embodiment, the sampling period is 15 minutes, i.e. 96 points/day cross section; the weather and temperature data is then processed into data samples X ═ X (X) at the same frequency as the historical load dataweather,Xtemperature,Xload);
S1.2, cleaning data samples
Performing data cleaning on the data sample X, replacing an abnormal value in the data sample with a missing value, and then performing proximity interpolation processing at the missing value to obtain a complete data sample X' (X)weather′,Xtemperature′,Xload′);
S1.3, Z-score normalization
And (5) carrying out standardization processing on the complete data sample to obtain a standard data sample X ″ (X ═ X)weather″,Xtemperature″,Xload"), the Z-score normalization formula is as follows:
Figure BDA0002789039670000061
wherein x is a standard value, x is a data sample value before processing, mu is an average value of the data sample, and sigma is a standard deviation of the data sample; the standardization processing can unify the magnitude and dimension of data, so that the generalization capability of the model is enhanced;
s1.4, time of addition feature
Taking the time corresponding to the standard data sample, including the week and time characteristics, as a fourth characteristic X in the form of One-hot codingdateAdded to X ", a standard data sample with time signature X" (X)weather″,Xtemperature″,Xload″,Xdate);
S1.5, constructing a standard data set
On X', a standard dataset was constructed using a sliding window method.
S1.5.1 sliding window method for generating t frame of standard data set
A segment of data { X '″ of length n preceding time t is taken at X't-n,x″′t-n+1,...,x″′t-1And taking load data x' at the time t as a training sample of the modelloadAs a training label of the model, the training sample and the training label are correspondingly matched and used as one frame of the standard data set;
s1.5.2, repeating S1.5.1 steps, sliding backwards in sequence by M steps on the X' to construct a standard data set with the length of M frames;
s2, constructing a load probability prediction model
Constructing a time domain convolution-based neural network as a load probability prediction model, wherein the time domain convolution neural network is formed by serially connecting an input layer, a plurality of residual blocks and an output layer, the output layer is formed by connecting three full-connection layers in parallel, and the number of quantiles corresponds to the number of the full-connection layers;
as shown in fig. 2, the ith residual block has two input branches, wherein the left input branch is the output of the (i-1) th residual block, which is input to the ith residual block after being respectively calculated by convolutional layer, weight normalization, ReLU activation function and neuron immediate deactivation (drop); the right input branch is to directly input the output of the i-1 th residual block to the i-th residual block after 1 × 1 convolution; wherein the convolution layer is a convolution kernel formed by a plurality of dilation causal convolutions;
s3 training load probability prediction model
S3.1, randomly selecting a frame of data from the standard data set, inputting the frame of data into a time domain convolution neural network model, converting the frame of data into a tensor form through an input layer, and inputting the tensor form into the serially connected residual blocks;
s3.2, in the time domain convolution neural network model, the input tensor of the ith residual block is set to be Z(i-1)I ═ 1,2, …, K denotes the number of residual blocks; in the left branch of the ith residual block, the tensor Z(i-1)Performing feature extraction through a convolution kernel formed by a plurality of dilation causal convolutions, and then sequentially performing batch normalization, layer normalization, ReLU activation function and neuron random inactivation to obtain the output tensor of the left branch
Figure BDA0002789039670000071
In the right branch of the i-th residual block, the tensor Z(i-1)Convolving by 1 × 1 to make its output tensor
Figure BDA0002789039670000072
Matching the dimensionality of the output tensors of the left branch, and then adding the output tensors of the two branches to obtain the output of the ith residual block
Figure BDA0002789039670000073
In this embodiment, the function f(s) of the dilated causal convolution is:
Figure BDA0002789039670000074
wherein the content of the first and second substances,
Figure BDA0002789039670000075
representing an input tensor; (i) denotes a convolution filter, i ═ 0,1,2, …, λ -1, λ being the size of the convolution filter; s-d · i denotes the range of each step of convolution over the element s in the input tensor, d being the expansion factor.
S3.3, repeating the step S3.2 until the last residual block outputs Z(K)Finally, Z is(K)The outputs of the three full-connection layers connected in parallel are respectively recorded as
Figure BDA0002789039670000076
And
Figure BDA0002789039670000077
s3.4, calculating the upper quantile q at the t moment through a Pinball quantile formulathLower quantile (1-q)thAnd 0.5thLoss value L corresponding to quantilet,q(ut,q),Lt,(1-q)(ut,(1-q)) And Lt,0.5(ut,0.5);
In this embodiment, to predict the 90% probability load interval, 90% and 10% Pinball quantiles are provided, as shown in fig. 3, so as to obtain two corresponding load curves, and a middle area of the two load curves is the 90% probability load prediction interval. In addition, the output corresponding to 50% of Pinball quantiles can be increased, when the Pinball quantile is 50%, the output corresponding to the model is a deterministic prediction result, and the residual distribution of deterministic prediction can improve the performance of the probabilistic prediction model. During the model training process, the 50% Pinball quantile is taken as one of the loss functions due to the mechanism of back propagation, and the gradient of the model can be guided to be reduced towards the direction beneficial to increasing the probability prediction performance.
S3.4.1, setting the preset probability of Pinball quantiles as q;
s3.4.2, calculating the upper quantile q at the time tthLower quantile (1-q)thAnd 0.5thQuantile prediction error
Figure BDA0002789039670000081
And
Figure BDA0002789039670000082
ytis the real load value;
s3.4.3, calculating the upper quantile q at the time tthLower quantile (1-q)thAnd 0.5thLoss value L corresponding to quantilet,q(ut,q),Lt,(1-q)(ut,(1-q)) And Lt,0.5(ut,0.5);
Figure BDA0002789039670000083
Figure BDA0002789039670000084
Figure BDA0002789039670000085
S3.5, carrying out weighted average on the loss values;
Figure BDA0002789039670000086
wherein the content of the first and second substances,
Figure BDA0002789039670000087
respectively represent Lt,q(ut,q),Lt,(1-q)(ut,(1-q)) And Lt,0.5(ut,0.5) The weight of (c);
s3.6, repeating the steps S3.1-S3.5, comparing the weighted average loss obtained by each round of training, executing a learning rate automatic attenuation strategy when the weighted average loss does not decrease in the P round of training, and executing an early stop strategy when the weighted average loss does not decrease in the 3 x P round of training, namely automatically stopping the training to obtain a load probability prediction model after the training is finished; otherwise, entering step S3.7;
the purpose of adopting the learning rate automatic attenuation strategy is as follows: when the learning rate is low during model training, the updating speed of the parameters is seriously reduced; when the learning rate is high, the model can vibrate in the gradient descending process, so that the model falls into local optimum;
the purpose of adopting the early-stop strategy is: the more training iterations of the model, the better the fitting effect of the model to a specific sample. However, an excessive number of iterations may result in model overfitting and excessive computational overhead. The early stopping strategy enables the training process to be automatically stopped under the condition that the loss of the model on the verification set is not reduced any more;
s3.7, updating weight parameters in the time domain convolution neural network model in the negative gradient direction of the weighted average loss function by using a batch gradient descent algorithm according to the negative gradient direction, and then returning to the step S3.1;
s4, load probability real-time prediction
Collecting load data of a user in real time and weather and temperature data related to the load data, constructing a frame of standard data according to the method in the step S1, and inputting the frame of standard data into a trained load probability prediction model so as to output a load probability prediction interval
Figure BDA0002789039670000091
And deterministic prediction results
Figure BDA0002789039670000092
In order to accurately and effectively evaluate the probability prediction interval, the PICP and PINAW indexes are used as evaluation criteria, and are respectively defined as follows:
Figure BDA0002789039670000093
Figure BDA0002789039670000094
in the formula: n is the number of check samples, and xi is the predicted target value if the predicted target value is included in the upper and lower limits of interval prediction t1, otherwise ξt=0;
Figure BDA0002789039670000095
Predicting a section for the load probability; r is an actual value change range and is used for carrying out normalization processing on the prediction interval;
table 1 compares the probability prediction results of the method of the present invention with those of other methods in the case of a 95% preset probability Prediction Interval (PINC), from which it can be concluded that, in the case of PICP approximating PINC, the performance of TCN is superior to CNN and LSTM, where the probability prediction interval width (PINAW) of LSTM is larger than TCN although smaller than CNN, thus demonstrating that the method of the present invention has better probability prediction performance. Fig. 4 is a corresponding probability prediction curve.
TABLE 1. probability prediction performance of different models
Figure BDA0002789039670000096
Table 2 compares the probability prediction results of the method of the present invention under different preset probability Prediction Intervals (PINC), and the Pinball quantile function is a non-parameter estimation method, and can conveniently output Prediction Intervals (PI) corresponding to different PINCs by adjusting quantiles. The probability predictions for different PINCs are shown in fig. 5. The width of the PI with 99% of the PINC is the largest, while the width of the PI with the lowest PINC is the smallest. Fig. 5 is a corresponding probability prediction curve.
TABLE 2 Performance of Pinball-TCN under different PINCs
Figure BDA0002789039670000101
Although illustrative embodiments of the present invention have been described above to facilitate the understanding of the present invention by those skilled in the art, it should be understood that the present invention is not limited to the scope of the embodiments, and various changes may be made apparent to those skilled in the art as long as they are within the spirit and scope of the present invention as defined and defined by the appended claims, and all matters of the invention which utilize the inventive concepts are protected.

Claims (2)

1. A user side load probability prediction method based on a time domain convolution neural network is characterized by comprising the following steps:
(1) data acquisition and preprocessing
(1.1) collecting historical load, temperature and weather data
Active power of a user is collected as historical load data X according to a fixed period within a certain time rangeloadSimultaneously collecting weather types X in corresponding time rangesweatherAnd a temperature value XtemperatureThe weather and temperature data is then processed into data samples X ═ X (X) at the same frequency as the historical load dataweather,Xtemperature,Xload);
(1.2) data sample cleaning
Performing data cleaning on the data sample X, replacing an abnormal value in the data sample with a missing value, and then performing proximity interpolation processing at the missing value to obtain a complete data sample X' (X)weather′,Xtemperature′,Xload′);
(1.3), Z-score normalization
And (5) carrying out standardization processing on the complete data sample to obtain a standard data sample X ″ (X ═ X)weather″,Xtemperature″,Xload"), the Z-score normalization formula is as follows:
Figure RE-FDA0002875358890000011
wherein x is a standard value, x is a data sample value before processing, mu is an average value of the data sample, and sigma is a standard deviation of the data sample;
(1.4) addition time characteristics
Taking the time corresponding to the standard data sample, including the week and time characteristics, as a fourth characteristic X in the form of One-hot codingdateAdded to X ", a standard data sample with time signature X" (X)weather″,Xtemperature″,Xload″,Xdate);
(1.5) construction of Standard data set
On X', a standard dataset was constructed using a sliding window method.
(1.5.1) generating the t frame of the standard data set by a sliding window method
A segment of data { X '″ of length n preceding time t is taken at X't-n,x″′t-n+1,...,x″′t-1And taking load data x' at the time t as a training sample of the modelloadAs a training label of the model, the training sample and the training label are correspondingly matched and used as one frame of the standard data set;
(1.5.2) repeating the step (1.5.1), and sequentially sliding backwards for M steps on the X' to construct a standard data set with the length of M frames;
(2) and constructing a load probability prediction model
Constructing a time domain convolution-based neural network as a load probability prediction model, wherein the time domain convolution neural network is formed by serially connecting an input layer, a plurality of residual blocks and an output layer, the output layer is formed by connecting three full-connection layers in parallel, and the number of quantiles corresponds to the number of the full-connection layers;
the ith residual block is provided with two input branches, wherein the left input branch is the input of the output of the (i-1) th residual block to the ith residual block after the computation of the convolutional layer, the weight normalization, the ReLU activation function and the neuron immediate inactivation (dropout) respectively; the right input branch is to directly input the output of the i-1 th residual block to the i-th residual block after 1 × 1 convolution; wherein the convolution layer is a convolution kernel formed by a plurality of dilation causal convolutions;
(3) training load probability prediction model
(3.1) randomly selecting a frame of data from the standard data set, inputting the frame of data into a time domain convolutional neural network model, converting the frame of data into a tensor form through an input layer, and inputting the tensor form into the serially connected residual blocks;
(3.2) in the time domain convolution neural network model, the input tensor of the ith residual block is set as Z(i-1)I ═ 1,2, …, K denotes the number of residual blocks; in the left branch of the ith residual block, the tensor Z(i-1)Performing feature extraction through a convolution kernel formed by a plurality of dilation causal convolutions, and then sequentially performing batch normalization, layer normalization, ReLU activation function and neuron random inactivation to obtain the output tensor of the left branch
Figure RE-FDA0002875358890000021
In the right branch of the i-th residual block, the tensor Z(i-1)Convolving by 1 × 1 to make its output tensor
Figure RE-FDA0002875358890000022
Matching the dimensionality of the output tensors of the left branch, and then adding the output tensors of the two branches to obtain the output of the ith residual block
Figure RE-FDA0002875358890000023
(3.3) repeating the step (3.2) until the last residual block output Z(K)Finally, Z is(K)The outputs of the three full-connection layers connected in parallel are respectively recorded as
Figure RE-FDA0002875358890000024
And
Figure RE-FDA0002875358890000025
(3.4) calculating the upper quantile q at the time t by a Pinball quantile formulathLower quantile (1-q)thAnd 0.5thLoss value L corresponding to quantilet,q(ut,q),Lt,(1-q)(ut,(1-q)) And Lt,0.5(ut,0.5)
(3.4.1) setting the preset probability of the Pinball quantile as q;
(3.4.2) calculating the upper quantile q at time tthLower quantile (1-q)thAnd 0.5thQuantile prediction error
Figure RE-FDA0002875358890000026
And
Figure RE-FDA0002875358890000027
ytis the real load value;
(3.4.3) calculating the upper quantile q at the time tthLower quantile (1-q)thAnd 0.5thLoss value L corresponding to quantilet,q(ut,q),Lt,(1-q)(ut,(1-q)) And Lt,0.5(ut,0.5);
Figure RE-FDA0002875358890000031
Figure RE-FDA0002875358890000032
Figure RE-FDA0002875358890000033
(3.5) carrying out weighted average on the loss values;
Figure RE-FDA0002875358890000034
wherein the content of the first and second substances,
Figure RE-FDA0002875358890000035
respectively represent Lt,q(ut,q),Lt,(1-q)(ut,(1-q)) And Lt,0.5(ut,0.5) The weight of (c).
(3.6) repeating the steps (3.1) - (3.5), comparing the weighted average loss obtained by each round of training, executing a learning rate automatic attenuation strategy when the weighted average loss does not decrease in the P round of training, and executing an early stop strategy when the weighted average loss does not decrease in the 3 x P round of training, namely automatically stopping the training to obtain a trained load probability prediction model; otherwise, entering the step (3.7);
(3.7) updating the weight parameters in the time domain convolution neural network model in the negative gradient direction of the weighted average loss function by using a batch gradient descent algorithm according to the negative gradient direction, and then returning to the step (3.1);
(4) load probability real-time prediction
Collecting load data of a user in real time and weather and temperature data related to the load data, constructing a frame of standard data according to the method in the step (1), and inputting the frame of standard data into a trained load probability prediction model so as to output a load probability prediction interval
Figure RE-FDA0002875358890000036
And deterministic prediction results
Figure RE-FDA0002875358890000037
2. The method for predicting the load probability on the user side based on the time-domain convolutional neural network as claimed in claim 1, wherein the function f(s) of the dilation causal convolution is:
Figure RE-FDA0002875358890000038
wherein the content of the first and second substances,
Figure RE-FDA0002875358890000039
presentation inputA tensor; (i) denotes a convolution filter, i ═ 0,1,2, …, λ -1, λ being the size of the convolution filter; s-d · i denotes the range of each step of convolution over the element s in the input tensor, d being the expansion factor.
CN202011308724.4A 2020-11-20 2020-11-20 User side load probability prediction method based on time domain convolutional neural network Pending CN112396234A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011308724.4A CN112396234A (en) 2020-11-20 2020-11-20 User side load probability prediction method based on time domain convolutional neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011308724.4A CN112396234A (en) 2020-11-20 2020-11-20 User side load probability prediction method based on time domain convolutional neural network

Publications (1)

Publication Number Publication Date
CN112396234A true CN112396234A (en) 2021-02-23

Family

ID=74607611

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011308724.4A Pending CN112396234A (en) 2020-11-20 2020-11-20 User side load probability prediction method based on time domain convolutional neural network

Country Status (1)

Country Link
CN (1) CN112396234A (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113256031A (en) * 2021-06-25 2021-08-13 国网江西省电力有限公司供电服务管理中心 Self-learning optimization method based on resident demand response strategy
CN113919232A (en) * 2021-10-25 2022-01-11 北京航天创智科技有限公司 Photovoltaic power station power prediction method and system based on recurrent neural network
CN115409296A (en) * 2022-11-02 2022-11-29 国网江西省电力有限公司电力科学研究院 Method for predicting net load probability of active power distribution network
CN116008733A (en) * 2023-03-21 2023-04-25 成都信息工程大学 Single-phase grounding fault diagnosis method based on integrated deep neural network
CN116596044A (en) * 2023-07-18 2023-08-15 华能山东发电有限公司众泰电厂 Power generation load prediction model training method and device based on multi-source data
CN117239739A (en) * 2023-11-13 2023-12-15 国网冀北电力有限公司 Method, device and equipment for predicting user side load by knowledge big model
CN117977584A (en) * 2024-04-02 2024-05-03 山东大学 Power load probability prediction method, system, medium, device and program product

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113256031A (en) * 2021-06-25 2021-08-13 国网江西省电力有限公司供电服务管理中心 Self-learning optimization method based on resident demand response strategy
CN113256031B (en) * 2021-06-25 2021-10-26 国网江西省电力有限公司供电服务管理中心 Self-learning optimization method based on resident demand response strategy
CN113919232A (en) * 2021-10-25 2022-01-11 北京航天创智科技有限公司 Photovoltaic power station power prediction method and system based on recurrent neural network
CN115409296A (en) * 2022-11-02 2022-11-29 国网江西省电力有限公司电力科学研究院 Method for predicting net load probability of active power distribution network
CN116008733A (en) * 2023-03-21 2023-04-25 成都信息工程大学 Single-phase grounding fault diagnosis method based on integrated deep neural network
CN116596044A (en) * 2023-07-18 2023-08-15 华能山东发电有限公司众泰电厂 Power generation load prediction model training method and device based on multi-source data
CN116596044B (en) * 2023-07-18 2023-11-07 华能山东泰丰新能源有限公司 Power generation load prediction model training method and device based on multi-source data
CN117239739A (en) * 2023-11-13 2023-12-15 国网冀北电力有限公司 Method, device and equipment for predicting user side load by knowledge big model
CN117239739B (en) * 2023-11-13 2024-02-02 国网冀北电力有限公司 Method, device and equipment for predicting user side load by knowledge big model
CN117977584A (en) * 2024-04-02 2024-05-03 山东大学 Power load probability prediction method, system, medium, device and program product

Similar Documents

Publication Publication Date Title
CN112396234A (en) User side load probability prediction method based on time domain convolutional neural network
CN108009674A (en) Air PM2.5 concentration prediction methods based on CNN and LSTM fused neural networks
CN111210633A (en) Short-term traffic flow prediction method based on deep learning
CN112990556A (en) User power consumption prediction method based on Prophet-LSTM model
Wang et al. A compound framework for wind speed forecasting based on comprehensive feature selection, quantile regression incorporated into convolutional simplified long short-term memory network and residual error correction
CN112350876A (en) Network flow prediction method based on graph neural network
CN110751318B (en) Ultra-short-term power load prediction method based on IPSO-LSTM
CN110705743A (en) New energy consumption electric quantity prediction method based on long-term and short-term memory neural network
CN106909933A (en) A kind of stealing classification Forecasting Methodology of three stages various visual angles Fusion Features
CN113554466B (en) Short-term electricity consumption prediction model construction method, prediction method and device
CN114282443B (en) Residual service life prediction method based on MLP-LSTM supervised joint model
CN111079989A (en) Water supply company water supply amount prediction device based on DWT-PCA-LSTM
CN111461463A (en) Short-term load prediction method, system and equipment based on TCN-BP
CN114169434A (en) Load prediction method
CN115544890A (en) Short-term power load prediction method and system
CN115034485A (en) Wind power interval prediction method and device based on data space
CN114218872A (en) Method for predicting remaining service life based on DBN-LSTM semi-supervised joint model
CN115905857A (en) Non-invasive load decomposition method based on mathematical morphology and improved Transformer
CN116187835A (en) Data-driven-based method and system for estimating theoretical line loss interval of transformer area
CN111785093A (en) Air traffic flow short-term prediction method based on fractal interpolation
CN111930601A (en) Deep learning-based database state comprehensive scoring method and system
Zhang Short-term power load forecasting based on SAPSO-CNN-LSTM model considering autocorrelated errors
Guo et al. Short-Term Water Demand Forecast Based on Deep Neural Network:(029)
CN116632834A (en) Short-term power load prediction method based on SSA-BiGRU-Attention
CN113762591B (en) Short-term electric quantity prediction method and system based on GRU and multi-core SVM countermeasure learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20210223