CN114819352A - Load prediction method based on MIC-TCN-LSTM, storage medium and computer system - Google Patents
Load prediction method based on MIC-TCN-LSTM, storage medium and computer system Download PDFInfo
- Publication number
- CN114819352A CN114819352A CN202210460380.1A CN202210460380A CN114819352A CN 114819352 A CN114819352 A CN 114819352A CN 202210460380 A CN202210460380 A CN 202210460380A CN 114819352 A CN114819352 A CN 114819352A
- Authority
- CN
- China
- Prior art keywords
- data
- lstm
- tcn
- output
- gate
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
- G06Q10/04—Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
- G06Q10/06—Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
- G06Q10/063—Operations research, analysis or management
- G06Q10/0639—Performance analysis of employees; Performance analysis of enterprise or organisation operations
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Business, Economics & Management (AREA)
- Human Resources & Organizations (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Strategic Management (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Economics (AREA)
- Artificial Intelligence (AREA)
- Mathematical Physics (AREA)
- Molecular Biology (AREA)
- Computational Linguistics (AREA)
- Software Systems (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Entrepreneurship & Innovation (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Development Economics (AREA)
- Marketing (AREA)
- Game Theory and Decision Science (AREA)
- Operations Research (AREA)
- Quality & Reliability (AREA)
- Tourism & Hospitality (AREA)
- General Business, Economics & Management (AREA)
- Educational Administration (AREA)
- Evolutionary Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
The invention discloses a load prediction method based on MIC-TCN-LSTM, a storage medium and a computer system, wherein the method comprises the following steps: s1, preprocessing the original data and then carrying out normalization processing; s2, performing correlation analysis on the data by using the maximum correlation coefficient MIC, and providing data with small or irrelevant correlation; s3, dividing a data set formed by the data processed in the step 2 into a training set and a testing set by using a sliding window; s4, constructing a TCN-LSTM model, and sending the training set in the step 3 into the TCN-LSTM model for training; and S5, predicting the test set by using the model trained in the step 4 to finally obtain a prediction result, and outputting the prediction result. The method and the device accurately predict the load in the power grid, and effectively improve the prediction accuracy. The method provided by the invention is simple and easy to realize.
Description
Technical Field
The invention relates to the field of load prediction methods, in particular to a load prediction method based on MIC-TCN-LSTM, a storage medium and a computer system.
Background
MIC-Maximal information coefficient, originated from a paper published in Science in 2011. MIC is used to measure the degree of correlation between two variables X and Y, and the strength of linearity or nonlinearity, and is commonly used for feature selection for machine learning. Compared with other correlation methods, the MIC has higher accuracy and is an excellent calculation method for data correlation. According to the nature of MIC, MIC has universality, fairness and symmetry. 1) Universality: when the sample size is large enough to contain most of the information of the sample, various interesting associations can be captured, and the method is not limited to a specific function type, such as a linear function, an exponential function or a periodic function, or can be balanced to cover all functional relations. Complex relationships between general variables can be modeled not just by a single function, but by superimposed functions. For a function with better universality, the starting points of the different types of association relations should be close. And is close to 1. 2) Fairness: when the sample size is large enough, similar coefficients can be given for the correlation relations of different types of single noise with similar degrees.
A Time Convolutional Network (TCN) is a neural network algorithm proposed in 2018 for analyzing time series data. The TCN introduces one-dimensional convolution, causal convolution, expansion convolution and residual error network, and solves the problem of feature extraction of long-term time sequence information. TCN is an efficient feature extraction and expression algorithm that can accurately model and estimate complex multi-feature, multi-scale nonlinear data. The load prediction algorithm based on the TCN can accurately simulate the complex relation between the load and the relevant characteristics of the load, and high-precision prediction is realized. TCNs represent a further advantage in the time series prediction task compared to traditional Recurrent Neural Network (RNN) based algorithms: 1) TCNs are good at capturing time dependence because convolution can capture local information; 2) the size of the receptive field of the TCN model can be flexibly adjusted to obtain more input characteristics.
The existing power grid load prediction precision is not high, and a method with high prediction precision is urgently needed to solve the problem.
Disclosure of Invention
In order to solve the existing problems, the invention provides a load prediction method, a storage medium and a computer system based on MIC-TCN-LSTM, and the specific scheme is as follows:
a load prediction method, a storage medium and a computer system based on MIC-TCN-LSTM, wherein the method comprises the following steps:
s1, preprocessing the original data and then carrying out normalization processing;
s2, carrying out correlation analysis on the data by utilizing the maximum correlation coefficient MIC, and providing data with small or irrelevant correlation;
s3, dividing a data set formed by the data processed in the step 2 into a training set and a testing set by using a sliding window;
s4, constructing a TCN-LSTM model, and sending the training set in the step 3 into the TCN-LSTM model for training;
and S5, predicting the test set by using the model trained in the step 4 to finally obtain a prediction result, and outputting the prediction result.
Preferably, the method for preprocessing the raw data in step 1 includes: and processing the original data into an ideal format, and filling missing items in the data by using the average values of the previous and next two days.
Preferably, the normalized data is data to be processed limited between 0 and 1, and the normalized data is processed as followsWherein x' is after normalizationData, x min And x max The minimum and maximum values of the load data, respectively.
Preferably, the method for performing correlation analysis on data in step 2 includes: the method for solving the correlation between other data including temperature, humidity, precipitation, date type and load data by using the maximum correlation information number MIC for measuring the correlation degree between two variables X and Y, linear or non-linear intensity and selecting the characteristics between the variables, and discarding unnecessary variables, wherein the basic principle of the maximum correlation information number MIC utilizes the mutual information concept which is described by using the following equationWherein x and y are two linked random variables, p (x, y) is a joint probability density distribution function, the two random variables are quantized into a scatter diagram, then are continuously divided by small squares, and then the falling probability P (X) and P (Y) in each square grid are calculated, thereby calculating the joint probability density distribution
The correlation between the load variable and the other variables is obtained, and the variables having a small correlation with the load are discarded.
Preferably, the TCN-LSTM model constructed in step 4 includes an input layer, a TCN layer, an LSTM layer, and an output layer, where data are sequentially transmitted.
Preferably, the TCN layer comprises a number of residual error networks including an extended causal convolution layer, a weight normalization layer WeightNorm and a regularization layer Dropout, one branch of the residual error networks performing a transformation operation F on the input X, the other branch performing a simple 1 × 1 transformation to maintain consistency of the number of feature maps in parallel with the existing branch, and the output X (h) of the h residual error module may be represented as
X (h) =ReLU(F(X (h-1) )+X (h-1) )
Wherein ReLU represents an activation function as shown in the following formula
Preferably, the calculation process of the LSTM layer includes:
SA1, forgetting gate determines which information will be forgotten according to input state, thereby obtaining output state f (t) of forgetting gate
f (t) =σ(W fa α (t-1) +W fx x (t) +b f )
Where x (t) is the input at the current time, α (t-1) is the cell state at the previous time, and W fa 、W fx Is the forgetting gate weight coefficient; b f Is a forgetting gate bias; σ is a forgetting gate activation function;
SA2, update gate to determine whether the cell state is updated; the updating of the cell state comprises updating the output state i (t) of the gate and the tan h function activation outputThe product of (a) and the product of the last cell state c (t-1) and the forgetting gate f (t) has an output expression of
i (t) =σ(W ia α (t-1) +W ix x (t) +b i )
In the formula, W ia ,W ix ,W ca ,W cx Is to update the weight coefficients of the gate; bi, bc are the bias of the update gate; σ, tanh is the activation function of the update gate; c (t) is the updated cell state, the output gate determines whether to pass the current state to the next time,
o (t) =σ(W oa α (t-1) +W ox x (t) +b o )
α (t) =o (t) ×tanh(c (t) )
where o (t) is the hidden state of the current cell, W oa ,W ox Is the output gate cell state update weight coefficient; b o Is output gate cell state bias; tanh is the output gate activation function; alpha (t) is the hidden state of the neural network at the current moment,
the hidden state of the final output of the SA3, LSTM network will retain all the past output information to calculate its output, whose output expression is:
in the formula, W y Is the weight of the LSTM network; b y Is an offset of the LSTM network;is a predicted value calculated by the LSTM network; g (z) is an activation function.
The invention also discloses a computer readable storage medium, wherein a computer program is stored on the medium, and after the computer program runs, the load prediction method based on the MIC-TCN-LSTM is executed.
The invention also discloses a computer system, which comprises a processor and a storage medium, wherein the storage medium is stored with a computer program, and the processor reads the computer program from the storage medium and runs the computer program to execute the MIC-TCN-LSTM-based load prediction method.
The invention has the beneficial effects that:
the method and the device accurately predict the load in the power grid, and effectively improve the prediction accuracy. The method provided by the invention is simple and easy to realize.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, and it is obvious that the drawings in the following description are some embodiments of the present invention, and those skilled in the art can also obtain other drawings according to the drawings without creative efforts.
FIG. 1 is a flow chart of a method of the present invention;
FIG. 2 is a schematic view of a sliding window;
FIG. 3 is a schematic diagram of the structure of TCN.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, but not all, embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
A MIC-TCN-LSTM based load prediction method, storage medium and computer system, as shown in fig. 1, wherein the method comprises the steps of:
s1, preprocessing the original data and then carrying out normalization processing;
s2, carrying out correlation analysis on the data by utilizing the maximum correlation coefficient MIC, and providing data with small or irrelevant correlation;
s3, dividing a data set formed by the data processed in the step 2 into a training set and a testing set by using a sliding window;
s4, constructing a TCN-LSTM model, and sending the training set in the step 3 into the TCN-LSTM model for training;
s5, predicting the test set by using the TCN-LSTM model trained in the step 4, detecting the prediction accuracy, finally obtaining a prediction result, and outputting the prediction result.
The method for preprocessing the original data in the step 1 comprises the following steps: and processing the original data into an ideal format, and filling missing items in the data by using the average values of the previous and next two days.
Normalizing data limits the data to be processed to be between 0 and 1, and normalizing is used for facilitating the subsequent data processing and ensuring that the convergence is accelerated during the operation of a program. The normalization processing method is as followsWhere x' is the data after normalization, x min And x max The minimum and maximum values of the load data, respectively.
The method for carrying out correlation analysis on the data in the step 2 comprises the following steps: the method for calculating the correlation between other data including temperature, humidity, precipitation, date type and load data by using the maximum correlation information number MIC, discarding unnecessary variables, wherein the maximum correlation information number MIC is used for measuring the correlation degree between two variables X and Y, and the linear or nonlinear strength is used for selecting the characteristics between the variables, the basic principle of the maximum correlation information number MIC is utilized to the mutual information concept, and the mutual information concept uses the following equation to explain the mutual information conceptWherein x and y are two linked random variables, p (x, y) is a joint probability density distribution function, the two random variables are quantized into a scatter diagram, then the scatter diagram is continuously divided by small squares, and then the probabilities P (X) and P (Y) falling into each square grid are calculated, so that the joint probability density distribution is calculated, the larger the data volume is, the better the MIC effect is, and the specific method is as follows:
the correlation between the load variable and the other variables is obtained according to the above method, and the variables having a small correlation with the load are discarded.
The TCN-LSTM model constructed in the step 4 comprises an input layer, a TCN layer, an LSTM layer and an output layer, wherein data are transmitted in sequence.
The TCN layer includes several residual error networks including a dilated causal convolution layer, a weight normalization layer Weightnorm and a regularization layer Dropout, one branch of the residual error networks performing a transformation operation F on the input X, the other branch performing a simple 1X 1 transformation to maintain consistency of the number of feature maps in parallel with the existing branch, the output X (h) of the h residual error module can be represented as
X (h) =ReLU(F(X (h-1) )+X (h-1) )
Wherein ReLU represents an activation function as shown in the following formula
And an LSTM network is added behind the TCN network, so that the accuracy of prediction is further improved.
The LSTM layer is an improvement of the recurrent neural network. The key of the LSTM layer is the cell state, and information in the cell state is updated and deleted through a forgetting gate, an updating gate and an output gate. The calculation process of the LSTM layer comprises the following steps:
SA1, forgetting gate determines which information will be forgotten according to input state, thereby obtaining output state f (t) of forgetting gate
f (t) =σ(W fa α (t-1) +W fx x (t) +b f )
Where x (t) is an input at the current time, α (t-1) is a cell state at the previous time, and Wfa and Wfx are forgetting gate weight coefficients; bf is the forgotten gate bias; σ is a forgetting gate activation function;
SA2, update Gate determining whether cell State is updated(ii) a The updating of the cell state comprises updating the output state i (t) of the gate and activating the output by the tanh functionThe product of (a) and the product of the last-moment cell state c (t-1) and the forgetting gate f (t), and the output expression is
i (t) =σ(W ia α (t-1) +W ix x (t) +b i )
Wherein Wia, Wix, Wca, Wcx are weight coefficients of the update gate; bi, bc are the bias of the update gate; σ, tanh is the activation function of the update gate; c (t) is the updated cell state, the output gate determines whether to pass the current state to the next time,
o (t) =σ(W oa α (t-1) +W ox x (t) +b o )
α (t) =o (t) ×tanh(c (t) )
where o (t) is the hidden state of the current cell, Woa, and Wox is the output gate cell state update weight coefficient; bo is the output gate cell state bias; tanh is the output gate activation function; alpha (t) is the hidden state of the neural network at the current moment,
the hidden state of the final output of the SA3, LSTM network will retain all the past output information to calculate its output, whose output expression is:
where Wy is the weight of the LSTM network; by is the bias of the LSTM network;is a predicted value calculated by the LSTM network; g (z) is an activation function.
The invention also discloses a computer readable storage medium, wherein a computer program is stored on the medium, and after the computer program runs, the load prediction method based on the MIC-TCN-LSTM is executed.
The invention also discloses a computer system, which comprises a processor and a storage medium, wherein the storage medium is stored with a computer program, and the processor reads the computer program from the storage medium and runs the computer program to execute the MIC-TCN-LSTM-based load prediction method. The method and the device accurately predict the load in the power grid, and effectively improve the prediction accuracy. The method provided by the invention is simple and easy to realize.
Those of skill would further appreciate that the various illustrative logical blocks, modules, circuits, and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, computer software, or combinations of both. To clearly illustrate this interchangeability of hardware and software, various illustrative components, blocks, modules, circuits, and steps have been described above generally in terms of their functionality. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the overall system. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
The various illustrative logical blocks, modules, and circuits described in connection with the embodiments disclosed herein may be implemented or performed with a general purpose processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or other programmable logic device, discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein. A general purpose processor may be a microprocessor, but in the alternative, the processor may be any conventional processor, controller, microcontroller, or state machine. A processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration.
The steps of a method or algorithm described in connection with the embodiments disclosed herein may be embodied directly in hardware, in a software module executed by a processor, or in a combination of the two. A software module may reside in RAM memory, flash memory, ROM memory, EPROM memory, EEPROM memory, registers, hard disk, a removable disk, a CD-ROM, or any other form of storage medium known in the art. An exemplary storage medium is coupled to the processor such the processor can read information from, and write information to, the storage medium. In the alternative, the storage medium may be integral to the processor. The processor and the storage medium may reside in an ASIC. The ASIC may reside in a user terminal. In the alternative, the processor and the storage medium may reside as discrete components in a user terminal.
In one or more exemplary embodiments, the functions described may be implemented in hardware, software, firmware, or any combination thereof. If implemented in software as a computer program product, the functions may be stored on or transmitted over as one or more instructions or code on a computer-readable medium. Computer-readable media includes both computer storage media and communication media including any medium that facilitates transfer of a computer program from one place to another. A storage media may be any available media that can be accessed by a computer. By way of example, and not limitation, such computer-readable media can comprise RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium that can be used to carry or store desired program code in the form of instructions or data structures and that can be accessed by a computer. Any connection is properly termed a computer-readable medium. For example, if the software is transmitted from a web site, server, or other remote source using a coaxial cable, fiber optic cable, twisted pair, Digital Subscriber Line (DSL), or wireless technologies such as infrared, radio, and microwave, then the coaxial cable, fiber optic cable, twisted pair, DSL, or wireless technologies such as infrared, radio, and microwave are included in the definition of medium. Disk (disk) and disc (disc), as used herein, includes Compact Disc (CD), laser disc, optical disc, Digital Versatile Disc (DVD), floppy disk and blu-ray disc where disks (disks) usually reproduce data magnetically, while discs (discs) reproduce data optically with lasers. Combinations of the above should also be included within the scope of computer-readable media.
The previous description of the disclosure is provided to enable any person skilled in the art to make or use the disclosure. Various modifications to the disclosure will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other variations without departing from the spirit or scope of the disclosure. Thus, the disclosure is not intended to be limited to the examples and designs described herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.
Although the present invention has been described in detail with reference to the foregoing embodiments, it should be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; and such modifications or substitutions do not depart from the spirit and scope of the corresponding technical solutions of the embodiments of the present invention.
Claims (9)
1. A load prediction method based on MIC-TCN-LSTM is characterized by comprising the following steps:
s1, preprocessing the original data and then carrying out normalization processing;
s2, carrying out correlation analysis on the data by utilizing the maximum correlation coefficient MIC, and providing data with small or irrelevant correlation;
s3, dividing a data set formed by the data processed in the step 2 into a training set and a testing set by using a sliding window;
s4, constructing a TCN-LSTM model, and sending the training set in the step 3 into the TCN-LSTM model for training;
and S5, predicting the test set by using the model trained in the step 4 to finally obtain a prediction result, and outputting the prediction result.
2. The method according to claim 1, wherein the method for preprocessing the raw data in step 1 is as follows: and processing the original data into an ideal format, and filling missing items in the data by using the average values of the previous and next two days.
4. The method of claim 1, wherein the step 2 of performing correlation analysis on the data comprises: the method for solving the correlation between other data including temperature, humidity, precipitation, date type and load data by using the maximum correlation information number MIC for measuring the correlation degree between two variables X and Y, linear or non-linear intensity and selecting the characteristics between the variables, and discarding unnecessary variables, wherein the basic principle of the maximum correlation information number MIC utilizes the mutual information concept which is described by using the following equationWherein x and y are two linked random variables, p (x, y) is a joint probability density distribution function, the two random variables are quantized into a scatter diagram, then are continuously divided by small squares, and the falling probabilities P (X) and P (Y) in each square grid are calculated, so that the joint probability density distribution is calculatedThe volume method is calculated as follows
The correlation between the load variable and the other variables is obtained, and the variables having a small correlation with the load are discarded.
5. The method of claim 1, wherein: the TCN-LSTM model constructed in the step 4 comprises an input layer, a TCN layer, an LSTM layer and an output layer, wherein data are transmitted in sequence.
6. The method of claim 5, wherein: the TCN layer includes several residual error networks including an augmented causal convolution layer, a weight normalization layer Weightnorm and a regularization layer Dropout, one branch of the residual error networks performing a transform operation F on an input X, the other branch to perform a simple 1X 1 transformation to maintain consistency of the number of feature maps in parallel with the existing branch, and the output X (h) of the h-th residual error module can be represented as
X (h) =ReLU(F(X (h-1) )+X (h-1) )
Wherein ReLU represents an activation function as shown in the following formula
7. The method of claim 6, wherein the calculation of the LSTM layer comprises:
SA1, forgetting gate determines which information will be forgotten according to input state, thereby obtaining output state f (t) of forgetting gate
f (t) =σ(W fa α (t-1) +W fx x (t) +b f )
Where x (t) is the input at the current time, α (t-1) is the cell state at the previous time, and W fa 、W fx Is the forgetting gate weight coefficient; b f Is a forgetting gate bias; σ is a forgetting gate activation function;
SA2, update gate to determine whether the cell state is updated; the updating of the cell state comprises updating the output state i (t) of the gate and activating the output by the tanh functionThe product of (a) and the product of the last-moment cell state c (t-1) and the forgetting gate f (t), and the output expression is
i (t) =σ(W ia α (t-1) +W ix x (t) +b i )
In the formula, W ia ,W ix ,W ca ,W cx Is to update the weight coefficients of the gate; bi, bc are the bias of the update gate; σ, tanh is the activation function of the update gate; c (t) is the updated cell state, the output gate determines whether to pass the current state to the next time,
o (t) =σ(W oa α (t-1) +W ox x (t) +b o )
α (t) =o (t) ×tanh(c (t) )
where o (t) is the hidden state of the current cell, W oa ,W ox Is the output gate cell state update weight coefficient; b o Is output gate cell state bias; tanh is the output gate activation function; alpha (t) is the hidden state of the neural network at the current moment,
the hidden state of the final output of the SA3, LSTM network will retain all the past output information to calculate its output, whose output expression is:
8. A computer-readable storage medium characterized by: a medium having stored thereon a computer program which, when executed, performs a MIC-TCN-LSTM based load prediction method as claimed in any one of claims 1 to 7.
9. A computer system, characterized by: comprising a processor, a storage medium having a computer program stored thereon, the processor reading and executing the computer program from the storage medium to perform the MIC-TCN-LSTM based load prediction method according to any one of claims 1 to 7.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210460380.1A CN114819352A (en) | 2022-04-28 | 2022-04-28 | Load prediction method based on MIC-TCN-LSTM, storage medium and computer system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210460380.1A CN114819352A (en) | 2022-04-28 | 2022-04-28 | Load prediction method based on MIC-TCN-LSTM, storage medium and computer system |
Publications (1)
Publication Number | Publication Date |
---|---|
CN114819352A true CN114819352A (en) | 2022-07-29 |
Family
ID=82510240
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210460380.1A Pending CN114819352A (en) | 2022-04-28 | 2022-04-28 | Load prediction method based on MIC-TCN-LSTM, storage medium and computer system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114819352A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115907131A (en) * | 2022-11-16 | 2023-04-04 | 国网宁夏电力有限公司经济技术研究院 | Method and system for building electric heating load prediction model in northern area |
-
2022
- 2022-04-28 CN CN202210460380.1A patent/CN114819352A/en active Pending
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115907131A (en) * | 2022-11-16 | 2023-04-04 | 国网宁夏电力有限公司经济技术研究院 | Method and system for building electric heating load prediction model in northern area |
CN115907131B (en) * | 2022-11-16 | 2024-07-12 | 国网宁夏电力有限公司经济技术研究院 | Method and system for constructing electric heating load prediction model in northern area |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
TWI769754B (en) | Method and device for determining target business model based on privacy protection | |
CN114422381B (en) | Communication network traffic prediction method, system, storage medium and computer equipment | |
CN111563706A (en) | Multivariable logistics freight volume prediction method based on LSTM network | |
CN110210644A (en) | The traffic flow forecasting method integrated based on deep neural network | |
CN111931983B (en) | Precipitation prediction method and system | |
CN116303786B (en) | Block chain financial big data management system based on multidimensional data fusion algorithm | |
CN114331122A (en) | Key person risk level assessment method and related equipment | |
CN114819352A (en) | Load prediction method based on MIC-TCN-LSTM, storage medium and computer system | |
CN113988357A (en) | High-rise building wind-induced response prediction method and device based on deep learning | |
CN115691652A (en) | Causal relationship prediction method based on multi-sample data | |
CN115100466A (en) | Non-invasive load monitoring method, device and medium | |
CN117392564B (en) | River water quality inversion method based on deep learning, electronic equipment and storage medium | |
CN114694379A (en) | Traffic flow prediction method and system based on self-adaptive dynamic graph convolution | |
CN113128612A (en) | Processing method of abnormal value in power data and terminal equipment | |
CN116596915A (en) | Blind image quality evaluation method based on multi-scale characteristics and long-distance dependence | |
CN116740586A (en) | Hail identification method, hail identification device, electronic equipment and computer readable storage medium | |
CN116384223A (en) | Nuclear equipment reliability assessment method and system based on intelligent degradation state identification | |
CN115907775A (en) | Personal credit assessment rating method based on deep learning and application thereof | |
CN113776675B (en) | Multispectral radiation temperature measurement inversion calculation method based on generalized inverse-neural network, computer and storage medium | |
CN115936196A (en) | Monthly rainfall model prediction method based on time sequence convolution network | |
CN112183848B (en) | Power load probability prediction method based on DWT-SVQR integration | |
CN112132310A (en) | Power equipment state estimation method and device based on improved LSTM | |
CN117610696B (en) | Runoff prediction method for crossing data sets by utilizing different attributes | |
CN118172799B (en) | Pedestrian re-identification method, device and storage medium for millimeter wave radar | |
CN110728615B (en) | Steganalysis method based on sequential hypothesis testing, terminal device and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |