CN113887812B - Clustering-based small sample load prediction method, device, equipment and storage medium - Google Patents

Clustering-based small sample load prediction method, device, equipment and storage medium Download PDF

Info

Publication number
CN113887812B
CN113887812B CN202111200796.1A CN202111200796A CN113887812B CN 113887812 B CN113887812 B CN 113887812B CN 202111200796 A CN202111200796 A CN 202111200796A CN 113887812 B CN113887812 B CN 113887812B
Authority
CN
China
Prior art keywords
power load
clustering
data
predicted
discrete wavelet
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111200796.1A
Other languages
Chinese (zh)
Other versions
CN113887812A (en
Inventor
陈东
张海
汪启元
陈致晖
吴辰晔
沈灯鸿
刘之亮
赵晨
张然
王波
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Power Grid Co Ltd
Electric Power Dispatch Control Center of Guangdong Power Grid Co Ltd
Original Assignee
Guangdong Power Grid Co Ltd
Electric Power Dispatch Control Center of Guangdong Power Grid Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Power Grid Co Ltd, Electric Power Dispatch Control Center of Guangdong Power Grid Co Ltd filed Critical Guangdong Power Grid Co Ltd
Priority to CN202111200796.1A priority Critical patent/CN113887812B/en
Publication of CN113887812A publication Critical patent/CN113887812A/en
Application granted granted Critical
Publication of CN113887812B publication Critical patent/CN113887812B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • G06F18/232Non-hierarchical techniques
    • G06F18/2321Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
    • G06F18/23213Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions with fixed number of clusters, e.g. K-means clustering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/06Energy or water supply
    • HELECTRICITY
    • H02GENERATION; CONVERSION OR DISTRIBUTION OF ELECTRIC POWER
    • H02JCIRCUIT ARRANGEMENTS OR SYSTEMS FOR SUPPLYING OR DISTRIBUTING ELECTRIC POWER; SYSTEMS FOR STORING ELECTRIC ENERGY
    • H02J3/00Circuit arrangements for ac mains or ac distribution networks
    • H02J3/003Load forecast, e.g. methods or systems for forecasting future load demand
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y04INFORMATION OR COMMUNICATION TECHNOLOGIES HAVING AN IMPACT ON OTHER TECHNOLOGY AREAS
    • Y04SSYSTEMS INTEGRATING TECHNOLOGIES RELATED TO POWER NETWORK OPERATION, COMMUNICATION OR INFORMATION TECHNOLOGIES FOR IMPROVING THE ELECTRICAL POWER GENERATION, TRANSMISSION, DISTRIBUTION, MANAGEMENT OR USAGE, i.e. SMART GRIDS
    • Y04S10/00Systems supporting electrical power generation, transmission or distribution
    • Y04S10/50Systems or methods supporting the power network operation or management, involving a certain degree of interaction with the load-side end user applications

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • Theoretical Computer Science (AREA)
  • Economics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Human Resources & Organizations (AREA)
  • Strategic Management (AREA)
  • Evolutionary Computation (AREA)
  • Tourism & Hospitality (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Business, Economics & Management (AREA)
  • Artificial Intelligence (AREA)
  • Marketing (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Primary Health Care (AREA)
  • Molecular Biology (AREA)
  • Public Health (AREA)
  • Power Engineering (AREA)
  • Quality & Reliability (AREA)
  • Operations Research (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Game Theory and Decision Science (AREA)
  • Water Supply & Treatment (AREA)
  • Computing Systems (AREA)
  • Development Economics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Probability & Statistics with Applications (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The application discloses a small sample load prediction method, a device, equipment and a storage medium based on clustering, wherein the prediction method comprises the following steps: extracting characteristics of the historical power load and the power load to be predicted to obtain a characteristic vector; integrating and clustering the historical power load and the power load to be predicted according to the obtained feature vector to obtain a clustering result; adopting a wavelet noise reduction algorithm to reduce noise of the clustering result, and carrying out averaging treatment on the noise-reduced data to obtain time sequence data with preset length; inputting time sequence data with preset length into a second-order long-short-term memory neural network to obtain a prediction result of the power load; the second-order long-short-term memory neural network is trained through data of historical power loads and power loads to be predicted. The method still has excellent prediction performance under the condition that the power load to be predicted is scarce.

Description

Clustering-based small sample load prediction method, device, equipment and storage medium
Technical Field
The present disclosure relates to the field of power load prediction technologies, and in particular, to a clustering-based small sample load prediction method, apparatus, device, and storage medium.
Background
A method of predicting time-series data (hereinafter, referred to as time-series data) has been widely used in the electric power market. Traditional statistical models have advantages in terms of interpretability but are less performing in terms of prediction accuracy, etc. In recent years, algorithms based on machine learning have gained a very high degree of attention in the field of time series prediction by virtue of higher prediction accuracy.
Another closely related technique is small sample learning. Generally, a large-scale training set is essential for deep learning, and a small sample learning purpose is to improve the existing deep learning algorithm so as to achieve an excellent training effect on the small-scale training set.
However, while deep learning based models have higher accuracy in time series data prediction, these models are less efficient for data utilization. When the training set is small, deep learning-based models often fail to accurately predict future trends in the time series data.
Meanwhile, when a new user joins the power grid, the traditional prediction model cannot effectively obtain priori knowledge capable of supporting load prediction of the new user from historical data of other users so as to improve prediction accuracy.
Disclosure of Invention
The application provides a clustering-based small sample load prediction method, device, equipment and storage medium, which are used for solving the problem of low prediction result precision when a training set of power load is smaller in the prior art.
In order to solve the above technical problems, the present application proposes a small sample load prediction method based on clustering, including: extracting characteristics of the historical power load and the power load to be predicted to obtain a characteristic vector; integrating and clustering the historical power load and the power load to be predicted according to the obtained feature vector to obtain a clustering result; adopting a wavelet noise reduction algorithm to reduce noise of the clustering result, and carrying out averaging treatment on the noise-reduced data to obtain time sequence data with preset length; inputting time sequence data with preset length into a second-order long-short-term memory neural network to obtain a prediction result of the power load; the second-order long-short-term memory neural network is trained through data of historical power loads and power loads to be predicted. Optionally, feature extraction is performed on the historical power load and the power load to be predicted to obtain feature vectors, including: discrete wavelet analysis and statistical analysis based on seasonality, skewness and sample entropy are carried out on the historical power load and the power load to be predicted, so that feature vectors are extracted.
Optionally, denoising the clustering result by adopting a wavelet denoising algorithm, including: setting the discrete wavelet transformation order layer number; wherein the number of levels of the discrete wavelet transform is set to a maximum value that is not affected by the edge effect; and denoising the clustering result by using discrete wavelet transform.
Optionally, the historical power load and the power load to be predicted are clustered in an integrated manner according to the obtained feature vector, so as to obtain a clustering result, including: and obtaining a clustering result based on the K-means algorithm, hierarchical clustering, neighbor propagation algorithm and Gaussian mixture model.
In order to solve the above technical problem, the present application proposes a small sample load prediction device based on clustering, including: the characteristic extraction module is used for extracting characteristics of the historical power load and the power load to be predicted to obtain a characteristic vector; the clustering integration module is used for integrating and clustering the historical power load and the power load to be predicted according to the obtained feature vector to obtain a clustering result; the noise reduction processing module is used for reducing noise of the clustering result by adopting a wavelet noise reduction algorithm, and carrying out averaging processing on the noise-reduced data to obtain time sequence data with preset length; the prediction result module is used for inputting time sequence data with preset length into the second-order long-short-term memory neural network to obtain a prediction result of the power load; the second-order long-short-term memory neural network is trained through data of historical power loads and power loads to be predicted. Optionally, the feature extraction module is further configured to: discrete wavelet analysis and statistical analysis based on seasonality, skewness and sample entropy are carried out on the historical power load and the power load to be predicted, so that feature vectors are extracted.
Optionally, the noise reduction processing module is further configured to: setting the discrete wavelet transformation order layer number; wherein the number of levels of the discrete wavelet transform is set to a maximum value that is not affected by the edge effect; and denoising the clustering result by using discrete wavelet transform.
Optionally, the cluster integration module is further configured to: and obtaining a clustering result based on the K-means algorithm, hierarchical clustering, neighbor propagation algorithm and Gaussian mixture model.
In order to solve the above technical problems, the present application proposes an electronic device, including a memory and a processor, where the memory is connected to the processor, and the memory stores a computer program, and the computer program implements the cluster-based small sample load prediction method when executed by the processor.
To solve the above technical problem, the present application proposes a computer readable storage medium storing a computer program, which when executed implements the clustering-based small sample load prediction method described above.
The application provides a clustering-based small sample load prediction method, a device, equipment and a storage medium, wherein a stable clustering result can be obtained by acquiring a feature vector and integrating and clustering a historical power load and a power load to be predicted according to the acquired feature vector; the wavelet denoising algorithm is adopted to denoise the clustering result, and the denoised data is subjected to the averaging treatment, so that the prediction accuracy can be remarkably improved; inputting time sequence data with preset length into a second-order long-short-term memory neural network to obtain a prediction result of the power load; the second-order long-short-term memory neural network is trained by historical power load and data of power load to be predicted, so that the prediction of a small sample is reliable. According to the method, the historical power load is used for mixing the power load to be predicted, the feature extraction and the integrated clustering result are combined to provide priori knowledge for the second-order LSTM, and then the model is finely adjusted by using a small amount of load data of the power load to be predicted, so that excellent prediction performance is still achieved under the condition that the power load to be predicted is scarce.
Drawings
In order to more clearly illustrate the technical solutions of the present application, the drawings that are needed in the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present application, and that other drawings can be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a flow diagram of one embodiment of a cluster-based small sample load prediction method of the present application;
FIG. 2 is a flow chart of an embodiment of feature vector extraction according to the present application;
FIG. 3 is a schematic diagram of one embodiment of an integrated cluster architecture of the present application;
FIG. 4 is a schematic diagram of one embodiment of a second-order long-term memory neural network of the present application;
FIG. 5 is a graph showing the impact of data granularity, training set sample number and time series period T prediction error;
FIG. 6 is a schematic structural diagram of an embodiment of a cluster-based small sample load prediction device of the present application;
FIG. 7 is a schematic diagram of an embodiment of an electronic device of the present application;
FIG. 8 is a schematic diagram of an embodiment of a computer-readable storage medium of the present application.
Detailed Description
In order to better understand the technical solutions of the present application, the clustering-based small sample load prediction method, apparatus, device and storage medium provided in the present application are described in further detail below with reference to the accompanying drawings and detailed description.
The technical problem to be solved by the application mainly comprises two aspects:
the first problem to be solved is to obtain a priori knowledge available for deep learning models in unlabeled historical data of grid users. The classical method of obtaining a priori knowledge from unlabeled data is cluster analysis. Although this is intuitive, how to perform data reduction and pattern discovery on high-dimensional time series data through feature extraction, and providing a more stable clustering result is a problem worthy of optimization. According to the method, discrete wavelet analysis and statistical analysis are carried out on the power load data, a comprehensive feature extraction template is provided for the problem of high-dimensional unlabeled time sequence data clustering, and a stable solution is provided for clustering analysis through the result of integrating a plurality of clusters. This helps to improve the compactness and interpretability of the cluster analysis.
The second problem to be solved is to combine historical load data with new user load data for small sample load prediction. The method carries out noise reduction based on wavelet decomposition on the historical data clustering result after averaging to obtain representative prototype time sequence data. The prototype time sequence data is used for providing priori knowledge for the algorithm to train a basic model, and fine tuning is carried out on the basic model through new user load data. This can significantly improve the accuracy of the deep learning algorithm in time sequential data prediction.
Referring to fig. 1, fig. 1 is a schematic flow chart of an embodiment of a small sample load prediction method based on clustering, in this implementation, the small sample load prediction method based on clustering may include:
s110: and extracting characteristics of the historical power load and the power load to be predicted to obtain characteristic vectors.
In this step, the user power load rule and characteristics need to be analyzed: discrete wavelet analysis and statistical analysis based on seasonality, skewness and sample entropy are carried out on the historical power load and the power load to be predicted, so that feature vectors are extracted. Referring to fig. 2, fig. 2 is a flowchart illustrating an embodiment of feature vector extraction.
When discrete wavelet analysis is performed on time series data, a part corresponding to a new user data timestamp in historical data is decomposed into an L-layer wavelet coefficient balance tree by utilizing discrete wavelet packet decomposition. The total energy of each layer of the balanced tree is defined as E. For the j-th layer contain N j Wavelet coefficient set d of number j Its discrete wavelet energy DWE can be expressed as:
Figure GDA0004275547030000041
the DWE was base 10 log to remove the correlation between discrete wavelet energies of each layer:
Figure GDA0004275547030000042
on the basis, discrete cosine transform DCT is further carried out on LWE coefficients of each layer in the balance tree to obtain cepstral coefficients WCC of the original time sequence data:
Figure GDA0004275547030000043
and performing dimension reduction integration on the three coefficients of DWE, LWE and WCC through principal component analysis PCA to obtain a new low-dimensional feature space. In addition, the original time sequence is subjected to statistical analysis on the basis of the characteristic space based on discrete wavelet analysis. First, the time series is listed as X by STL seasonal decomposition t Decomposition into seasonal S t Trend T t Residual E t Three additive components and define two indicators describing the periodicity and the trending of the time series, respectively:
Figure GDA0004275547030000044
Figure GDA0004275547030000045
secondly, in order to measure the non-gaussian property of the random process of the time sequence data, namely the thick tail characteristic of the random probability distribution, the skewness skew of the original time sequence data is calculated:
Figure GDA0004275547030000046
the non-linear characteristics of the time series data are measured using sample entropy. For time series data X t ={x 1 ,x 2 ,..x N Dividing the original sequence into N-m+1 segment sub-sequences with a window size of length m:
X m (i)={x i ,x i+1 ,…,x i+m-1 },1≤i≤N-m+1;
for two different subsequences with sequence numbers i and j, the distance between the two is calculated as:
Figure GDA0004275547030000047
a threshold r is designated, all the sub-sequence pairs divided by windows with the length of m and with the interval smaller than r are counted and recorded as N m The method comprises the steps of carrying out a first treatment on the surface of the Counting the number of pairs of all subsequences cut by a window with length of m+1 and with a spacing less than r, and marking the number as N m+1 . Then for a finite number N, the temporal data sample entropy is expressed as:
Figure GDA0004275547030000051
finally, to measure the long-term correlation of the time series, the autocorrelation of the time series is measured using a non-linear hurst index. For time sequence data after Gaussian regularization
Figure GDA0004275547030000052
Calculate X' t Is the cumulative sum sequence Y of (2) t Wherein the ith cumulative sum is denoted +.>
Figure GDA0004275547030000053
Thus, the hurst index can be expressed as:
Figure GDA0004275547030000054
s120: and integrating and clustering the historical power load and the power load to be predicted according to the obtained feature vector to obtain a clustering result.
After the feature space with lower user history load data dimension is obtained, a widely used clustering device comprising K-means, hierarchical clustering, neighbor propagation clustering, gaussian mixture model and the like is used for obtaining a stable clustering result, and the result of the clustering device is integrated into a stable clustering solution through a similarity partitioning algorithm CSPA based on clustering. Referring to fig. 3, fig. 3 is a schematic diagram illustrating an embodiment of an integrated cluster structure according to the present application.
Unsupervised clustering provides prototype data for the predictive model to learn a priori knowledge of the predictive task. The algorithm adopts an integrated clustering mode as a clustering mode, and learns and integrates clustering results of an original data set based on a K-means algorithm, hierarchical clustering, a neighbor propagation algorithm and a Gaussian mixture model to obtain data division capable of better reflecting the internal structure of the data set. The superiority of the integrated clustering on the prediction algorithm is confirmed by designing a demonstration experiment and using the prediction accuracy as an index to compare the integrated clustering with the independent clustering methods.
S130: and adopting a wavelet denoising algorithm to denoise the clustering result, and carrying out averaging treatment on the denoised data to obtain time sequence data with preset length.
Integrated clustering utilizes limited data to predict targetHistorical data marked with the same time series data characteristics are aggregated. In order to generate a priori knowledge available to the predictive model, a prototype timing data generator (hereinafter referred to as a generator) receives a plurality of pieces of historical timing data corresponding to the clustering result, and converts the historical timing data into prototype timing data available for the predictive model to learn through averaging and wavelet noise reduction. The generator first denoises the clustering result using a wavelet denoising algorithm. Specifically, the number of levels of the discrete wavelet transform is set to a maximum value that is not affected by the edge effect. The signal length is denoted as l x The filter length is denoted as l f The discrete wavelet transform order layer number can be expressed as:
Figure GDA0004275547030000055
the discrete wavelet transformed time series data will be passed through a hard threshold function to achieve noise reduction, where the threshold is denoted as T, the input signal is denoted as x, and the output signal is denoted as ρ T (x):
Figure GDA0004275547030000061
The plurality of pieces of historical time sequence data subjected to discrete wavelet noise reduction are further averaged, and a generator generates prototype time sequence data X with a specific length c
S140: inputting time sequence data with preset length into a second-order long-short-term memory neural network to obtain a prediction result of the power load; the second-order long-short-term memory neural network is trained through data of historical power loads and power loads to be predicted.
In order to fully utilize the prior knowledge collected by the feature extraction and the integrated clustering to train a prediction model, the traditional long-short-period memory neural network is expanded into a second-order long-short-period memory neural network (hereinafter referred to as a second-order LSTM). Referring to fig. 4, fig. 4 is a schematic diagram of an embodiment of a second-order long-term memory neural network according to the present application.
The second-order long-short-term memory neural network comprises a first-order LSTM and a second-order LSTM, wherein the historical power load is used for performing basic class training on the first-order LSTM, and the power load to be predicted is used for performing model fine tuning on the second-order LSTM, so that the trained second-order long-short-term memory neural network capable of predicting data is obtained.
According to the above steps, the time series data after the wavelet decomposition and noise reduction is further obtained from the historical power load and the power load to be predicted, and the true value X is obtained in one step test Finally, the prediction result X can be obtained by combining pred
Specifically, according to the integrated clustering result, historical data X of a plurality of users belonging to the same class as the predicted target is used for S 1,X S 2,…X S n and a small amount of training data of the predicted target
Figure GDA0004275547030000062
The second order LSTM will train based on both, respectively. Stage one, second order LSTM utilization generator for X S 1,X S 2,…X S n generated prototype timing data X c Training, and marking updated network weight as theta 0 Based on the ability to adapt the neural network quickly to new prediction tasks. In the second stage, the second-order LSTM utilizes a small amount of training data with limited prediction targets to finely adjust the network weight, and the updated network weight is recorded as theta 1 The trained second-order neural network has the future load data of the predicted target +.>
Figure GDA0004275547030000063
Is used for the prediction ability of the model (C).
RMSE is a commonly used accuracy index for the predicted result of time series data, if the true value of time series data is denoted as p x The predicted value is recorded as
Figure GDA0004275547030000064
The RMSE may be written in the form:
Figure GDA0004275547030000065
to measure the comprehensive performance of the prediction algorithm on multiple pieces of time sequence data, the average MRMSE of RMSE results based on different pieces of time sequence data is used as an index to represent the prediction accuracy:
Figure GDA0004275547030000066
where M represents the number of timing data used to calculate the mean of the RMSE, and n represents the length of a single timing data. Fig. 5 shows the effect of data granularity M, training set sample number N and time series period T on prediction error.
In time series data prediction, granularity and periodicity of training data have significant influence on model prediction accuracy. Intuitively, the model performs better with larger granularity and provided training data covering one or more cycles of the time series. And performing verification test by using sine time sequence data with additional Gaussian noise, and representing the prediction accuracy by using MRMSE as an index. Under the condition of controlling other variables, the prediction error is inversely proportional to the data granularity M.
MRMSE∝O(N -1 );
The number of data is measured by the number of samples of the training set of the prediction model, T/M represents the ratio of period to granularity, and when N=PT/M and P is any positive integer, the prediction error reaches the local minimum lower bound.
In summary, the embodiment provides a feature extraction template based on wavelet packet decomposition and statistical analysis for the problem of non-label long-term time series data cluster analysis, and provides a stable solving method for cluster analysis by an integrated clustering method; in order to obtain priori knowledge with available deep learning model and strong universality from the historical time sequence data, the clustering result of the historical time sequence data is averaged and wavelet decomposition noise reduction based on a hard threshold value is carried out; and meanwhile, the deep learning model trained through priori knowledge is finely adjusted by using small sample data with unknown labels, so that a prediction model which is customized for the small sample data and has high accuracy is obtained. The present embodiment proposes the minimum lower bound of the small sample data amount required to achieve the optimal model prediction accuracy at a specific granularity by studying the relationship between the time series data granularity and the small sample data amount.
Based on the above-mentioned small sample load prediction method based on clustering, the present application further provides a small sample load prediction device based on clustering, please refer to fig. 6, fig. 6 is a schematic structural diagram of an embodiment of the small sample load prediction device based on clustering, in this embodiment, the small sample load prediction device 100 based on clustering may include a feature extraction module 110, a cluster integration module 120, a noise reduction processing module 130, and a prediction result module 140.
Specifically, the feature extraction module 110 is configured to perform feature extraction on the historical power load and the power load to be predicted, so as to obtain a feature vector.
The clustering integration module 120 is configured to integrate and cluster the historical power load and the power load to be predicted according to the obtained feature vector, and obtain a clustering result.
The noise reduction processing module 130 is configured to reduce noise of the clustering result by using a wavelet noise reduction algorithm, and perform a mean processing on the noise reduced data to obtain time sequence data with a preset length.
The prediction result module 140 is configured to input time sequence data with a preset length into the second-order long-short-term memory neural network to obtain a prediction result of the power load; the second-order long-short-term memory neural network is trained through data of historical power loads and power loads to be predicted.
Optionally, the feature extraction module 110 is further configured to: discrete wavelet analysis and statistical analysis based on seasonality, skewness and sample entropy are carried out on the historical power load and the power load to be predicted, so that feature vectors are extracted.
Optionally, the noise reduction processing module 130 is further configured to: setting the discrete wavelet transformation order layer number; wherein the number of levels of the discrete wavelet transform is set to a maximum value that is not affected by the edge effect; and denoising the clustering result by using discrete wavelet transform.
Optionally, the cluster integration module 120 is further configured to: and obtaining a clustering result based on the K-means algorithm, hierarchical clustering, neighbor propagation algorithm and Gaussian mixture model.
Based on the clustering-based small sample load prediction method, the application also provides an electronic device, as shown in fig. 7, and fig. 7 is a schematic structural diagram of an embodiment of the electronic device. The electronic device 200 may comprise a memory 21 and a processor 22, the memory 21 being connected to the processor 22, the memory 21 having stored therein a computer program which, when executed by the processor 22, implements the method of any of the embodiments described above. The steps and principles of the method are described in detail in the above method, and are not described in detail herein.
In the present embodiment, the processor 22 may also be referred to as a CPU (central processing unit ). The processor 22 may be an integrated circuit chip having signal processing capabilities. Processor 22 may also be a general purpose processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or other programmable logic device, discrete gate or transistor logic, discrete hardware components. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
Based on the clustering-based small sample load prediction method, the application also provides a computer-readable storage medium. Referring to fig. 8, fig. 8 is a schematic structural diagram of an embodiment of a computer readable storage medium of the present application. The computer readable storage medium 300 has stored thereon a computer program 31 which, when executed by a processor, implements the method of any of the embodiments described above. The steps and principles of the method are described in detail in the above method, and are not described in detail herein.
Further, the computer readable storage medium 300 may be a usb disk, a removable hard disk, a read-only memory (ROM), a random access memory (random access memory, RAM), a magnetic tape, or a compact disc, etc. which may store the program code.
The ability to quickly adapt to time series prediction tasks in small sample scenarios with limited data volume is an important challenge for power load prediction and other practical applications. The application provides a small sample time sequence prediction method based on second-order LSTM. According to the method, the existing power load record is utilized through integrated clustering, so that the forecasting task of the time sequence with unknown labels and sparse data size is effectively solved. Numerous studies have shown that this approach can be significantly better than its baseline on both major electrical load data sets. In addition, the performance of FSL-LSTM is demonstrated in terms of both small sample data size and data granularity.
It is to be understood that the specific embodiments described herein are for purposes of illustration only and are not limiting. Further, for ease of description, only some, but not all, of the structures associated with this application are shown in the drawings. The step numbers used herein are also for convenience of description only, and are not limiting as to the order in which the steps are performed. All other embodiments, which can be made by one of ordinary skill in the art without undue burden from the present disclosure, are within the scope of the present disclosure.
The terms "first," "second," and the like in this application are used for distinguishing between different objects and not for describing a particular sequential order. Furthermore, the terms "comprise" and "have," as well as any variations thereof, are intended to cover a non-exclusive inclusion. For example, a process, method, system, article, or apparatus that comprises a list of steps or elements is not limited to only those listed steps or elements but may include other steps or elements not listed or inherent to such process, method, article, or apparatus.
Reference herein to "an embodiment" means that a particular feature, structure, or characteristic described in connection with the embodiment may be included in at least one embodiment of the present application. The appearances of such phrases in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. Those of skill in the art will explicitly and implicitly appreciate that the embodiments described herein may be combined with other embodiments.
The foregoing description is only of embodiments of the present application, and is not intended to limit the scope of the patent application, and all equivalent structures or equivalent processes using the descriptions and the contents of the present application or other related technical fields are included in the scope of the patent application.

Claims (8)

1. A cluster-based small sample load prediction method, comprising:
extracting characteristics of the historical power load and the power load to be predicted to obtain a characteristic vector;
the feature extraction of the historical power load and the power load to be predicted is carried out to obtain a feature vector, and the feature vector comprises:
performing discrete wavelet analysis and statistical analysis based on seasonality, skewness and sample entropy on the historical power load and the power load to be predicted so as to extract a feature vector;
when discrete wavelet analysis is carried out on time series data, discrete wavelet packet decomposition is utilized to decompose a part corresponding to a new user data timestamp in historical data into L layers of wavelet coefficient balance; defining the total energy of each layer of the balance tree as E; for the j-th layer contain N j Wavelet coefficient set d of number j Its discrete wavelet energy DWE can be expressed as:
Figure FDA0004275547020000011
the DWE was base 10 log to remove the correlation between discrete wavelet energies of each layer:
Figure FDA0004275547020000012
on the basis, discrete cosine transform DCT is further carried out on LWE coefficients of each layer in the balance tree to obtain cepstral coefficients WCC of the original time sequence data:
Figure FDA0004275547020000013
performing dimension reduction integration on three coefficients of DWE, LWE and WCC through principal component analysis PCA to obtain a new low-dimension feature space;
according to the obtained feature vector, integrating and clustering the historical power load and the power load to be predicted to obtain a clustering result;
adopting a wavelet noise reduction algorithm to reduce noise of the clustering result, and carrying out averaging treatment on the noise-reduced data to obtain time sequence data with preset length;
inputting the time sequence data with the preset length into a second-order long-short-term memory neural network to obtain a prediction result of the power load; wherein the second-order long-short-term memory neural network is trained via the historical power load and the data of the power load to be predicted.
2. The cluster-based small sample load prediction method according to claim 1, wherein the denoising the cluster result using a wavelet denoising algorithm comprises:
setting the discrete wavelet transformation order layer number; wherein the number of levels of the discrete wavelet transform is set to a maximum value that is not affected by the edge effect;
and denoising the clustering result by using the discrete wavelet transform.
3. The clustering-based small sample load prediction method according to claim 2, wherein the integrating the historical power load and the power load to be predicted according to the feature vector to obtain a clustering result comprises:
and obtaining the clustering result based on a K-means algorithm, hierarchical clustering, a neighbor propagation algorithm and a Gaussian mixture model.
4. A cluster-based small sample load prediction apparatus, comprising:
the characteristic extraction module is used for extracting characteristics of the historical power load and the power load to be predicted to obtain a characteristic vector; the method is also used for carrying out discrete wavelet analysis and statistical analysis based on seasonality, skewness and sample entropy on the historical power load and the power load to be predicted so as to extract a feature vector;
specifically, discrete wavelet analysis and statistical analysis based on seasonality, skewness and sample entropy are carried out on the historical power load and the power load to be predicted so as to extract a feature vector;
when discrete wavelet analysis is carried out on time series data, discrete wavelet packet decomposition is utilized to decompose a part corresponding to a new user data timestamp in historical data into L layers of wavelet coefficient balance; defining the total energy of each layer of the balance tree as E; for the j-th layer contain N j Wavelet coefficient set d of number j Its discrete wavelet energy DWE can be expressed as:
Figure FDA0004275547020000021
the DWE was base 10 log to remove the correlation between discrete wavelet energies of each layer:
Figure FDA0004275547020000022
on the basis, discrete cosine transform DCT is further carried out on LWE coefficients of each layer in the balance tree to obtain cepstral coefficients WCC of the original time sequence data:
Figure FDA0004275547020000023
performing dimension reduction integration on three coefficients of DWE, LWE and WCC through principal component analysis PCA to obtain a new low-dimension feature space;
the clustering integration module is used for integrating and clustering the historical power load and the power load to be predicted according to the obtained feature vector to obtain a clustering result;
the noise reduction processing module is used for reducing noise of the clustering result by adopting a wavelet noise reduction algorithm, and carrying out averaging processing on the noise-reduced data to obtain time sequence data with preset length;
the prediction result module is used for inputting the time sequence data with the preset length into a second-order long-short-term memory neural network to obtain a prediction result of the power load; wherein the second-order long-short-term memory neural network is trained via the historical power load and the data of the power load to be predicted.
5. The cluster-based small sample load prediction apparatus of claim 4, wherein the noise reduction processing module is further configured to:
setting the discrete wavelet transformation order layer number; wherein the number of levels of the discrete wavelet transform is set to a maximum value that is not affected by the edge effect;
and denoising the clustering result by using the discrete wavelet transform.
6. The cluster-based small sample load prediction apparatus of claim 4, wherein the cluster integration module is further configured to:
and obtaining the clustering result based on a K-means algorithm, hierarchical clustering, a neighbor propagation algorithm and a Gaussian mixture model.
7. An electronic device comprising a memory and a processor, the memory being coupled to the processor, the memory storing a computer program that, when executed by the processor, implements the cluster-based small sample load prediction method of any of claims 1-3.
8. A computer readable storage medium, characterized in that a computer program is stored, which computer program, when executed, implements the cluster-based small sample load prediction method of any of claims 1-3.
CN202111200796.1A 2021-10-14 2021-10-14 Clustering-based small sample load prediction method, device, equipment and storage medium Active CN113887812B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111200796.1A CN113887812B (en) 2021-10-14 2021-10-14 Clustering-based small sample load prediction method, device, equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111200796.1A CN113887812B (en) 2021-10-14 2021-10-14 Clustering-based small sample load prediction method, device, equipment and storage medium

Publications (2)

Publication Number Publication Date
CN113887812A CN113887812A (en) 2022-01-04
CN113887812B true CN113887812B (en) 2023-07-07

Family

ID=79002953

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111200796.1A Active CN113887812B (en) 2021-10-14 2021-10-14 Clustering-based small sample load prediction method, device, equipment and storage medium

Country Status (1)

Country Link
CN (1) CN113887812B (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017004880A1 (en) * 2015-07-08 2017-01-12 中兴通讯股份有限公司 Method, device for behavior recognition and computer storage medium
CN108808657A (en) * 2018-05-30 2018-11-13 上海工程技术大学 A kind of Short-Term Load Forecasting of Electric Power System
CN110019543A (en) * 2017-09-12 2019-07-16 中兴通讯股份有限公司 A kind of method and device of Time Series Clustering
CN110650058A (en) * 2019-10-08 2020-01-03 河南省云安大数据安全防护产业技术研究院有限公司 Network traffic analysis method, device, storage medium and equipment
CN111079989A (en) * 2019-11-29 2020-04-28 武汉理工大学 Water supply company water supply amount prediction device based on DWT-PCA-LSTM
CN111461058A (en) * 2020-04-17 2020-07-28 福州大学 Diagnosis method and diagnosis system for power electronic converter fault
CN112561156A (en) * 2020-12-11 2021-03-26 国网江苏省电力有限公司南通供电分公司 Short-term power load prediction method based on user load mode classification

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7519488B2 (en) * 2004-05-28 2009-04-14 Lawrence Livermore National Security, Llc Signal processing method and system for noise removal and signal extraction

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017004880A1 (en) * 2015-07-08 2017-01-12 中兴通讯股份有限公司 Method, device for behavior recognition and computer storage medium
CN110019543A (en) * 2017-09-12 2019-07-16 中兴通讯股份有限公司 A kind of method and device of Time Series Clustering
CN108808657A (en) * 2018-05-30 2018-11-13 上海工程技术大学 A kind of Short-Term Load Forecasting of Electric Power System
CN110650058A (en) * 2019-10-08 2020-01-03 河南省云安大数据安全防护产业技术研究院有限公司 Network traffic analysis method, device, storage medium and equipment
CN111079989A (en) * 2019-11-29 2020-04-28 武汉理工大学 Water supply company water supply amount prediction device based on DWT-PCA-LSTM
CN111461058A (en) * 2020-04-17 2020-07-28 福州大学 Diagnosis method and diagnosis system for power electronic converter fault
CN112561156A (en) * 2020-12-11 2021-03-26 国网江苏省电力有限公司南通供电分公司 Short-term power load prediction method based on user load mode classification

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
基于小波变换的时间序列相似模式匹配;张海勤,蔡庆生;计算机学报(03);全文 *
基于能量谱分析的柴油机故障诊断;门吉芳;潘宏侠;;起重运输机械(09);全文 *

Also Published As

Publication number Publication date
CN113887812A (en) 2022-01-04

Similar Documents

Publication Publication Date Title
CN108734216A (en) Classification of power customers method, apparatus and storage medium based on load curve form
CN108770002B (en) Base station flow analysis method, device, equipment and storage medium
Fang et al. Graph regularised sparse NMF factorisation for imagery de‐noising
Lian et al. Identification of partially linear structure in additive models with an application to gene expression prediction from sequences
Cong et al. Fast and effective model order selection method to determine the number of sources in a linear transformation model
Tatinati et al. Hybrid method based on random convolution nodes for short-term wind speed forecasting
CN113887812B (en) Clustering-based small sample load prediction method, device, equipment and storage medium
CN113470799A (en) Intelligent editor of hospital comprehensive quality supervision platform
Feng et al. Spatiotemporal prediction based on feature classification for multivariate floating-point time series lossy compression
Parker et al. Nonlinear time series classification using bispectrum‐based deep convolutional neural networks
Ding et al. Time-varying Gaussian Markov random fields learning for multivariate time series clustering
CN112463964B (en) Text classification and model training method, device, equipment and storage medium
CN111008281B (en) Text classification method and device, computer equipment and storage medium
Devijver Model-based regression clustering for high-dimensional data. Application to functional data
Kumar et al. Novel affective features for multiscale prediction of emotion in music
Madsen et al. Modeling temporal structure in music for emotion prediction using pairwise comparisons
Jiang et al. A Bayesian Markov-switching model for sparse dynamic network estimation
Ding et al. Analysis of multistate autoregressive models
CN111242043A (en) Time series big data sparsification method and system
Wang et al. Stock price forecasting based on wavelet filtering and ensembled machine learning model
Xiao et al. Branch error reduction criterion-based signal recursive decomposition and its application to wind power generation forecasting
Christoforos et al. On time series representations for multi-label NILM
CN118171783B (en) Energy consumption prediction model training method, prediction method, device, medium and equipment
Zhang et al. Clustering Gaussian mixture reduction algorithm based on fuzzy adaptive resonance theory for extended target tracking
CN117033975A (en) Time series abnormality detection method, computer device, and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant