CN113485986B - Electric power data restoration method - Google Patents
Electric power data restoration method Download PDFInfo
- Publication number
- CN113485986B CN113485986B CN202110717117.1A CN202110717117A CN113485986B CN 113485986 B CN113485986 B CN 113485986B CN 202110717117 A CN202110717117 A CN 202110717117A CN 113485986 B CN113485986 B CN 113485986B
- Authority
- CN
- China
- Prior art keywords
- power data
- data
- gate
- output
- input
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/21—Design, administration or maintenance of databases
- G06F16/215—Improving data quality; Data cleansing, e.g. de-duplication, removing invalid entries or correcting typographical errors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Biophysics (AREA)
- Evolutionary Computation (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Computational Linguistics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Databases & Information Systems (AREA)
- Quality & Reliability (AREA)
- Supply And Distribution Of Alternating Current (AREA)
Abstract
Description
技术领域Technical Field
本发明涉及一种电力数据修复方法,属于电力数据检测修复技术领域。The invention relates to a power data repair method, belonging to the technical field of power data detection and repair.
背景技术Background technique
数字化技术不断发展进步,使得电力系统中产生了大量的电力数据。但是,电力数据常常因为外界干扰、传输错误、设备异常、网络延迟等原因而出现数据缺失的情况,这样会影响电力系统内处理数据的正确性与及时性。The continuous development and progress of digital technology has generated a large amount of power data in the power system. However, power data is often missing due to external interference, transmission errors, equipment abnormalities, network delays, etc., which will affect the accuracy and timeliness of data processing in the power system.
现有修复方法多采用传统机器学习方法,但是,面对目前复杂的日益结构复杂的电网结构,尤其是新能源发电系统与电动汽车的接入、需求响应机制的应用,导致传统机器学习方法无法应对高随机性的电力数据的修复。Existing repair methods mostly use traditional machine learning methods. However, faced with the increasingly complex power grid structure, especially the access of new energy power generation systems and electric vehicles, and the application of demand response mechanisms, traditional machine learning methods are unable to cope with the repair of highly random power data.
为了保证智能化电力系统的优化稳定运行,需要有完整正确的电力数据提供支持,避免影响电力系统安全稳定运行。本领域技术人员急需要对电力系统的数据进行更高精度的修复。In order to ensure the optimized and stable operation of the intelligent power system, complete and correct power data is needed to provide support to avoid affecting the safe and stable operation of the power system. Technical personnel in this field urgently need to repair the data of the power system with higher accuracy.
发明内容Summary of the invention
目的:为了克服现有技术中存在的不足,本发明提供一种电力数据修复方法。Purpose: In order to overcome the deficiencies in the prior art, the present invention provides a method for repairing power data.
技术方案:为解决上述技术问题,本发明采用的技术方案为:Technical solution: To solve the above technical problems, the technical solution adopted by the present invention is:
一种电力数据修复方法,包括如下步骤:A method for repairing power data comprises the following steps:
步骤S1、获取历史电力数据集,利用SOM神经网络对历史电力数据集中电力数据进行分类处理,得到电力数据类型。Step S1, obtaining a historical power data set, and using a SOM neural network to classify the power data in the historical power data set to obtain a power data type.
步骤S2、利用Pearson相关系数理论,对电力数据类型与影响因素进行相关性分析,获得电力数据类型满足关联性阈值的影响因素,将电力数据类型对应的影响因素作为特征值。Step S2: Use the Pearson correlation coefficient theory to perform correlation analysis on the power data type and the influencing factors, obtain the influencing factors of the power data type that meet the correlation threshold, and use the influencing factors corresponding to the power data type as feature values.
步骤S3、将电力数据类型与对应的特征值作为训练样本,对LSTM神经网络进行训练,获得训练好的LSTM神经网络。Step S3: Use the power data type and the corresponding characteristic value as training samples to train the LSTM neural network to obtain a trained LSTM neural network.
步骤S4、将缺失数据的影响因素输入训练好的LSTM神经网络,获得缺失数据的电力数据类型。Step S4: Input the influencing factors of the missing data into the trained LSTM neural network to obtain the power data type of the missing data.
步骤S5、根据缺失数据的电力数据类型采用不同的方法对数据进行修复。Step S5: Use different methods to repair the data according to the type of power data of the missing data.
作为优选方案,所述SOM神经网络由输入层和输出层两层组成,输入层与输出层之间采用全连接。在对输入电力数据进行分类时,输入层的各神经元协同工作,分别竞争输入电力数据的响应机会,从而得到输出的神经元。通过对权值的更新,输出层的神经元四周的权值都会进行相应的调整,电力数据经过调整后输入层的每一个神经元对特定类别的学习,实现输出层对输入电力数据的学习分类。权值取值范围通常为[0,1]。As a preferred solution, the SOM neural network consists of two layers, an input layer and an output layer, and the input layer and the output layer are fully connected. When classifying the input power data, the neurons in the input layer work together to compete for the response opportunities of the input power data, thereby obtaining the output neurons. By updating the weights, the weights around the neurons in the output layer will be adjusted accordingly. After the power data is adjusted, each neuron in the input layer learns a specific category, thereby realizing the learning and classification of the input power data by the output layer. The weight value range is usually [0,1].
作为优选方案,所述Pearson相关系数理论,计算公式如下:As a preferred solution, the Pearson correlation coefficient theory is calculated as follows:
式中:ρX.Y为相关性系数,范围为[0,1];X,Y分别为连续变量;n为连续变量样本的个数;和分别为连续变量的平均值;ρX.Y=0,表示变量间无关;ρX.Y>0,表示变量间正相关;ρX.Y<0,表示变量间负相关;|ρX,Y|越大说明变量间关联性越强,反之越弱。Where: ρ XY is the correlation coefficient, ranging from [0,1]; X, Y are continuous variables; n is the number of continuous variable samples; and are the average values of continuous variables respectively; ρ XY = 0, indicating that the variables are independent of each other; ρ XY > 0, indicating that the variables are positively correlated; ρ XY < 0, indicating that the variables are negatively correlated; the larger the |ρ X,Y |, the stronger the correlation between the variables, and vice versa.
作为优选方案,所述LSTM神经网络构建如下所示:As a preferred solution, the LSTM neural network is constructed as follows:
LSTM神经网络包括3个门结构实现对电力数据类型的预测,分别是遗忘门,输入门和输出门。The LSTM neural network includes three gate structures to realize the prediction of power data type, namely forget gate, input gate and output gate.
(1)遗忘门。主要用于对输入数据筛选,计算数据的保留程度,通过sigmoid神经层处理后输出一个[0,1]范围内的值,数值越大表示保留成分越多,反之越少。遗忘门计算公式如下:(1) Forget gate. It is mainly used to filter input data and calculate the degree of data retention. After processing through the sigmoid neural layer, it outputs a value in the range of [0,1]. The larger the value, the more components are retained, and vice versa. The calculation formula of the forget gate is as follows:
ft=σ(Wxfxt+Whfht-1+Bf)f t =σ(W xf x t +W hf h t-1 +B f )
式中:ft为遗忘门的输出;Wxf和Whf为遗忘门需要学习的网络参数;Bf为遗忘门的偏置;σ为sigmoid函数:Where: ft is the output of the forget gate; Wxf and Whf are the network parameters that the forget gate needs to learn; Bf is the bias of the forget gate; σ is the sigmoid function:
(2)输入门。主要用于对数据状态的更新,主要包括两部分信息,一部分是sigmoid函数选取需要保存的数据;另一部分是tanh函数把当前输入xt生成的新信息,C′t将两部分信息结合起来产生新的记忆状态,计算公式如下:(2) Input gate. It is mainly used to update the data state. It mainly includes two parts of information. One part is the data to be saved selected by the sigmoid function; the other part is the new information generated by the tanh function for the current input xt . C′t combines the two parts of information to generate a new memory state. The calculation formula is as follows:
it=δ(Wxixt+Whiht-1+Bi)i t =δ(W xi x t +W hi h t-1 +B i )
C′t=tanh(WxCxt+WhCht-1+BC)C′ t =tanh(W xC x t +W hC h t-1 +B C )
Ct=ft×Ct-1+it×C′t C t =f t ×C t-1 +i t ×C′ t
式中:it为输入门的输出;Wxi和Whi为输入门需要学习的网络参数;Bi为输入门的偏置;Ct′为tanh函数更新后的输入;Wxc和Whc为细胞状态需要学习的网络参数;Bc为细胞状态的偏置;Ct为细胞状态。Where: it is the output of the input gate; Wxi and Whi are the network parameters that need to be learned by the input gate; Bi is the bias of the input gate; Ct ′ is the input after the tanh function is updated; Wxc and Whc are the network parameters that need to be learned for the cell state; Bc is the bias of the cell state; Ct is the cell state.
(3)输出门。主要用于计算输入信息xt的输出程度。通过细胞状态中输出的信息经过tanh函数后与sigmoid函数做乘得到最终的隐藏单元ht,计算公式如下:(3) Output gate. It is mainly used to calculate the output level of input information xt . The information output in the cell state is multiplied by the sigmoid function after passing through the tanh function to obtain the final hidden unit ht . The calculation formula is as follows:
ot=σ(Wxoxt+Whoht-1+Bo)o t =σ(W xo x t +W ho h t-1 +B o )
ht=ot×tanh(Ct)h t = o t × tanh(C t )
式中:Ot为输出门的输出;ht为输出的隐藏单元;Wxo和Who为输出门的需要学习的网络参数;Bo为输出门的偏置。Where: Ot is the output of the output gate; ht is the hidden unit of the output; Wxo and Who are the network parameters of the output gate that need to be learned; Bo is the bias of the output gate.
作为优选方案,所述根据缺失数据的电力数据类型采用不同的方法对数据进行修复,包括如下步骤:As a preferred solution, the method of repairing the data using different methods according to the type of power data of the missing data includes the following steps:
当缺失数据的电力数据类型为电能数据时,采用分层均值填补法,将电能数据按每年的相同月份进行分层,取每个层的平均值代替缺失值。When the type of power data with missing data is electric energy data, the stratified mean filling method is used to stratify the electric energy data according to the same month of each year, and the average value of each layer is used to replace the missing value.
当缺失数据的电力数据类型为负荷数据时,采用最大期望算法,对缺失数据进行最大似然估计,并不断迭代,最终输出值代替缺失值。When the type of power data with missing data is load data, the maximum expectation algorithm is used to perform maximum likelihood estimation on the missing data, and iterates continuously to finally output the value to replace the missing value.
当缺失数据的电力数据类型为电流数据、电压数据或频率数据时,采用均值填补法,对前后两个观测值的平均值或者是非缺失数据的均值来代替缺失值。When the type of power data with missing data is current data, voltage data or frequency data, the mean filling method is used to replace the missing value with the average of the two previous and next observations or the mean of the non-missing data.
有益效果:本发明提供的一种电力数据修复方法,本方法考虑到电力数据的复杂非线性,利用神经网络的学习能力强,能够处理非线性问题的特性,实现电力数据的修复,可有效提升分类效率和准确率。Beneficial effect: The present invention provides a method for repairing power data. This method takes into account the complex nonlinearity of power data, utilizes the strong learning ability of neural networks, and can handle the characteristics of nonlinear problems to achieve power data repair, which can effectively improve classification efficiency and accuracy.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
图1是本发明的方法流程示意图。FIG. 1 is a schematic flow chart of the method of the present invention.
图2是本发明一实施例提供的分类模型SOM-LSTM神经网络的性能准确率对比图。FIG2 is a performance accuracy comparison chart of the SOM-LSTM neural network classification model provided by an embodiment of the present invention.
图3是本发明另一实施例提供的分类模型SOM-LSTM神经网络的性能准确率对比图。FIG3 is a performance accuracy comparison chart of a SOM-LSTM neural network classification model provided by another embodiment of the present invention.
具体实施方式Detailed ways
下面结合具体实施例对本发明作更进一步的说明。The present invention will be further described below in conjunction with specific embodiments.
以下结合附图和具体实施方式对本发明提出的一种电力数据修复方法作进一步详细说明。根据下面说明,本发明的优点和特征将更清楚。需要说明的是,附图采用非常简化的形式且均使用非精准的比例,仅用以方便、明晰地辅助说明本发明实施方式的目的。为了使本发明的目的、特征和优点能够更加明显易懂,请参阅附图。须知,本说明书所附图式所绘示的结构、比例、大小等,均仅用以配合说明书所揭示的内容,以供熟悉此技术的人士了解与阅读,并非用以限定本发明实施的限定条件,故不具技术上的实质意义,任何结构的修饰、比例关系的改变或大小的调整,在不影响本发明所能产生的功效及所能达成的目的下,均应仍落在本发明所揭示的技术内容能涵盖的范围内。The following is a further detailed description of a power data repair method proposed in the present invention in combination with the accompanying drawings and specific embodiments. According to the following description, the advantages and features of the present invention will be clearer. It should be noted that the drawings are in a very simplified form and use non-precise proportions, which are only used to conveniently and clearly assist in explaining the purpose of the embodiments of the present invention. In order to make the purposes, features and advantages of the present invention more obvious and easy to understand, please refer to the accompanying drawings. It should be noted that the structure, proportion, size, etc. illustrated in the drawings of this specification are only used to match the contents disclosed in the specification for people familiar with this technology to understand and read, and are not used to limit the limiting conditions for the implementation of the present invention, so it has no technical substantive significance. Any modification of the structure, change in the proportional relationship or adjustment of the size, without affecting the effects that the present invention can produce and the purposes that can be achieved, should still fall within the scope of the technical content disclosed in the present invention.
需要说明的是,在本文中,诸如第一和第二等之类的关系术语仅仅用来将一个实体或者操作与另一个实体或操作区分开来,而不一定要求或者暗示这些实体或操作之间存在任何这种实际的关系或者顺序。而且,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者设备不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者设备所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括所述要素的过程、方法、物品或者设备中还存在另外的相同要素。It should be noted that, in this article, relational terms such as first and second, etc. are only used to distinguish one entity or operation from another entity or operation, and do not necessarily require or imply any such actual relationship or order between these entities or operations. Moreover, the terms "include", "comprise" or any other variants thereof are intended to cover non-exclusive inclusion, so that a process, method, article or device including a series of elements includes not only those elements, but also other elements not explicitly listed, or also includes elements inherent to such process, method, article or device. In the absence of further restrictions, the elements defined by the sentence "comprise a ..." do not exclude the presence of other identical elements in the process, method, article or device including the elements.
结合附图1所示,本实施例提供的一种电力数据修复方法,包括如下步骤:As shown in FIG. 1 , the present embodiment provides a method for repairing power data, comprising the following steps:
步骤S1、获取历史电力数据集,利用SOM神经网络对历史电力数据集中电力数据进行分类处理,得到电力数据类型。Step S1, obtaining a historical power data set, and using a SOM neural network to classify the power data in the historical power data set to obtain a power data type.
步骤S2、利用Pearson相关系数理论,对电力数据类型与影响因素进行相关性分析,获得电力数据类型满足关联性阈值的影响因素,将电力数据类型对应的影响因素作为特征值。Step S2: Use the Pearson correlation coefficient theory to perform correlation analysis on the power data type and the influencing factors, obtain the influencing factors of the power data type that meet the correlation threshold, and use the influencing factors corresponding to the power data type as feature values.
步骤S3、将电力数据类型与对应的特征值作为训练样本,对LSTM神经网络进行训练,获得训练好的LSTM神经网络。Step S3: Use the power data type and the corresponding characteristic value as training samples to train the LSTM neural network to obtain a trained LSTM neural network.
步骤S4、将缺失数据的影响因素输入训练好的LSTM神经网络,获得缺失数据的电力数据类型。Step S4: Input the influencing factors of the missing data into the trained LSTM neural network to obtain the power data type of the missing data.
步骤S5、根据缺失数据的电力数据类型采用不同的方法对数据进行修复。Step S5: Use different methods to repair the data according to the type of power data of the missing data.
所述SOM神经网络由输入层和输出层两层组成,输入层与输出层之间采用全连接。在对输入电力数据进行分类时,输入层的各神经元协同工作,分别竞争输入电力数据的响应机会,从而得到输出的神经元。通过对权值的更新,输出层的神经元四周的权值都会进行相应的调整,电力数据经过调整后输入层的每一个神经元对特定类别的学习,实现输出层对输入电力数据的学习分类。权值取值范围通常为[0,1]。The SOM neural network consists of two layers, the input layer and the output layer, and the input layer and the output layer are fully connected. When classifying the input power data, the neurons in the input layer work together to compete for the response opportunities of the input power data, thereby obtaining the output neurons. By updating the weights, the weights around the neurons in the output layer will be adjusted accordingly. After the power data is adjusted, each neuron in the input layer learns a specific category, thereby realizing the learning and classification of the input power data in the output layer. The weight value range is usually [0,1].
所述Pearson相关系数理论,计算公式如下:The Pearson correlation coefficient theory, the calculation formula is as follows:
式中:ρX.Y为相关性系数,范围为[0,1];X,Y分别为连续变量;n为连续变量样本的个数;和分别为连续变量的平均值;ρX.Y=0,表示变量间无关;ρX.Y>0,表示变量间正相关;ρX.Y<0,表示变量间负相关;|ρX,Y|越大说明变量间关联性越强,反之越弱。Where: ρ XY is the correlation coefficient, ranging from [0,1]; X, Y are continuous variables; n is the number of continuous variable samples; and are the average values of continuous variables respectively; ρ XY = 0, indicating that the variables are independent of each other; ρ XY > 0, indicating that the variables are positively correlated; ρ XY < 0, indicating that the variables are negatively correlated; the larger the |ρ X,Y |, the stronger the correlation between the variables, and vice versa.
所述LSTM神经网络构建如下所示:The LSTM neural network is constructed as follows:
LSTM神经网络包括3个门结构实现对电力数据类型的预测,分别是遗忘门,输入门和输出门。The LSTM neural network includes three gate structures to realize the prediction of power data type, namely forget gate, input gate and output gate.
(1)遗忘门。主要用于对输入数据筛选,计算数据的保留程度,通过sigmoid神经层处理后输出一个[0,1]范围内的值,数值越大表示保留成分越多,反之越少。遗忘门计算公式如下:(1) Forget gate. It is mainly used to filter input data and calculate the degree of data retention. After processing through the sigmoid neural layer, it outputs a value in the range of [0,1]. The larger the value, the more components are retained, and vice versa. The calculation formula of the forget gate is as follows:
ft=σ(Wxfxt+Whfht-1+Bf)f t =σ(W xf x t +W hf h t-1 +B f )
式中:ft为遗忘门的输出;Wxf和Whf为遗忘门需要学习的网络参数;Bf为遗忘门的偏置;σ为sigmoid函数:Where: ft is the output of the forget gate; Wxf and Whf are the network parameters that the forget gate needs to learn; Bf is the bias of the forget gate; σ is the sigmoid function:
(2)输入门。主要用于对数据状态的更新,主要包括两部分信息,一部分是sigmoid函数选取需要保存的数据;另一部分是tanh函数把当前输入xt生成的新信息,C′t将两部分信息结合起来产生新的记忆状态,计算公式如下:(2) Input gate. It is mainly used to update the data state. It mainly includes two parts of information. One part is the data to be saved selected by the sigmoid function; the other part is the new information generated by the tanh function for the current input xt . C′t combines the two parts of information to generate a new memory state. The calculation formula is as follows:
it=δ(Wxixt+Whiht-1+Bi)i t =δ(W xi x t +W hi h t-1 +B i )
C′t=tanh(WxCxt+WhCht-1+BC)C′ t =tanh(W xC x t +W hC h t-1 +B C )
Ct=ft×Ct-1+it×C′t C t =f t ×C t-1 +i t ×C′ t
式中:it为输入门的输出;Wxi和Whi为输入门需要学习的网络参数;Bi为输入门的偏置;Ct′为tanh函数更新后的输入;Wxc和Whc为细胞状态需要学习的网络参数;Bc为细胞状态的偏置;Ct为细胞状态。Where: it is the output of the input gate; Wxi and Whi are the network parameters that need to be learned by the input gate; Bi is the bias of the input gate; Ct ′ is the input after the tanh function is updated; Wxc and Whc are the network parameters that need to be learned for the cell state; Bc is the bias of the cell state; Ct is the cell state.
(3)输出门。主要用于计算输入信息xt的输出程度。通过细胞状态中输出的信息经过tanh函数后与sigmoid函数做乘得到最终的隐藏单元ht,计算公式如下:(3) Output gate. It is mainly used to calculate the output level of input information xt . The information output in the cell state is multiplied by the sigmoid function after passing through the tanh function to obtain the final hidden unit ht . The calculation formula is as follows:
ot=σ(Wxoxt+Whoht-1+Bo)o t =σ(W xo x t +W ho h t-1 +B o )
ht=ot×tanh(Ct)h t = o t × tanh(C t )
式中:Ot为输出门的输出;ht为输出的隐藏单元;Wxo和Who为输出门的需要学习的网络参数;Bo为输出门的偏置。Where: Ot is the output of the output gate; ht is the hidden unit of the output; Wxo and Who are the network parameters of the output gate that need to be learned; Bo is the bias of the output gate.
所述根据缺失数据的电力数据类型采用不同的方法对数据进行修复,包括如下步骤:The method of repairing the data using different methods according to the type of power data of the missing data includes the following steps:
当缺失数据的电力数据类型为电能数据时,采用分层均值填补法,将电能数据按每年的相同月份进行分层,取每个层的平均值代替缺失值。When the type of power data with missing data is electric energy data, the stratified mean filling method is used to stratify the electric energy data according to the same month of each year, and the average value of each layer is used to replace the missing value.
当缺失数据的电力数据类型为负荷数据时,采用最大期望算法,对缺失数据进行最大似然估计,并不断迭代,最终输出值代替缺失值。When the type of power data with missing data is load data, the maximum expectation algorithm is used to perform maximum likelihood estimation on the missing data, and iterates continuously to finally output the value to replace the missing value.
当缺失数据的电力数据类型为电流数据、电压数据或频率数据时,采用均值填补法,对前后两个观测值的平均值或者是非缺失数据的均值来代替缺失值。When the type of power data with missing data is current data, voltage data or frequency data, the mean filling method is used to replace the missing value with the average of the two previous and next observations or the mean of the non-missing data.
如图2,3所示,为了验证所提的基于SOM-LSMT电力数据修复方法的有效性,本文实验仿真的计算机配置为Inter Intel(R)Core(TM)i5-8300H CPU@2.30GHz2.30GHz和NVIDIAGTX 1050Ti 4G显存,在MATLAB R2019a平台上进行仿真计算。As shown in Figures 2 and 3, in order to verify the effectiveness of the proposed SOM-LSMT-based power data repair method, the computer configuration of the experimental simulation in this paper is Inter Intel(R)Core(TM)i5-8300H CPU@2.30GHz2.30GHz and NVIDIAGTX 1050Ti 4G video memory, and the simulation calculation is carried out on the MATLAB R2019a platform.
本文选取2018年8月中国某地区的家庭电力数据以及相应气象站所采集到的温度数据,数据采集间隔为15分钟。选用平均绝对误差((mean absolute error,MAE)和均方根误差(root mean square error,RMSE)评价指标对电力数据修复结果进行分析。This paper selects household electricity data from a certain area in China in August 2018 and temperature data collected by the corresponding meteorological station, with a data collection interval of 15 minutes. The mean absolute error (MAE) and root mean square error (RMSE) evaluation indicators are used to analyze the power data repair results.
为了对所提出的基于SOM-LSTM模型的准确性进行分析,本文分别对典型工作日和非工作日的连续缺失电力数据SOM-LSTM模型与极限学习机(Extreme Learning Machine,ELM)模型、LSTM模型的修复结果绘制曲线进行对比,结果如图2和图3所示。In order to analyze the accuracy of the proposed SOM-LSTM model, this paper draws curves to compare the repair results of the SOM-LSTM model, the Extreme Learning Machine (ELM) model, and the LSTM model for continuous missing power data on typical working days and non-working days. The results are shown in Figures 2 and 3.
由图2和图3可知,针对典型工作日和非工作日的用户电压数据连续缺失进行修复时,LSTM神经网络的误差最大,且随着数据缺失的增加,精确度显著降低,说明该神经网络的拟合修复效果不足。对比LSTM模型和ELM模型,经过SOM神经网络分类后的数据进行预测修复的SOM-LSTM神经网络能够有效改善连续缺失数据修复准确率低的问题,并且在数据缺失较多的情况下依然保证较高水平的修复率。As shown in Figures 2 and 3, when repairing the continuous missing voltage data of users on typical working days and non-working days, the LSTM neural network has the largest error, and as the data missing increases, the accuracy decreases significantly, indicating that the fitting and repair effect of the neural network is insufficient. Compared with the LSTM model and the ELM model, the SOM-LSTM neural network that predicts and repairs the data after classification by the SOM neural network can effectively improve the problem of low repair accuracy of continuous missing data, and still maintain a high level of repair rate when there is a lot of data missing.
尽管本发明的内容已经通过上述优选实施例作了详细介绍,但应当认识到上述的描述不应被认为是对本发明的限制。在本领域技术人员阅读了上述内容后,对于本发明的多种修改和替代都将是显而易见的。因此,本发明的保护范围应由所附的权利要求来限定。Although the content of the present invention has been described in detail through the above preferred embodiments, it should be appreciated that the above description should not be considered as a limitation of the present invention. After reading the above content, it will be apparent to those skilled in the art that various modifications and substitutions of the present invention will occur. Therefore, the protection scope of the present invention should be limited by the appended claims.
Claims (7)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110717117.1A CN113485986B (en) | 2021-06-25 | 2021-06-25 | Electric power data restoration method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110717117.1A CN113485986B (en) | 2021-06-25 | 2021-06-25 | Electric power data restoration method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113485986A CN113485986A (en) | 2021-10-08 |
CN113485986B true CN113485986B (en) | 2024-08-02 |
Family
ID=77936302
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110717117.1A Active CN113485986B (en) | 2021-06-25 | 2021-06-25 | Electric power data restoration method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113485986B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114298374A (en) * | 2021-11-30 | 2022-04-08 | 国家电网有限公司 | Demand response capability determination method and system considering missing user power curve |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2015158198A1 (en) * | 2014-04-17 | 2015-10-22 | 北京泰乐德信息技术有限公司 | Fault recognition method and system based on neural network self-learning |
CN110097920A (en) * | 2019-04-10 | 2019-08-06 | 大连理工大学 | A kind of metabolism group shortage of data value fill method based on neighbour's stability |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3514908B1 (en) * | 2018-01-22 | 2022-02-09 | Hitachi Energy Switzerland AG | Methods and devices for condition classification of power network assets |
CN109091867B (en) * | 2018-07-26 | 2023-04-07 | 深圳市腾讯网络信息技术有限公司 | Operation control method, device, equipment and storage medium |
CN109820525A (en) * | 2019-01-23 | 2019-05-31 | 五邑大学 | A driving fatigue recognition method based on CNN-LSTM deep learning model |
CN110083699B (en) * | 2019-03-18 | 2021-01-12 | 中国科学院自动化研究所 | News popularity prediction model training method based on deep neural network |
CN110334726A (en) * | 2019-04-24 | 2019-10-15 | 华北电力大学 | A method for identifying and repairing abnormal data of electric load based on density clustering and LSTM |
CN111597080A (en) * | 2020-05-22 | 2020-08-28 | 广东省生态环境技术研究所 | Method for repairing underground water level missing data based on ground statistics and neural network |
-
2021
- 2021-06-25 CN CN202110717117.1A patent/CN113485986B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2015158198A1 (en) * | 2014-04-17 | 2015-10-22 | 北京泰乐德信息技术有限公司 | Fault recognition method and system based on neural network self-learning |
CN110097920A (en) * | 2019-04-10 | 2019-08-06 | 大连理工大学 | A kind of metabolism group shortage of data value fill method based on neighbour's stability |
Also Published As
Publication number | Publication date |
---|---|
CN113485986A (en) | 2021-10-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108959732B (en) | A Convolutional Neural Network Based Fault Type Identification Method for Transmission Lines | |
CN109086913B (en) | A method and system for transient stability assessment of power system based on deep learning | |
CN109063939B (en) | A Wind Speed Prediction Method and System Based on Neighborhood Gate Long Short-Term Memory Network | |
CN115688982B (en) | Building photovoltaic data complement method based on WGAN and whale optimization algorithm | |
CN106126906A (en) | Short-term wind speed forecasting method based on C C Yu ELM | |
Wang et al. | Wind power curve modeling with asymmetric error distribution | |
CN113361803A (en) | Ultra-short-term photovoltaic power prediction method based on generation countermeasure network | |
CN115935810A (en) | Power medium-term load forecasting method and system based on fusion features of attention mechanism | |
CN111585277B (en) | A dynamic safety assessment method of power system based on hybrid integrated model | |
CN114792156A (en) | Photovoltaic output power prediction method and system based on clustering of curve characteristic indexes | |
CN105354643A (en) | Risk prediction evaluation method for wind power grid integration | |
CN116559975A (en) | Multi-step long weather prediction method based on multi-element time sequence diagram neural network | |
CN115409317A (en) | Transformer area line loss detection method and device based on feature selection and machine learning | |
CN113485986B (en) | Electric power data restoration method | |
CN111091141B (en) | Photovoltaic backboard fault diagnosis method based on layered Softmax | |
CN106503793B (en) | A kind of neural network short-term wind speed forecasting method based on improvement difference algorithm | |
CN115860212A (en) | Risk prediction method and terminal for power distribution network | |
CN115392697A (en) | A data-model hybrid driven power system security assessment method and system | |
CN113762591B (en) | Short-term electric quantity prediction method and system based on GRU and multi-core SVM countermeasure learning | |
Zhang | Short‐Term Power Load Forecasting Based on SAPSO‐CNN‐LSTM Model considering Autocorrelated Errors | |
CN110543615A (en) | Interaction Analysis Method of Risk Factors Based on SPSS Interpretation Structure Model | |
CN106372440A (en) | Method and device for estimating self-adaptive robust state of distribution network through parallel computation | |
CN111628531B (en) | Data driving method for static voltage stability evaluation of power system | |
CN118365016A (en) | Method and system for predicting load of power exchange station by fusing Attention and CNN-GRU | |
CN118523425A (en) | Multi-source data-fused wind turbine generator prediction optimization control method and system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |