CN112465664B - AVC intelligent control method based on artificial neural network and deep reinforcement learning - Google Patents
AVC intelligent control method based on artificial neural network and deep reinforcement learning Download PDFInfo
- Publication number
- CN112465664B CN112465664B CN202011263523.7A CN202011263523A CN112465664B CN 112465664 B CN112465664 B CN 112465664B CN 202011263523 A CN202011263523 A CN 202011263523A CN 112465664 B CN112465664 B CN 112465664B
- Authority
- CN
- China
- Prior art keywords
- neural network
- function
- avc
- artificial neural
- agent
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 67
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 54
- 230000002787 reinforcement Effects 0.000 title claims abstract description 37
- 230000006870 function Effects 0.000 claims abstract description 76
- 230000009471 action Effects 0.000 claims abstract description 38
- 238000012549 training Methods 0.000 claims abstract description 19
- 230000008859 change Effects 0.000 claims abstract description 12
- 239000003795 chemical substances by application Substances 0.000 claims description 53
- 239000013598 vector Substances 0.000 claims description 39
- 238000004364 calculation method Methods 0.000 claims description 6
- 238000012546 transfer Methods 0.000 claims description 6
- 238000011217 control strategy Methods 0.000 claims description 3
- 238000009826 distribution Methods 0.000 claims description 3
- 239000011159 matrix material Substances 0.000 claims description 3
- 210000002569 neuron Anatomy 0.000 claims description 3
- 210000004205 output neuron Anatomy 0.000 claims description 3
- 239000000203 mixture Substances 0.000 claims 1
- 239000002994 raw material Substances 0.000 claims 1
- 230000004044 response Effects 0.000 claims 1
- 238000004422 calculation algorithm Methods 0.000 abstract description 6
- 238000010586 diagram Methods 0.000 description 11
- 230000008569 process Effects 0.000 description 11
- 238000004458 analytical method Methods 0.000 description 6
- 230000000694 effects Effects 0.000 description 6
- 230000003993 interaction Effects 0.000 description 6
- 238000005457 optimization Methods 0.000 description 5
- 238000012360 testing method Methods 0.000 description 4
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 description 3
- 238000011161 development Methods 0.000 description 3
- 230000018109 developmental process Effects 0.000 description 3
- 230000008447 perception Effects 0.000 description 3
- 230000007613 environmental effect Effects 0.000 description 2
- 230000005284 excitation Effects 0.000 description 2
- 230000007774 longterm Effects 0.000 description 2
- 238000005070 sampling Methods 0.000 description 2
- 230000003044 adaptive effect Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000007621 cluster analysis Methods 0.000 description 1
- 238000007418 data mining Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000012886 linear function Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000005065 mining Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 238000005192 partition Methods 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000036962 time dependent Effects 0.000 description 1
- 230000001052 transient effect Effects 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
- 238000012800 visualization Methods 0.000 description 1
- 230000002087 whitening effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q50/00—Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
- G06Q50/06—Energy or water supply
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
- G06F17/18—Complex mathematical operations for evaluating statistical data, e.g. average values, frequency distributions, probability functions, regression analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/23—Clustering techniques
- G06F18/232—Non-hierarchical techniques
- G06F18/2321—Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
- G06F18/23213—Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions with fixed number of clusters, e.g. K-means clustering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
- G06Q10/04—Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
-
- H—ELECTRICITY
- H02—GENERATION; CONVERSION OR DISTRIBUTION OF ELECTRIC POWER
- H02J—CIRCUIT ARRANGEMENTS OR SYSTEMS FOR SUPPLYING OR DISTRIBUTING ELECTRIC POWER; SYSTEMS FOR STORING ELECTRIC ENERGY
- H02J3/00—Circuit arrangements for ac mains or ac distribution networks
-
- H—ELECTRICITY
- H02—GENERATION; CONVERSION OR DISTRIBUTION OF ELECTRIC POWER
- H02J—CIRCUIT ARRANGEMENTS OR SYSTEMS FOR SUPPLYING OR DISTRIBUTING ELECTRIC POWER; SYSTEMS FOR STORING ELECTRIC ENERGY
- H02J3/00—Circuit arrangements for ac mains or ac distribution networks
- H02J3/12—Circuit arrangements for ac mains or ac distribution networks for adjusting voltage in ac networks by changing a characteristic of the network load
- H02J3/16—Circuit arrangements for ac mains or ac distribution networks for adjusting voltage in ac networks by changing a characteristic of the network load by adjustment of reactive power
-
- H—ELECTRICITY
- H02—GENERATION; CONVERSION OR DISTRIBUTION OF ELECTRIC POWER
- H02J—CIRCUIT ARRANGEMENTS OR SYSTEMS FOR SUPPLYING OR DISTRIBUTING ELECTRIC POWER; SYSTEMS FOR STORING ELECTRIC ENERGY
- H02J3/00—Circuit arrangements for ac mains or ac distribution networks
- H02J3/38—Arrangements for parallely feeding a single network by two or more generators, converters or transformers
- H02J3/46—Controlling of the sharing of output between the generators, converters, or transformers
- H02J3/50—Controlling the sharing of the out-of-phase component
-
- H—ELECTRICITY
- H02—GENERATION; CONVERSION OR DISTRIBUTION OF ELECTRIC POWER
- H02J—CIRCUIT ARRANGEMENTS OR SYSTEMS FOR SUPPLYING OR DISTRIBUTING ELECTRIC POWER; SYSTEMS FOR STORING ELECTRIC ENERGY
- H02J2203/00—Indexing scheme relating to details of circuit arrangements for AC mains or AC distribution networks
- H02J2203/20—Simulating, e g planning, reliability check, modelling or computer assisted design [CAD]
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02E—REDUCTION OF GREENHOUSE GAS [GHG] EMISSIONS, RELATED TO ENERGY GENERATION, TRANSMISSION OR DISTRIBUTION
- Y02E40/00—Technologies for an efficient electrical power generation, transmission or distribution
- Y02E40/30—Reactive power compensation
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02E—REDUCTION OF GREENHOUSE GAS [GHG] EMISSIONS, RELATED TO ENERGY GENERATION, TRANSMISSION OR DISTRIBUTION
- Y02E40/00—Technologies for an efficient electrical power generation, transmission or distribution
- Y02E40/70—Smart grids as climate change mitigation technology in the energy generation sector
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y04—INFORMATION OR COMMUNICATION TECHNOLOGIES HAVING AN IMPACT ON OTHER TECHNOLOGY AREAS
- Y04S—SYSTEMS INTEGRATING TECHNOLOGIES RELATED TO POWER NETWORK OPERATION, COMMUNICATION OR INFORMATION TECHNOLOGIES FOR IMPROVING THE ELECTRICAL POWER GENERATION, TRANSMISSION, DISTRIBUTION, MANAGEMENT OR USAGE, i.e. SMART GRIDS
- Y04S10/00—Systems supporting electrical power generation, transmission or distribution
- Y04S10/50—Systems or methods supporting the power network operation or management, involving a certain degree of interaction with the load-side end user applications
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Business, Economics & Management (AREA)
- Economics (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Health & Medical Sciences (AREA)
- Strategic Management (AREA)
- Human Resources & Organizations (AREA)
- Power Engineering (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Software Systems (AREA)
- Evolutionary Biology (AREA)
- Mathematical Analysis (AREA)
- Bioinformatics & Computational Biology (AREA)
- Pure & Applied Mathematics (AREA)
- Probability & Statistics with Applications (AREA)
- Computing Systems (AREA)
- Mathematical Optimization (AREA)
- Molecular Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Computational Mathematics (AREA)
- Marketing (AREA)
- Operations Research (AREA)
- Biomedical Technology (AREA)
- Tourism & Hospitality (AREA)
- General Business, Economics & Management (AREA)
- Quality & Reliability (AREA)
- Public Health (AREA)
- Water Supply & Treatment (AREA)
- Primary Health Care (AREA)
- Entrepreneurship & Innovation (AREA)
Abstract
The invention discloses an AVC intelligent control method based on an artificial neural network and deep reinforcement learning, which comprises the steps of dividing a transformer substation into different sub-control areas by combining a situation prediction result of reactive load of a power grid and a reactive load change rule of a new energy grid-connected point; optimizing an action utility function based on a Bellman equation and a minimized loss function, and obtaining a decision metric function by combining the action utility function; training the agent by optimizing decision model parameters of the agent using the gradient of the decision metric function; and inputting the situation prediction results of different sub-regions and the reactive change rule of the new energy into the intelligent agent, and calculating the voltage control quantity of the power system through the intelligent agent to control the reactive voltage of the power grid. The invention trains the intelligent agent by combining the artificial neural network and the multi-agent reinforcement learning algorithm of the deterministic strategy, thereby improving the active control capability of the reactive voltage.
Description
Technical Field
The invention relates to the technical field of power control, in particular to an AVC intelligent control method based on an artificial neural network and deep reinforcement learning.
Background
In recent years, in the operation control process of the power system, the large-scale power failure accidents caused by insufficient situation perception are increased day by day in all countries in the world, and the wide-area situation perception of the power system is paid more and more attention; electric power system wide area situation perception includes through gathering wide area electric wire netting steady state and developments, electric quantity and non-electric quantity information: analyzing, understanding and evaluating equipment state information, power grid steady-state data information, power grid dynamic data information, power grid transient fault information, power grid operating environment information and the like by means of wide-area dynamic safety monitoring, data mining, dynamic parameter identification, super real-time simulation, visualization and the like, and further predicting the power grid development situation; the application of situation awareness technology in power systems is still in the beginning stage, and the situation awareness has been listed as one of the technical fields of preferential support of smart grids by mechanisms such as the U.S. federal energy management commission and the national standards and technical society.
With the rapid development of large-scale new energy access and alternating current-direct current hybrid power grids, the uncertainty of the source-load double sides is enhanced, the reactive voltage problem of the system is increasingly prominent, and the challenge is brought to the safe operation of the power grids; at present, reactive power optimization control belongs to system global optimization under a short time scale, control decisions do not have initiative and predictability, and the influence of uncertainty of new energy and reactive load on reactive voltage control under a long time scale is not fully considered, so that reactive power equipment is frequently adjusted, and the overall control effect under the long time scale is not ideal.
Disclosure of Invention
This section is for the purpose of summarizing some aspects of embodiments of the invention and to briefly introduce some preferred embodiments. In this section, as well as in the abstract and the title of the invention of this application, simplifications or omissions may be made to avoid obscuring the purpose of the section, the abstract and the title, and such simplifications or omissions are not intended to limit the scope of the invention.
The present invention has been made in view of the above-mentioned conventional problems.
Therefore, the invention provides an AVC intelligent control method based on an artificial neural network and deep reinforcement learning, which can avoid reactive voltage risks and solve the problem of poor reactive voltage active control effect.
In order to solve the technical problems, the invention provides the following technical scheme: the method comprises the steps of dividing the transformer substation into different sub-control areas by combining a situation prediction result of the reactive load of a power grid and a reactive load change rule of a new energy grid-connected point; optimizing an action utility function based on a Bellman equation and a minimized loss function, and obtaining a decision metric function by combining the action utility function; training the agent by optimizing decision model parameters of the agent using the gradient of the decision metric function; and inputting the situation prediction results of different sub-regions and the reactive change rule of the new energy into the intelligent agent, and calculating the voltage control quantity of the power system through the intelligent agent to control the reactive voltage of the power grid.
As a preferred scheme of the AVC intelligent control method based on the artificial neural network and the deep reinforcement learning, the AVC intelligent control method comprises the following steps: the situation prediction result comprises the steps of constructing a deep neural network regression model based on a deep artificial neural network, and integrating a plurality of regression load results of the deep neural network regression model to obtain the situation prediction result of the reactive load.
As a preferred scheme of the AVC intelligent control method based on the artificial neural network and the deep reinforcement learning, the AVC intelligent control method comprises the following steps: the method for constructing the deep neural network regression model comprises the following steps of constructing a regression model structure based on reactive load data characteristics and by considering climate environment, season, regional distribution, user load and power grid scheduling control strategies:
wherein k is the order; x is the number of(k)A k-order hidden layer node unit vector is obtained; y is(k)Is a k-order output node vector; u. of(k)Is an input vector of order k;inputting a vector for a k-order feedback state;is a k-order feedback state vector;outputting vectors for k-order hidden layers; omegaiA connection weight matrix of each layer, i is 1, 2, 3, 4, 5, 6; g () is the transfer function of the output neuron; f () is the transfer function of the middle layer neurons.
As a preferred scheme of the AVC intelligent control method based on the artificial neural network and the deep reinforcement learning, the AVC intelligent control method comprises the following steps: the regression model structure may further include,
wherein x is(k-1)Is a k-1 order hidden layer node unit vector;is a feedback state vector of k-1 order; u. of(k-1)Is an input vector of order k-1;inputting a vector for a feedback state of k-1 order; y is(k-1)Is a k-1 order output node vector;outputting vectors for k-1 order hidden layer; eta, b,Is a self-feedback gain factor.
As a preferred scheme of the AVC intelligent control method based on the artificial neural network and the deep reinforcement learning, the AVC intelligent control method comprises the following steps: the minimizing a loss function includes defining the minimizing a loss function:
wherein,for taking the independent variable as a training parameterA time-dependent minimum loss function, E is an expected value, s is a current system state, s' is an environmental state at a next moment, a is a selected action in a corresponding state,as a pool of experiences, yiTo pass through the Bellman equation pairAnd estimating the true value.
As a preferred scheme of the AVC intelligent control method based on the artificial neural network and the deep reinforcement learning, the AVC intelligent control method comprises the following steps: the estimated true values of the values may include,
wherein r isiIs the return value obtained in the ith iteration; mu is a decision value, gamma represents a decay rate, and gamma is in the range of 0,1];Qi' is a function of the Q value of the next state target Critic network; s' is the next state entered by taking the action a at the system state s; a' is according to the target Actor network in the system state sAn act of selecting;is a parameter of the target Actor network;is a parameter of the target Critic network.
As a preferred scheme of the AVC intelligent control method based on the artificial neural network and the deep reinforcement learning, the AVC intelligent control method comprises the following steps: the parameters of the target Critic network and the parameters of the target Actor network comprise parameters passing through the actual Actor networkAnd (3) updating parameters:
where τ controls the update rate.
As a preferred scheme of the AVC intelligent control method based on the artificial neural network and the deep reinforcement learning, the AVC intelligent control method comprises the following steps: the decision metric function includes defining the action utility function Qi(s, a) is the expectation of the sum of the rewards subsequently obtained by the agents in the ith area after the action a is executed in the system state s, and then the decision metric function is:
as a preferred scheme of the AVC intelligent control method based on the artificial neural network and the deep reinforcement learning, the AVC intelligent control method comprises the following steps: the gradient of the decision metric function comprises the decision function parameters of the i-th regional agentThe gradient of (d) is:
wherein,computing the sign of the gradient for the function; a isiAn action value representing the ith iteration;iterating the gradient i times for the action utility function,for the target Actor networkThe gradient is iterated i times.
As a preferred scheme of the AVC intelligent control method based on the artificial neural network and the deep reinforcement learning, the AVC intelligent control method comprises the following steps: the voltage control quantity comprises a calculation formula based on Newton-Raphson power flow, wherein the calculation formula of the voltage control quantity is as follows:
wherein U is the voltage control mass, MiFor modulation of voltage source converters, UdThe fundamental voltage of the dc node.
The invention has the beneficial effects that: the reactive voltage future situation prediction is formed based on the analysis of data samples of new energy and reactive load, the reactive voltage of a power grid is controlled through an intelligent agent, and meanwhile, the intelligent agent is trained by combining an artificial neural network and a multi-agent reinforcement learning algorithm of a deterministic strategy, so that the active control capability of the reactive voltage is improved.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings needed to be used in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without inventive exercise. Wherein:
FIG. 1 is a schematic flowchart of an AVC intelligent control method based on an artificial neural network and deep reinforcement learning according to a first embodiment of the present invention;
fig. 2 is a schematic diagram of a transformer substation and a substation system region division of the AVC intelligent control method based on an artificial neural network and deep reinforcement learning according to the first embodiment of the present invention;
fig. 3 is a schematic diagram of an Actor network structure of an AVC intelligent control method based on an artificial neural network and deep reinforcement learning according to a first embodiment of the present invention;
FIG. 4 is a schematic structural diagram of a criticic network of the AVC intelligent control method based on an artificial neural network and deep reinforcement learning according to the first embodiment of the present invention;
FIG. 5 is a schematic diagram of an agent training process of an AVC intelligent control method based on an artificial neural network and deep reinforcement learning according to a first embodiment of the present invention;
FIG. 6 is a schematic diagram of the operation flow of an agent in an AVC intelligent control method based on an artificial neural network and deep reinforcement learning according to a first embodiment of the present invention;
fig. 7 is a schematic diagram of a loss function curve of an Actor network of an AVC intelligent control method based on an artificial neural network and deep reinforcement learning according to a second embodiment of the present invention;
FIG. 8 is a graph illustrating a loss function curve of a criticc network of an AVC intelligent control method based on an artificial neural network and deep reinforcement learning according to a second embodiment of the present invention;
FIG. 9 is a diagram illustrating the variation of the total reward function and the action times of the AVC intelligent control method based on artificial neural network and deep reinforcement learning according to the second embodiment of the present invention with the training process;
fig. 10 is a schematic diagram of voltage amplitudes of nodes before and after the intelligent agent controls in a certain operating state according to the AVC intelligent control method based on the artificial neural network and the deep reinforcement learning according to the second embodiment of the present invention;
fig. 11 is a schematic diagram of a loss function curve of an Actor network in consideration of new energy output fluctuation according to a second embodiment of the AVC intelligent control method based on an artificial neural network and deep reinforcement learning;
fig. 12 is a schematic diagram of a loss function curve of the Critic network in consideration of new energy output fluctuation in the AVC intelligent control method based on an artificial neural network and deep reinforcement learning according to the second embodiment of the present invention;
fig. 13 is a schematic diagram of the action times of each agent in consideration of new energy fluctuation according to the AVC intelligent control method based on an artificial neural network and deep reinforcement learning according to the second embodiment of the present invention.
Detailed Description
In order to make the aforementioned objects, features and advantages of the present invention comprehensible, specific embodiments accompanied with figures are described in detail below, and it is apparent that the described embodiments are a part of the embodiments of the present invention, not all of the embodiments. All other embodiments, which can be obtained by a person skilled in the art without making creative efforts based on the embodiments of the present invention, shall fall within the protection scope of the present invention.
In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention, but the present invention may be practiced in other ways than those specifically described and will be readily apparent to those of ordinary skill in the art without departing from the spirit of the present invention, and therefore the present invention is not limited to the specific embodiments disclosed below.
Furthermore, reference herein to "one embodiment" or "an embodiment" means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one implementation of the invention. The appearances of the phrase "in one embodiment" in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments.
The present invention will be described in detail with reference to the drawings, wherein the cross-sectional views illustrating the structure of the device are not enlarged partially in general scale for convenience of illustration, and the drawings are only exemplary and should not be construed as limiting the scope of the present invention. In addition, the three-dimensional dimensions of length, width and depth should be included in the actual fabrication.
Meanwhile, in the description of the present invention, it should be noted that the terms "upper, lower, inner and outer" and the like indicate orientations or positional relationships based on the orientations or positional relationships shown in the drawings, and are only for convenience of describing the present invention and simplifying the description, but do not indicate or imply that the referred device or element must have a specific orientation, be constructed in a specific orientation and operate, and thus, cannot be construed as limiting the present invention. Furthermore, the terms first, second, or third are used for descriptive purposes only and are not to be construed as indicating or implying relative importance.
The terms "mounted, connected and connected" in the present invention are to be understood broadly, unless otherwise explicitly specified or limited, for example: can be fixedly connected, detachably connected or integrally connected; they may be mechanically, electrically, or directly connected, or indirectly connected through intervening media, or may be interconnected between two elements. The specific meanings of the above terms in the present invention can be understood in specific cases to those skilled in the art.
Example 1
Referring to fig. 1 to 6, a first embodiment of the present invention provides an AVC intelligent control method based on an artificial neural network and deep reinforcement learning, including:
s1: and dividing the transformer substation into different sub-control areas by combining the situation prediction result of the reactive load of the power grid and the reactive load change rule of the new energy grid-connected point.
(1) And constructing a deep neural network regression model based on the deep artificial neural network, and integrating a plurality of regression load results of the deep neural network regression model to further obtain a situation prediction result of the reactive load.
It should be noted that, before the deep neural network regression model is constructed, the load data is preprocessed by methods including denoising, normalization, whitening and the like, massive reactive load data is integrated, error load data is eliminated, and a reactive load data set with a complete structure, a standard format and a low error rate is generated.
Specifically, the constructing of the deep neural network regression model comprises,
based on the reactive load data characteristics, a regression model structure is constructed by considering climate environment, season, regional distribution, user load and a power grid dispatching control strategy, and input information of a middle layer of the structure comprises an input layer, an input bearing layer, a middle bearing layer and an output bearing layer; the input of the output layer comprises an intermediate layer and an intermediate receiving layer, and the mathematical formula corresponding to the regression model is as follows:
wherein k is the order; x is the number of(k)A k-order hidden layer node unit vector is obtained; y is(k)Is a k-order output node vector; u. of(k)Is an input vector of order k;inputting a vector for a k-order feedback state;is a k-order feedback state vector;outputting vectors for k-order hidden layers; omegaiA connection weight matrix of each layer, i is 1, 2, 3, 4, 5, 6; g () is the transfer function of the output neuron; f () is the transfer function of the middle layer neurons.
wherein x is(k-1)Is a k-1 order hidden layer node unit vector;is a feedback state vector of k-1 order; u. of(k-1)Is an input vector of order k-1;inputting a vector for a feedback state of k-1 order; y is(k-1)Is a k-1 order output node vector;a k-1 order hidden layer output vector; eta (eta is more than or equal to 0),Is a self-feedback gain factor.
It should be noted that, in this embodiment, g () uses a linear function, and f () uses a Sigmoid function;
sigmoid function is shown as follows:
further, integrating a plurality of regression load results of the regression model based on the preprocessed data set to obtain a situation prediction value of the reactive load.
(2) Because reactive power compensation devices installed in different new energy plants are different and reactive voltage control methods of the reactive power compensation devices are different, firstly, reactive power characteristics of different reactive power sources are analyzed based on the near-area actual situation of the new energy plants; and (3) integrating the output and load characteristics of the new energy based on a cluster analysis method, and equivalently obtaining the equivalent load characteristic of the near region of the system energy field station to obtain the fluctuation rule of the node voltage under the equivalent load characteristic.
Specifically, Clustering Analysis (Clustering Analysis) is an Analysis method for grouping according to the principle of maximizing intra-class similarity and minimizing inter-class similarity of objects, and also belongs to a descriptive mining task.
The embodiment adopts K-means to perform partition clustering on the data.
Dividing a data set D into K classes, and evaluating the cluster quality by using the sum of squared errors, wherein the classes are defined as follows:
wherein E represents the sum of the squares of the errors for all objects of the data set; p represents a point of a given data object in space; dist (x, y) represents the Euclidean distance in space from point x to point y.
And secondly, determining the value of the cluster number k by adopting an elbow method.
Finding out an inflection point through SSE (sum of the squared errors), wherein the K value at the moment is the value obtained; calculating the SSE:
wherein, CiIs the ith cluster, p is CiSample point of (1), miIs CiThe center of mass of;
region division referring to fig. 2, each is controlled by 2 different agents.
S2: and optimizing the action utility function based on the Bellman equation and the minimized loss function, and combining the action utility function to obtain a decision metric function.
Defining a minimization loss function:
wherein,for taking the independent variable as a training parameterThe function of the minimum loss of time,e is the expected value, s is the current system state, s' is the environmental state at the next moment, a is the selected action in the corresponding state,as a pool of experiences, yiIs to pass through a Bellman equation pairAnd estimating the true value.
In particular, the method comprises the following steps of,
wherein r isiThe return value obtained in the ith iteration is obtained; mu is a decision value; gamma denotes the decay rate and gamma is in [0,1 ]]When γ is 0, only immediate return is considered and no long-term return is considered, and when γ is 1, the system considers both long-term return and immediate return as equally important; qi' is a function of the Q value of the next state target Critic network; (ii) a s' is the next state entered by taking action a at system state s; a' is a network according to a target Actor in a system state sAn act of selecting;is a parameter of the target Actor network;is a parameter of the target Critic network.
Further, updating the parameters of the target Critic network and the parameters of the target Actor network:
the embodiment adopts an Adaptive motion estimation (Adam) optimization algorithm to update parameters;
wherein, it is required to be noted that: momentum gradient descent part (exponentially weighted average) in Adam optimization algorithm:
vdw=β1vdw+(1-β1)dW
vdb=β1vdb+(1-β1)db
RMSprop section (exponentially weighted average of squared versions) in Adam optimization algorithm:
Sdw=β2Sdw+(1-β2)dW2
Sdb=β2Sdb+(1-β2)db2
wherein, beta1Is the first torch, beta2Is a second torch;
wherein τ controls the update rate, and τ < 1 is usually satisfied.
Still further, an action utility function Q is definedi(s, a) is the expectation of the sum of the rewards subsequently obtained by agents in the ith zone after performing action a in system state s:
Qi(s,a)=E(r(s,a)+γmaxQi(s′,a′))
wherein r (s, a) is the return value after executing action a under the system state s, Qi(s ', a') is the goodness of taking action a 'under system state s';
the decision metric function is then:
s3: the agent is trained by optimizing decision model parameters of the agent using the gradient of the decision metric function.
And optimizing the decision model of the agent in the ith area through the gradient of the decision metric function to finish the training of the agent.
wherein,computing the sign of the gradient for the function; a isiAn action value representing the ith iteration;the gradient is iterated i times for the action utility function,for target Actor networkThe gradient is iterated i times.
S4: and inputting the situation prediction results of different sub-regions and the reactive change rule of the new energy into the intelligent agent, and calculating the voltage control quantity of the power system through the intelligent agent to control the reactive voltage of the power grid.
Based on Newton Raphson power flow calculation, the calculation formula of the voltage control quantity is as follows:
wherein U is the voltage control mass, MiFor modulation of voltage source converters, UdThe fundamental voltage of the dc node.
Example 2
In order to verify and explain the technical effect adopted in the method, the embodiment selects the new energy power station which does not generate output fluctuation and the new energy power station which generates output fluctuation to carry out voltage control comparison test, and compares the test results by a scientific demonstration means to verify the real effect of the method.
(1) Analysis of voltage control result when no output fluctuation occurs in new energy power station
Firstly, analyzing the effect of the invention on the voltage control of the power system under the condition that the output of the new energy power station is relatively stable; under the condition, the active output and the load of each generator set (including a new energy generator set) in the power system are kept near relatively stable values in the whole voltage real-time control process, so that the active output and the load of the generator are considered to be kept unchanged in the process of interaction of each agent and the power system environment, and only the change of the generator terminal voltage caused by excitation regulation of the generator is considered.
The method comprises the steps of generating power system operation state data samples through random sampling, training the agents in two areas by using the first 70% of group operation states of which the node load change ranges are 0.8-1.2 times of rated load, and using the last 30% of group operation states of which the node load change ranges are 0.7-1.3 times of rated load as a verification set of a regression model.
As can be seen from fig. 7 and 8, with the progress of the training process, the loss function of the Actor network of the agent first rises obviously, then falls, and finally converges to a stable value; this shows that the parameter initialization of the neural network is random, and the output of the Actor network cannot effectively regulate the generator terminal voltage in the training early stage, so that the voltage of the power system is out of limit, and the loss function is high; however, with the continuous update of the neural network parameters, after the generator terminal voltage is set according to the output of the Actor network, the voltage level of the power system is effectively controlled, and the loss function is continuously reduced, which shows that the training algorithm provided by the method can effectively train the regression model.
It can be known by comparing the loss function curves between the two agents that the drop speed of the Actor network loss function of agent 1 is significantly faster than that of agent 2, and when the Critic network loss function of agent 1 converges, the fluctuation degree is significantly lower than that of agent 2, which means that the number of nodes in area 1 controlled by agent 1 is less than that in area 2 controlled by agent 2, and the control action of agent 2 can be used for controlling the node voltage of area 1, and the node voltage of area 2 is controlled by agent 2 only, indicating that for the training strategy proposed by the method, the number of nodes is less, the more controllable node voltage units are in the area, and the corresponding agent model is easier to train.
In fig. 9, the gray line in the left graph is the total reward curve obtained by each agent during each screen interaction, the black line is the smoothed total reward curve, and the black dotted points in the right graph represent the required action times when the control voltage of each agent is not exceeded during each screen interaction; therefore, as can be seen from fig. 9, in the training process, the total reward obtained by each agent is continuously increased, and the number of actions required by each agent to control the voltage to be not out-of-limit is continuously reduced, which means that after each agent is continuously trained, the number of actions required to be executed from the control voltage to be not out-of-limit is as small as possible, and when the training is completed and the test is performed, the agent can execute only one or two actions to prevent the voltage from being out-of-limit.
Taking a certain running state in the test, calculating the voltage sum and the average value of each node in each intelligent agent control area before and after control, and displaying the result visually as shown in fig. 10; in fig. 10, dotted dots indicate upper and lower limits of node voltages, scattered gray dots indicate voltages of respective nodes before control, black dotted lines indicate average values of voltages of respective nodes before control, and triangular dotted lines indicate average values of voltages of respective nodes after control; before control, the voltage of each node is integrally higher, and the voltage of each node is higher than the upper limit; after the intelligent agent control, the voltage of each node moves towards the direction of the voltage reference value 1.0, and the average value of the voltage is close to 1.0, which shows that the node voltage is effectively controlled from the out-of-limit.
(2) Voltage control result analysis considering output fluctuation of new energy power station
Under the condition of considering new energy fluctuation, the uncertainty of the output of the new energy unit is enhanced, so in the process of real-time voltage control, the output of the new energy unit should be regarded as a variable quantity, that is, in the process of interaction between each intelligent agent and the power system environment, the active output of the new energy unit is considered to be changed, and meanwhile, the change of the generator terminal voltage caused by generator excitation regulation is considered.
Similarly, the running state data sample of the power system is generated through random sampling, but in the interaction process of the intelligent agent and the environment, the load power of the node 2 and the output of the wind turbine generator of the node 3 need to be dynamically adjusted, and in each step of the interaction, the adjustment range of the randomly generated load and the output of the wind turbine generator is increased to 0.5-1.3 times of the rated power relative to the generation of the sample, so that the uncertainty of stronger output of the new energy turbine generator is reflected.
Fig. 11 and 12 show loss function curves of the Actor network and the Critic network after the new energy output fluctuation is considered, and comparing the convergence conditions of the loss functions in fig. 7 and 8, it can be seen that the drop speed of the loss function of the Actor network is lower than that when the new energy fluctuation is not considered, and the convergence value of the loss function is higher than that when the new energy fluctuation is not considered; the loss function of the criticic network is reduced after being trained for a certain number of times, but the loss function is difficult to converge to a stable value, and the fluctuation with a large amplitude is kept, so that the model is more difficult to train under the condition of considering the fluctuation of new energy.
Fig. 13 shows the number of actions required when the node voltage of each agent control area is not out of limit in consideration of new energy fluctuation, and as can be seen from comparison with fig. 9, the number of control actions required by each agent is relatively more, and can be as high as 50 or more; however, with the progress of the training process, the number of actions required for controlling the voltage not to exceed the limit can be continuously reduced, and finally, the number can be basically controlled to be less than 5, which shows that although the difficulty of model training is higher when the fluctuation of new energy is considered, the model with the control effect can still be obtained through training.
It should be noted that the above-mentioned embodiments are only for illustrating the technical solutions of the present invention and not for limiting, and although the present invention has been described in detail with reference to the preferred embodiments, it should be understood by those skilled in the art that modifications or equivalent substitutions may be made on the technical solutions of the present invention without departing from the spirit and scope of the technical solutions of the present invention, which should be covered by the claims of the present invention.
Claims (7)
1. An AVC intelligent control method based on an artificial neural network and deep reinforcement learning is characterized in that: comprises the steps of (a) preparing a mixture of a plurality of raw materials,
dividing the transformer substation into different sub-control areas by combining the situation prediction result of the reactive load of the power grid and the reactive load change rule of the new energy grid-connected point;
optimizing an action utility function based on a Bellman equation and a minimized loss function, and obtaining a decision metric function by combining the action utility function;
training the agent by optimizing decision model parameters of the agent using the gradient of the decision metric function;
inputting situation prediction results of different sub-control areas and a new energy reactive power change rule into the intelligent agent, and calculating the voltage control quantity of the power system through the intelligent agent to control the reactive voltage of the power grid;
wherein the step of obtaining the situation prediction result comprises:
constructing a deep neural network regression model based on a deep artificial neural network, and integrating a plurality of regression load results of the deep neural network regression model to further obtain a situation prediction result of the reactive load;
the constructing of the deep neural network regression model comprises the following steps:
based on the reactive load data characteristics, considering the climate environment, season, regional distribution, user load and power grid dispatching control strategy, constructing the deep neural network regression model:
wherein k is the order; x is the number of(k)A k-order hidden layer node unit vector is obtained; y is(k)Is a k-order output node vector; u. of(k)Is an input vector of order k;inputting a vector for a k-order feedback state;is a k-order feedback state vector;outputting vectors for k-order hidden layers; omegaiA connection weight matrix of each layer, i is 1, 2, 3, 4, 5, 6; g () is the transfer function of the output neuron; f () is the transfer function of the middle layer neurons;
the deep neural network regression model further includes,
wherein x is(k-1)Is a k-1 order hidden layer node unit vector;is a feedback state vector of k-1 order; u. of(k-1)Is an input vector of order k-1;inputting a vector for a feedback state of k-1 order; y is(k-1)Is a k-1 order output node vector;outputting vectors for k-1 order hidden layer; eta, b,Is a self-feedback gain factor.
2. The AVC intelligent control method based on artificial neural network and deep reinforcement learning of claim 1, wherein: the function for minimizing the loss comprises,
defining the minimization of loss function:
wherein,for taking the independent variable as a training parameterA time minimum loss function, E is an expected value, s is the current system state, s' is the environment state at the next moment, and a is the phaseIn response to the action selected in the state,as a pool of experiences, yiTo pass through the Bellman equation pairAnd estimating the true value.
3. The AVC intelligent control method based on artificial neural network and deep reinforcement learning of claim 2, wherein: the estimated true values include, for example,
wherein r isiThe return value obtained in the ith iteration is obtained; mu is a decision value, gamma represents a decay rate, and gamma is in the range of 0,1];Qi' is a function of the Q value of the next state target Critic network; s' is the next state entered by taking the action a at the system state s; a' is according to the target Actor network in the system state sAn act of selecting;is a parameter of the target Actor network;is a parameter of the target Critic network.
4. The AVC intelligent control method based on artificial neural network and deep reinforcement learning of claim 3, wherein: the parameters of the target Critic network and the parameters of the target Actor network comprise,
where τ controls the update rate.
5. The AVC intelligent control method based on artificial neural network and deep reinforcement learning of claim 4, wherein: the decision metric function includes at least one of,
defining the action utility function Qi(s, a) is the expectation of the sum of the rewards subsequently obtained by the agents in the ith area after the action a is executed in the system state s, and then the decision metric function is:
6. the AVC intelligent control method based on artificial neural network and deep reinforcement learning of claim 5, wherein: the gradient of the decision metric function includes,
7. The AVC intelligent control method based on artificial neural network and deep reinforcement learning of claim 6, wherein: the voltage control amount may include a voltage control amount,
based on Newton Raphson power flow calculation, the calculation formula of the voltage control quantity is as follows:
wherein U is the voltage control quantity, MiFor modulation of voltage source converters, UdThe fundamental voltage of the dc node.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011263523.7A CN112465664B (en) | 2020-11-12 | 2020-11-12 | AVC intelligent control method based on artificial neural network and deep reinforcement learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011263523.7A CN112465664B (en) | 2020-11-12 | 2020-11-12 | AVC intelligent control method based on artificial neural network and deep reinforcement learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112465664A CN112465664A (en) | 2021-03-09 |
CN112465664B true CN112465664B (en) | 2022-05-03 |
Family
ID=74825674
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011263523.7A Active CN112465664B (en) | 2020-11-12 | 2020-11-12 | AVC intelligent control method based on artificial neural network and deep reinforcement learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112465664B (en) |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113283156B (en) * | 2021-03-29 | 2023-09-15 | 北京建筑大学 | Energy-saving control method for subway station air conditioning system based on deep reinforcement learning |
CN112924177B (en) * | 2021-04-02 | 2022-07-19 | 哈尔滨理工大学 | Rolling bearing fault diagnosis method for improved deep Q network |
CN113300379B (en) * | 2021-05-08 | 2022-04-29 | 武汉大学 | Electric power system reactive voltage control method and system based on deep learning |
CN113363997B (en) * | 2021-05-28 | 2022-06-14 | 浙江大学 | Reactive voltage control method based on multi-time scale and multi-agent deep reinforcement learning |
CN113489015B (en) * | 2021-06-17 | 2024-01-26 | 清华大学 | Multi-time-scale reactive voltage control method for power distribution network based on reinforcement learning |
CN113725863A (en) * | 2021-07-30 | 2021-11-30 | 国家电网有限公司 | Power grid autonomous control and decision method and system based on artificial intelligence |
CN114400675B (en) * | 2022-01-21 | 2023-04-07 | 合肥工业大学 | Active power distribution network voltage control method based on weight mean value deep double-Q network |
CN115081702A (en) * | 2022-06-14 | 2022-09-20 | 国网信息通信产业集团有限公司 | Power load prediction method with interpretable characteristic, system and terminal |
Citations (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103490431A (en) * | 2013-09-29 | 2014-01-01 | 华南理工大学 | Power distribution network voltage reactive power optimization method based on BART algorithm |
CN106709820A (en) * | 2017-01-11 | 2017-05-24 | 中国南方电网有限责任公司电网技术研究中心 | Power system load prediction method and device based on deep belief network |
CN106713935A (en) * | 2017-01-09 | 2017-05-24 | 杭州电子科技大学 | Fast method for HEVC (High Efficiency Video Coding) block size partition based on Bayes decision |
CN107257133A (en) * | 2017-06-12 | 2017-10-17 | 浙江群力电气有限公司 | A kind of idle work optimization method, device and AVC systems |
CN107423839A (en) * | 2017-04-17 | 2017-12-01 | 湘潭大学 | A kind of method of the intelligent building microgrid load prediction based on deep learning |
CN107634866A (en) * | 2017-10-27 | 2018-01-26 | 朱秋华 | A kind of distribution network communication system performance estimating method and device |
CN108495129A (en) * | 2018-03-22 | 2018-09-04 | 北京航空航天大学 | The complexity optimized method and device of block partition encoding based on deep learning method |
WO2018187632A1 (en) * | 2017-04-05 | 2018-10-11 | Carnegie Mellon University | Deep learning methods for estimating density and/or flow of objects, and related methods and software |
CN108964023A (en) * | 2018-06-29 | 2018-12-07 | 国网上海市电力公司 | A kind of busbar voltage situation short term prediction method and system for power grid |
CN109343341A (en) * | 2018-11-21 | 2019-02-15 | 北京航天自动控制研究所 | It is a kind of based on deeply study carrier rocket vertically recycle intelligent control method |
CN109698556A (en) * | 2019-02-25 | 2019-04-30 | 深圳市广前电力有限公司 | The control method and logical construction of smart grid AVC substation system interface |
CN110087092A (en) * | 2019-03-11 | 2019-08-02 | 西安电子科技大学 | Low bit-rate video decoding method based on image reconstruction convolutional neural networks |
KR20190109868A (en) * | 2018-03-19 | 2019-09-27 | 삼성전자주식회사 | System and control method of system for processing sound data |
CN110474339A (en) * | 2019-08-07 | 2019-11-19 | 国网福建省电力有限公司 | A kind of electric network reactive-load control method based on the prediction of depth generation load |
CN110535146A (en) * | 2019-08-27 | 2019-12-03 | 哈尔滨工业大学 | The Method for Reactive Power Optimization in Power of Policy-Gradient Reinforcement Learning is determined based on depth |
CN110545416A (en) * | 2019-09-03 | 2019-12-06 | 国家广播电视总局广播电视科学研究院 | ultra-high-definition film source detection method based on deep learning |
CN110676842A (en) * | 2019-09-23 | 2020-01-10 | 南方电网科学研究院有限责任公司 | Power distribution network reconstruction and modeling solving method and device for minimally removing fault area |
CN110738010A (en) * | 2019-10-17 | 2020-01-31 | 湖南科技大学 | Wind power plant short-term wind speed prediction method integrated with deep learning model |
CN110866640A (en) * | 2019-11-11 | 2020-03-06 | 山东科技大学 | Power load prediction method based on deep neural network |
CN110958680A (en) * | 2019-12-09 | 2020-04-03 | 长江师范学院 | Energy efficiency-oriented unmanned aerial vehicle cluster multi-agent deep reinforcement learning optimization method |
CN111130053A (en) * | 2020-01-08 | 2020-05-08 | 华南理工大学 | Power distribution network overcurrent protection method based on deep reinforcement learning |
CN111460650A (en) * | 2020-03-31 | 2020-07-28 | 北京航空航天大学 | Unmanned aerial vehicle end-to-end control method based on deep reinforcement learning |
CN111884213A (en) * | 2020-07-27 | 2020-11-03 | 国网北京市电力公司 | Power distribution network voltage adjusting method based on deep reinforcement learning algorithm |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080291999A1 (en) * | 2007-05-24 | 2008-11-27 | Julien Lerouge | Method and apparatus for video frame marking |
US7802286B2 (en) * | 2007-07-24 | 2010-09-21 | Time Warner Cable Inc. | Methods and apparatus for format selection for network optimization |
US9806991B2 (en) * | 2015-01-21 | 2017-10-31 | Cisco Technology, Inc. | Rendering network policy and monitoring compliance |
-
2020
- 2020-11-12 CN CN202011263523.7A patent/CN112465664B/en active Active
Patent Citations (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103490431A (en) * | 2013-09-29 | 2014-01-01 | 华南理工大学 | Power distribution network voltage reactive power optimization method based on BART algorithm |
CN106713935A (en) * | 2017-01-09 | 2017-05-24 | 杭州电子科技大学 | Fast method for HEVC (High Efficiency Video Coding) block size partition based on Bayes decision |
CN106709820A (en) * | 2017-01-11 | 2017-05-24 | 中国南方电网有限责任公司电网技术研究中心 | Power system load prediction method and device based on deep belief network |
WO2018187632A1 (en) * | 2017-04-05 | 2018-10-11 | Carnegie Mellon University | Deep learning methods for estimating density and/or flow of objects, and related methods and software |
CN107423839A (en) * | 2017-04-17 | 2017-12-01 | 湘潭大学 | A kind of method of the intelligent building microgrid load prediction based on deep learning |
CN107257133A (en) * | 2017-06-12 | 2017-10-17 | 浙江群力电气有限公司 | A kind of idle work optimization method, device and AVC systems |
CN107634866A (en) * | 2017-10-27 | 2018-01-26 | 朱秋华 | A kind of distribution network communication system performance estimating method and device |
KR20190109868A (en) * | 2018-03-19 | 2019-09-27 | 삼성전자주식회사 | System and control method of system for processing sound data |
CN108495129A (en) * | 2018-03-22 | 2018-09-04 | 北京航空航天大学 | The complexity optimized method and device of block partition encoding based on deep learning method |
CN108964023A (en) * | 2018-06-29 | 2018-12-07 | 国网上海市电力公司 | A kind of busbar voltage situation short term prediction method and system for power grid |
CN109343341A (en) * | 2018-11-21 | 2019-02-15 | 北京航天自动控制研究所 | It is a kind of based on deeply study carrier rocket vertically recycle intelligent control method |
CN109698556A (en) * | 2019-02-25 | 2019-04-30 | 深圳市广前电力有限公司 | The control method and logical construction of smart grid AVC substation system interface |
CN110087092A (en) * | 2019-03-11 | 2019-08-02 | 西安电子科技大学 | Low bit-rate video decoding method based on image reconstruction convolutional neural networks |
CN110474339A (en) * | 2019-08-07 | 2019-11-19 | 国网福建省电力有限公司 | A kind of electric network reactive-load control method based on the prediction of depth generation load |
CN110535146A (en) * | 2019-08-27 | 2019-12-03 | 哈尔滨工业大学 | The Method for Reactive Power Optimization in Power of Policy-Gradient Reinforcement Learning is determined based on depth |
CN110545416A (en) * | 2019-09-03 | 2019-12-06 | 国家广播电视总局广播电视科学研究院 | ultra-high-definition film source detection method based on deep learning |
CN110676842A (en) * | 2019-09-23 | 2020-01-10 | 南方电网科学研究院有限责任公司 | Power distribution network reconstruction and modeling solving method and device for minimally removing fault area |
CN110738010A (en) * | 2019-10-17 | 2020-01-31 | 湖南科技大学 | Wind power plant short-term wind speed prediction method integrated with deep learning model |
CN110866640A (en) * | 2019-11-11 | 2020-03-06 | 山东科技大学 | Power load prediction method based on deep neural network |
CN110958680A (en) * | 2019-12-09 | 2020-04-03 | 长江师范学院 | Energy efficiency-oriented unmanned aerial vehicle cluster multi-agent deep reinforcement learning optimization method |
CN111130053A (en) * | 2020-01-08 | 2020-05-08 | 华南理工大学 | Power distribution network overcurrent protection method based on deep reinforcement learning |
CN111460650A (en) * | 2020-03-31 | 2020-07-28 | 北京航空航天大学 | Unmanned aerial vehicle end-to-end control method based on deep reinforcement learning |
CN111884213A (en) * | 2020-07-27 | 2020-11-03 | 国网北京市电力公司 | Power distribution network voltage adjusting method based on deep reinforcement learning algorithm |
Also Published As
Publication number | Publication date |
---|---|
CN112465664A (en) | 2021-03-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112465664B (en) | AVC intelligent control method based on artificial neural network and deep reinforcement learning | |
Lin et al. | An improved moth-flame optimization algorithm for support vector machine prediction of photovoltaic power generation | |
CN102270309B (en) | Short-term electric load prediction method based on ensemble learning | |
CN109787236A (en) | A kind of power system frequency Tendency Prediction method based on deep learning | |
CN113363998B (en) | Power distribution network voltage control method based on multi-agent deep reinforcement learning | |
CN112507614B (en) | Comprehensive optimization method for power grid in distributed power supply high-permeability area | |
CN113554466B (en) | Short-term electricity consumption prediction model construction method, prediction method and device | |
Wan et al. | Data-driven hierarchical optimal allocation of battery energy storage system | |
CN113471982B (en) | Cloud edge cooperation and power grid privacy protection distributed power supply in-situ voltage control method | |
Roukerd et al. | Probabilistic-possibilistic flexibility-based unit commitment with uncertain negawatt demand response resources considering Z-number method | |
CN114362175B (en) | Wind power prediction method and system based on depth certainty strategy gradient algorithm | |
CN109242136A (en) | A kind of micro-capacitance sensor wind power Chaos-Genetic-BP neural network prediction technique | |
CN117973644B (en) | Distributed photovoltaic power virtual acquisition method considering optimization of reference power station | |
CN103618315B (en) | A kind of line voltage idle work optimization method based on BART algorithm and super-absorbent wall | |
CN114784823A (en) | Micro-grid frequency control method and system based on depth certainty strategy gradient | |
Li et al. | Short term prediction of photovoltaic power based on FCM and CG-DBN combination | |
CN118484666B (en) | Energy storage power station evaluation method and system for source network load multi-element application | |
CN115764900A (en) | Distribution network resource aggregation method and system considering congestion mitigation benefit | |
CN112685900B (en) | Power load simulation method for representing impact load power characteristics | |
El-Tamaly et al. | Impact of interconnection photovoltaic/wind system with utility on their reliability using a fuzzy scheme | |
CN109615142A (en) | A kind of wind farm wind velocity combination forecasting method based on wavelet analysis | |
CN112365074A (en) | Artificial intelligence decision-making method based on power grid regulation and control data | |
Ouyang et al. | Compound improved Harris hawks optimization for global and engineering optimization | |
CN111799820A (en) | Double-layer intelligent hybrid zero-star cloud energy storage countermeasure regulation and control method for power system | |
Zicheng et al. | Minimum inertia demand estimation of new power system considering diverse inertial resources based on deep neural network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |