CN106026084A - AGC power dynamic distribution method based on virtual generation tribe - Google Patents
AGC power dynamic distribution method based on virtual generation tribe Download PDFInfo
- Publication number
- CN106026084A CN106026084A CN201610479264.9A CN201610479264A CN106026084A CN 106026084 A CN106026084 A CN 106026084A CN 201610479264 A CN201610479264 A CN 201610479264A CN 106026084 A CN106026084 A CN 106026084A
- Authority
- CN
- China
- Prior art keywords
- clan
- power
- delta
- virtual
- agc
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- H—ELECTRICITY
- H02—GENERATION; CONVERSION OR DISTRIBUTION OF ELECTRIC POWER
- H02J—CIRCUIT ARRANGEMENTS OR SYSTEMS FOR SUPPLYING OR DISTRIBUTING ELECTRIC POWER; SYSTEMS FOR STORING ELECTRIC ENERGY
- H02J3/00—Circuit arrangements for ac mains or ac distribution networks
-
- H—ELECTRICITY
- H02—GENERATION; CONVERSION OR DISTRIBUTION OF ELECTRIC POWER
- H02J—CIRCUIT ARRANGEMENTS OR SYSTEMS FOR SUPPLYING OR DISTRIBUTING ELECTRIC POWER; SYSTEMS FOR STORING ELECTRIC ENERGY
- H02J3/00—Circuit arrangements for ac mains or ac distribution networks
- H02J3/04—Circuit arrangements for ac mains or ac distribution networks for connecting networks of the same frequency but supplied from different sources
-
- H—ELECTRICITY
- H02—GENERATION; CONVERSION OR DISTRIBUTION OF ELECTRIC POWER
- H02J—CIRCUIT ARRANGEMENTS OR SYSTEMS FOR SUPPLYING OR DISTRIBUTING ELECTRIC POWER; SYSTEMS FOR STORING ELECTRIC ENERGY
- H02J3/00—Circuit arrangements for ac mains or ac distribution networks
- H02J3/38—Arrangements for parallely feeding a single network by two or more generators, converters or transformers
- H02J3/46—Controlling of the sharing of output between the generators, converters, or transformers
-
- H—ELECTRICITY
- H02—GENERATION; CONVERSION OR DISTRIBUTION OF ELECTRIC POWER
- H02J—CIRCUIT ARRANGEMENTS OR SYSTEMS FOR SUPPLYING OR DISTRIBUTING ELECTRIC POWER; SYSTEMS FOR STORING ELECTRIC ENERGY
- H02J2203/00—Indexing scheme relating to details of circuit arrangements for AC mains or AC distribution networks
- H02J2203/20—Simulating, e g planning, reliability check, modelling or computer assisted design [CAD]
Abstract
The invention discloses an AGC power dynamic distribution method based on a virtual generation tribe. According to the AGC power dynamic distribution method based on the virtual generation tribe, an automatic generation control (AGC) power distribution framework based on the virtual generation tribe (VGT) is firstly established, and a state discrete set and a motion discrete set are determined; a total real-time power grid generation power instruction of a present control period area is acquired, a value function matrix and selection motion of each virtual generation tribe are initialized by utilizing migration learning; motion selection is carried out according to leader strategy, and a tribe power is calculated; an instant reward function value of each virtual generation tribe is acquired; consistency weight, self-learning weight and a value function matrix of each virtual generation tribe in the present state are updated; whether an algorithm satisfies a convergence condition is determined. Through the method, under the AGC power distribution framework based on the VGT, in combination with QD learning and migration learning, AGC control time requirements can be satisfied, and the method is more suitable for solving an AGC dynamic power distribution problem of a large-scale complex power grid having properties of relatively strong randomness and nondeterminacy.
Description
Technical field
The present invention relates to power system Automatic Generation Control technical field, particularly relate to migrate Q based on harmonious property
The virtual generating clan AGC power dynamic allocation method of study, the method is applicable to interconnected network AGC scattered control system power
Random optimization dynamically distribute.
Background technology
AGC (Automatic Generation Control, Automatic Generation Control) is as EMS (Energy
Management System, EMS) one of important step, it is ensured that the frequency of interconnected network and interconnection are handed over
Change power and be maintained at rated value.In general, AGC is broadly divided into two processes: the 1) tracking of general power instruction, actual electric network is adjusted
Degree center main PI to be used controller, also has scholar to propose the intelligent control method such as fuzzy control, intensified learning;2) general power refers to
The distribution of order, often carries out merit according to engineering experience or in the method for identical variable capacity ratio fixed allocation in real system
Rate is distributed, for reducing regulation expense and improving CPS (Control Performance Standard, control performance standard), remaining
The scholars such as great waves have employed a series of nitrification enhancement such as Q study, multistep backtracking Q (λ) method, improvement layering Q study and carry out merit
Rate dynamically distributes optimization.But, above-mentioned all of allocation algorithm is all centralized optimized algorithm, when AGC unit scale increases,
Its effect of optimization is consequently increased also with decline, convergence time, it is difficult to meet the time scale requirements in 4~16 seconds of AGC.Another
Aspect, centralized optimized algorithm needs to be acquired the service data of each AGC unit, is also easily caused communication blocking.
To this end, applicant " the virtual generating clan harmonious property algorithm that interconnected network AGC power dynamically distributes " (in
State's electrical engineering journal) to propose power based on VGT (Virtual Generation Tribe, virtual generating clan) dynamic
Distribute harmonious property algorithm, respectively with regulation expense and time-to-climb for coherency state variable, efficiently solve AGC power
The Decentralized Autonomous problem of distribution, but its consistency algorithm is simple single order consistency algorithm, and use the distribution of two-layer power,
Stronger to the dependency of Optimized model, it is easier to be absorbed in locally optimal solution.For improving consistency algorithm at dynamic random environment
Adaptability, Moura teaches at " QD Learning:a collaborative distributed strategy for
multi‐agent reinforcement learning through consensus+innovations》(IEEE
Transactions on Signal Processing) carry out consistency algorithm and classical intensified learning Q algorithm highly having
Machine merges, it is proposed that the how virtual clan's intensified learning QD learning algorithm that generates electricity of a kind of brand-new distributing.But, QD learns
As conventional machines learning algorithm, learning new task when, do not utilize learning experience in the past and result so that
When algorithm carries out the new optimization task of enquiry learning every time, will take a substantial amount of time, this also cannot be applied to ultra-short term control
The AGC power distribution of time scale.
In recent years, vaild act and the knowledge of task is optimized for making full use of history, to improve of nitrification enhancement
Practising efficiency, many scholars expand in-depth study to transfer learning.Fachantidis A, Partalas I,
The scholars such as Tsoumakas G are at " Transferring task models in reinforcement learning
Agents " (Neurocomputing) point out that the legacy information that transfer learning is intended to utilize in the past history to optimize task processes newly
Optimization task, wherein the quality of migration performance depends entirely on the similarity of new task and historic task, actually distinct
Optimization task also tends to interrelated.
Summary of the invention
It is an object of the invention to overcome the deficiencies in the prior art, propose a kind of based on the migration Q study of harmonious property
The virtual generating clan AGC power dynamic allocation method of (consensus transfer Q-learning, CTQ).The method exists
Under virtual generating clan control framework, it is adjacent clan in each clan and carries out the interactive consistency calculating of value function matrix
After, clan leader can work in coordination with to self-organizing the generated output of each clan, thus reaches the effect of " Decentralized Autonomous is concentrated and coordinated "
Really;Effectively utilize history optimization information to carry out quick power and dynamically distribute optimization, want meeting the control time scale of AGC
Ask.
The present invention migrates the virtual generating clan AGC power dynamic allocation method of Q study based on harmonious property, including two
Individual process: 1) the general power instruction of regional power grid is assigned to each virtual generating clan VGT;2) by each virtual generating clan
The general power instruction of VGT is assigned to its generating set group.CTQ algorithm is applied to first process and carries out power distribution, second
Assigning process is in large scale due to unit, for improve optimize calculate speed, still use applicant " interconnected network AGC power move
The virtual generating clan harmonious property algorithm of state distribution " consistent the time-to-climb of simple unit in (Proceedings of the CSEE)
Property algorithm carries out power distribution.Based on the transfer of behavior in transfer learning, the present invention proposes a kind of CTQ algorithm, and uses merit
The linear similarity factor of rate distribution optimization task, the AGC power being organically applied to virtual generating clan dynamically distributes, effectively
Solve the distributing optimization problem that complicated large scale electric network AGC power dynamically distributes, reduce the same of AGC unit regulation expense
Time, the control performance standard of regional power grid can be improved.
The purpose of the present invention is achieved through the following technical solutions:
Virtual generating clan AGC power dynamic allocation method based on CTQ, comprises the following steps:
(1) build AGC power distribution frame based on virtual generating clan VGT, determine the leader of virtual generating clan VGT
Person;
(2) state set S and action discrete set A is determinedQ;
(3) real-time total generated output instruction Δ P of current control period regional power grid is gathered∑, it is current state,
Transfer learning is utilized to initialize value function matrix and the selection action of each virtual generating clan;
(4) the policy selection action issued according to leader calculates i-th virtual generating clan VGTiRegulation power Δ
Pi;
(5) by Δ Pi→RiCarry out reward function value mapping calculation, it is thus achieved that the reward function immediately of each virtual generating clan
Value Ri, wherein Δ PiIt is i-th virtual generating clan VGTiRegulation power, RiIt is i-th virtual generating clan VGTiAward
Function;
(6) concordance weight and the self study weight of each virtual generating clan under current state are updated;
(7) each void under current control period is updated according to the award value immediately of each virtual generating clan of current control period
Send out the value function matrix Q of electricity clani;
(8) the value function matrix Q of leader's+1 iteration of kth in current control period is judgedi k+1Value with kth time iteration
Jacobian matrix Qi kTwo norms of difference whether less than an infinitesimal positive number ε, i.e. | | Qi k+1‐Qi k||2<ε;If being unsatisfactory for, then
Returning step (5), if meeting, turning next step;
(9) solve the selection action under current state, and then obtain i-th virtual generating clan VGTiRegulation power
ΔPi;
(10) according to i-th virtual generating clan VGTiRegulation power Δ Pi, by time-to-climb consistency algorithm try to achieve
I-th virtual generating clan VGTiW unit unit regulation power Δ Piw, and when the next one controls cycle arrival,
Return step (3).
AGC power distribution frame based on virtual generating clan in described step 1, under this framework, traditional area electricity
Net is divided into several manor electrical networks according to geographical distribution, and the most virtual generating clan VGT, actually in AGC and power plant
Power controls to increase between (Plant Controller, PLC) a new power generation dispatching and key-course, is by the electrical network of manor
The generating set group that large power plant PLC, actively distribution AGC, microgrid AGC and load control system are constituted.Leader is used between VGT
Person's follower's pattern communicates cooperation, the leader i.e. control centre of regional power grid, is responsible for power disturbance balance;Follower
The dispatching terminal of the most common VGT, main being responsible for interacts collaborative with leader.
The state discrete collection S of described step 2 is to be instructed Δ P by each total generated output∑Constituted as a state
's.
Action discrete set A in described step 2QIt is made up of several combination of actions.When there is y action policy, n
During individual virtual generating clan, action discrete set AQIt is the matrix of a y × n, can be expressed as follows:
AQ=[a1 a2 … ay]=[(λ11,λ12,…,λ1n),
(λ21,λ22,…,λ2n),…,(λy1,λy2,…,λyn)]
Wherein, ayFor action discrete set AQY-th action;λynFor taking y-th action policy is distributed to virtual
The distribution power factor of electricity clan n.
Initial selected action in described step 3 is action discrete set AQMiddle any action.
According to currently optimizing task in described step 3, transfer learning is utilized to initialize the value letter of each virtual generating clan
Before matrix number, need to set up value function source matrix.The inventive method instructs Δ P with total generated output∑Degree of closeness conduct
The relativity evaluation index of Different Optimization task.First, total generated output is instructed Δ P∑Carry out interval range division, interval point
It is not:
Wherein,Represent η total generated output instruction Δ P respectively∑Left and right end points;Z is total generating
Power instruction Δ P∑Interval number.
Gather total generated output instruction Δ P of current control period∑, it is assumed that it falls in the η load disturbance interval, then
The dependency that current optimization task and left and right end points optimize task is respectively as follows:
Wherein, rleft、rrightRepresent current respectively and optimize the relevant of task end points left and right to interval corresponding optimization task
Property, 0≤rleft≤ 1,0≤rright≤ 1, rleft+rright=1.
Therefore the initial value function matrix of the optimization task under current state is:
Qi=rleftQi,left+rrightQi,rightI=1,2 ..., n
In formula: Qi,left、Qi,rightRepresent that virtual generating clan i is at interval left and right end points correspondence optimal value function square respectively
Battle array, can be by learning acquisition (see flow chart 2) in advance.
The policy selection action that leader in described step 4 issues, is greedy strategy π*, as follows:
Wherein, QleaderRepresent the state action value function matrix of leader, skRepresent the state of kth time iteration;A' is
Refer to motion space AQAny one interior action.
I-th virtual generating clan VGT in described step 4iRegulation power Δ PiCan be calculated by equation below:
ΔPi=λyiΔPΣ
Wherein, λyiFor taking y-th action policy is distributed to the distribution power factor of clan of virtual generating clan i, and
MeetConstraint.
Reward function value R immediately in described step 5i(sk,ak,sk+1) can design as follows:
Wherein, μ1、μ2Be respectively regulation the goal of cost and time-to-climb target weight coefficient, and μ1>=0, μ2>=0, adjust
Joint the goal of cost refer to minimize whole control area electrical network in the adjustment cost sum of all AGC units, time-to-climb target
The time-to-climb of being to minimize all units maximum;CiwFor the adjustment cost coefficient of w unit in clan i;ΔPiwExpressed portion
Fall the generated output instruction of w unit in i;Represent the regulations speed of w unit in clan i;N is that clan is the most individual
Number;miThe total number of unit for clan i;ΩiRepresent the set of the clan adjacent with clan i.
Owing to the calculating of reward function value needs first to determine the regulation power of clan's each unit internal, and time-to-climb
Under consistency algorithm, the unit regulation power under different clan's power situations determines that.Therefore, method of least square can be used
Regulation power Δ P to clan iiAnd RiCarry out mapping relations matching, to accelerate the calculating of reward function value.
The concordance weight beta of each the virtual generating clan in described step 6k(sk,ak) and self study weight αk(sk,ak)
It is updated as follows:
Wherein, NO(sk,ak) represent that state action is to (sk,ak) algorithm explore optimizing occurrence number;o1、o2、τ1、τ2
It is respectively normal number, and τ to be met1∈(1/2,1)、0<τ2<τ1‐1/(2+ε1) constraint, wherein ε1It it is an infinitesimal positive number.
The value function matrix Q of each virtual generating clan under current control period in described step 7iUpdate as follows:
Wherein, CQ(sk,ak) represent that virtual generating clan i is at skExecution action a under statekTime with adjacent virtual generate electricity clan
Between harmonious property update item, be also to characterize the maximum different qualities that CTQ learns with single virtual generating clan Q;IQ(sk,
ak) represent that virtual generating clan i is in state skLower execution action akTime self study update item, learn with single virtual generating clan Q
Update mechanism is the same.Wherein, CQ(sk,ak)、IQ(sk,ak) iteration update respectively as follows:
Wherein, sk、sk+1Represent the state of kth time and k+1 iteration respectively;It is that virtual generating clan i is at sk
Execution action a under statekQ-value;ΩiK () is the collection when kth time iteration with the virtual i adjacent virtual generating clan of clan that generates electricity
Close.
Time-to-climb consistency algorithm in described step 10 comprises the following steps, and asks for an interview " the interconnected network of the author in detail
The virtual generating clan harmonious property algorithm that AGC power dynamically distributes " (Proceedings of the CSEE):
1. to i-th virtual generating clan VGTiW the power of the assembling unit time-to-climb tiwGenerating clan virtual with i-th
VGTiPower offset value Δ Perror-iInitialize, wherein
2. gather the real-time running data of current control period regional power grid, instruct Δ P including total generated output∑, i-th
Virtual generating clan VGTiRegulation power Δ PiAnd the real-time active power of output of each unit, and calculate virtual of i-th
Electricity clan VGTiPower offset value Δ Perror-i, simultaneously according to i-th virtual generating clan VGTiRegulation power Δ PiDetermine
Regulations speed directionWherein, regulations speed directionDetermine according to following formula:
In formula:Represent i-th virtual generating clan VGTiThe w AGC unit rising regulation rate limit;Represent i-th virtual generating clan VGTiThe w AGC unit decline regulations speed limit.
3. according to current control period i-th virtual generating clan VGTiThe power of the assembling unit time-to-climb tiwEmpty with i-th
Send out electricity clan VGTiPower offset value Δ Perror-iCarry out concordance calculating;
I-th virtual generating clan VGTiThe power of the w AGC unit time-to-climb consistent update as follows:
Meanwhile, for ensure VGT power-balance, head time-to-climb should update as follows:
Wherein, μiRepresent the power error regulatory factor of clan i, μi> 0.
4. power of the assembling unit Δ P is calculatediw;
Power of the assembling unit Δ PiwCan be calculated by following formula:
5. power of the assembling unit Δ P is judgediwThe most out-of-limit.If power of the assembling unit Δ PiwOut-of-limit, calculate power of the assembling unit Δ the most respectively
PiwWith t time-to-climb of the power of the assembling unitiw, and update row stochastic matrix element;Otherwise, next step is turned;
If power of the assembling unit Δ PiwOut-of-limit, the power Δ P of unitiwAnd time-to-climb tiwCan recalculate as follows:
Wherein,Represent VGT respectivelyiThe minimum and maximum spare capacity of power of w unit.
If power of the assembling unit Δ PiwOut-of-limit, with connection weight all vanishing of unit w, it may be assumed that awj=0, j=1,2, L, mi, with
Shi Jinhang corresponding row stochastic matrix element updates.
6. VGT is updatediPower offset value Δ Perror-i;
7. precision is judged.If | Δ Perror-i| < εi, wherein εiIt is an infinitesimal positive number, then obtains power of the assembling unit Δ Pim,
Otherwise return step 3..
The present invention has such advantages as relative to prior art and effect:
(1) to be that one has the how virtual generating clan of " Decentralized Autonomous, concentrate coordinate " characteristic strong for the CTQ method of the present invention
Change learning algorithm, the control framework of virtual generating clan can be fused to well, fill for information network from now on and energy networks
Intensive energy the Internet AGC decentralized coordinated control is divided to provide new thinking.
(2) the CTQ method of the present invention is after the transfer learning introducing task dependencies, can be quickly carried out the most excellent
Change, ensure the stability of global convergence simultaneously, while the control time scale meeting AGC requires, the control of CPS can be improved
Performance processed, reduces the regulation expense of AGC.
(3) how virtual the CTQ method of the present invention is strong as a kind of generating clan with self study and Cooperative Study ability
Change learning algorithm, be more applicable for solving having stronger randomness, probabilistic large-scale complex power grid AGC power dynamic
Assignment problem.
Accompanying drawing explanation
Fig. 1 is the flow chart of present invention AGC power dynamic allocation method based on virtual generating clan.
Fig. 2 is Guangdong Power Grid VGT communication network topology figure in the present embodiment.
Detailed description of the invention
For being more fully understood that the present invention, below in conjunction with embodiment and accompanying drawing, the invention will be further described, but this
Bright embodiment is not limited to this.
Embodiment
In the present embodiment based on Guangdong Power Grid LOAD FREQUENCY Controlling model, 93 the machine components participating in AGC frequency modulation
Being 6 VGT, wherein the communication network topology of VGT is as in figure 2 it is shown, the relevant parameter of 93 units is referring specifically to table 1.In example
The connection weight a having communication for informationijIt is set to 1.For the void migrating Q study in model based on harmonious property in the present embodiment
Send out electricity clan AGC power dynamic allocation method to comprise the following steps (Fig. 2):
(1) based on Guangdong Power Grid LOAD FREQUENCY Controlling model, according to the location distribution situation in Guangdong, with region
Electrical network inner high voltage interconnection is electrical network border, manor, 93 units participating in AGC frequency modulation is divided into 6 VGT, builds based on VGT
AGC power distribution frame.Wherein, VGT1~VGT6Represent North Guangdong, south, west of Guangdong Province Yuexi, west of Guangdong Province Yuexi, Pearl River Delta, south, East Guangdong and Guangdong respectively
6 manor electrical networks such as east, and with the VGT of hair electricity core the most4As leader.
(2) state discrete collection S and action discrete set A is determinedQ
The state discrete collection S wherein determined in the present embodiment is to be instructed Δ P by each total generated output∑As a shape
State is constituted, unit MW.
The action discrete set A determined in the present embodimentQFor:
A=[(0,0,0,0,0,1), (0.1,0,0,0,0.9) ..., (1,0,0,0,0,0)];
A total of 2568 discrete movement.(action discrete set is exactly the situation of exhaustive all distribution in fact, due to amount of calculation
Problem, the precision of distribution factor gets 0.1)
(3) real-time total generated output that the AGC general power PI controller of current control period regional power grid sends is gathered
Instruction Δ P∑, it is current state, initializes selection action, and according to currently optimizing task, utilize transfer learning to initialize
The value function matrix of each virtual generating clan.
In the present embodiment, initial selected action is action discrete set AQMiddle any action.
According to currently optimizing task, before utilizing transfer learning to initialize the value function matrix of each virtual generating clan,
Need to set up sufficient value function source matrix.The inventive method is appointed using the degree of closeness of general power instruction size as Different Optimization
The relativity evaluation index of business.Interval is respectively as follows:
In the present embodiment, general power instruction size is carried out interval range division, interval be respectively as follows: [1500,
1250), [1250,1000), [1000,750) ..., (750,1000], (1000,1250], (1250,1500] } MW, always
Totally 12 migrate interval, optimize originating task 12 altogether.Wherein, 1500MW is corresponding Guangdong Power Grid maximum single failure (direct current
One pole locking) under load disturbance size.
Gather total generated output instruction Δ P of current control period∑, it is assumed that it falls in the η load disturbance interval, then
The dependency that current optimization task and left and right end points optimize task is respectively as follows:
Wherein, rleft、rrightRepresent current respectively and optimize the relevant of task end points left and right to interval corresponding optimization task
Property, 0≤rleft≤ 1,0≤rright≤ 1, rleft+rright=1.In the present embodiment, the control cycle is 8s.
Therefore the initial value function matrix of the optimization task under current state is:
Qi=rleftQi,left+rrightQi,rightI=1,2 ..., n
Wherein, Qi,left、Qi,rightRepresent that virtual generating clan i is at interval left and right end points correspondence optimal value function square respectively
Battle array, can be by learning acquisition in advance.(see flow chart 2)
(4) the policy selection action issued according to leader calculates clan regulation power Δ Pi。
In the present embodiment, the policy selection action that leader issues, it is greedy strategy π*, as follows:
Wherein, QleaderRepresent the state action value function matrix of leader, skRepresent the state of kth time iteration;A' is
Refer to motion space AQAny one interior action.
In the present embodiment, clan regulation power Δ PiGreedy strategy π issued according to leader*Obtain selection action, and then
Can be calculated by equation below:
ΔPi=λyiΔPΣ
Wherein, λyiFor taking y-th action policy is distributed to the distribution power factor of virtual generating clan i, and to expire
FootConstraint.
Nitrification enhancement is exactly the probability i.e. action policy of action discrete set of exhaustive all of distribution, each action
One group of scale factor is constituted in fact.
(5) by Δ Pi→RiCarry out reward function value mapping calculation, it is thus achieved that the reward function immediately of each virtual generating clan
Value Ri(sk,ak,sk+1)。
In the present embodiment, reward function value R immediatelyi(sk,ak,sk+1) can design as follows:
Wherein, μ1、μ2Be respectively regulation the goal of cost and time-to-climb target weight coefficient, and μ1>=0, μ2>=0, adjust
Joint the goal of cost refer to minimize whole control area electrical network in the adjustment cost sum of all AGC units, time-to-climb target
The time-to-climb of being to minimize all units maximum.In the present embodiment, to regulation the goal of cost and time-to-climb target preference
Unanimously, thus μ1、μ2All it is set to 0.5;CiwFor the adjustment cost coefficient of w unit in clan i;ΔPiwRepresent that in clan i, w is individual
The generated output instruction of unit;Represent the regulations speed of w unit in clan i;N is the total number of clan, the present embodiment
In, n is 6;miThe total number of unit for clan i;ΩiRepresent the set of the clan adjacent with clan i.
Owing to the calculating of reward function value needs first to determine the regulation power of clan's each unit internal, and time-to-climb
Under consistency algorithm, the unit regulation power under different clan's power situations determines that.Therefore, method of least square can be used
To clan power Δ PiAnd RiCarry out mapping relations matching, to accelerate the calculating of reward function value.
(6) concordance weight and the self study weight of each virtual generating clan under current state are updated.
The concordance weight beta of each virtual generating clank(sk,ak) and self study weight αk(sk,ak) be updated as follows:
Wherein, NO(sk,ak) represent that state action is to (sk,ak) algorithm explore optimizing occurrence number;o1、o2、τ1、τ2
It is respectively normal number, and τ to be met1∈(1/2,1)、0<τ2<τ1‐1/(2+ε1) constraint, wherein ε1It it is an infinitesimal positive number.
In the present embodiment, through substantial amounts of simulating, verifying, o1Take 0.2, o2Take 0.8, τ1Take 0.55, τ2Take 0.005.
(7) each void under current control period is updated according to the award value immediately of each virtual generating clan of current control period
Send out the value function matrix Q of electricity clani。
In the present embodiment, the value function matrix Q of each virtual generating clan under current control periodiUpdate as follows:
Wherein, CQRepresent the harmonious property renewal item that virtual generating clan i and adjacent virtual generate electricity between clan, be also
Characterize the maximum different qualities of CTQ and single virtual generating clan Q study;IQRepresent that the self study of virtual generating clan i updates item,
As single virtual generating clan Q study update mechanism.Wherein, CQ、IQIteration update respectively as follows:
Wherein, sk、sk+1Represent the state of kth time and k+1 iteration respectively;It is that virtual generating clan i is at sk
Execution action a under statekQ-value;ΩiK () is the collection when kth time iteration with the virtual i adjacent virtual generating clan of clan that generates electricity
Close.
(8) the value function matrix Q of leader's+1 iteration of kth in current control period is judgedi k+1Value with kth time iteration
Jacobian matrix Qi kTwo norms of difference whether less than an infinitesimal positive number ε, i.e. | | Qi k+1‐Qi k||2<ε.At the present embodiment
In, ε takes 0.001.
(9) solve the selection action under current state, and then obtain clan regulation power Δ Pi。
(10) power Δ P is regulated according to clani, by time-to-climb consistency algorithm try to achieve power of the assembling unit Δ Piw, and
When the next control cycle arrives, return step (3).Time-to-climb the specifically comprising the following steps that of consistency algorithm
1. to i-th virtual generating clan VGTiW the power of the assembling unit time-to-climb tiwGenerating clan virtual with i-th
VGTiPower offset value Δ Perror-iInitialize, wherein
2. gather the real-time running data of current control period regional power grid, instruct Δ P including total generated output∑, i-th
Virtual generating clan VGTiRegulation power Δ PiAnd the real-time active power of output of each unit, and calculate virtual of i-th
Electricity clan VGTiPower offset value Δ Perror-i, simultaneously according to i-th virtual generating clan VGTiRegulation power Δ PiDetermine
Regulations speed direction;
Regulations speed directionDetermine according to following formula:
In formula:Represent i-th virtual generating clan VGTiThe w AGC unit rising regulation rate limit;Represent i-th virtual generating clan VGTiThe w AGC unit decline regulations speed limit, machine in the present embodiment
GroupNumerical value be shown in Table 1.
3. according to current control period i-th virtual generating clan VGTiThe power of the assembling unit time-to-climb tiwEmpty with i-th
Send out electricity clan VGTiPower offset value Δ Perror-iCarry out concordance calculating;
In the present embodiment, it is similar to therewith using VGT1 as object of study, the analysis of other VGT.As shown in table 1, it is assumed that
G1 is the head of VGT1, and other unit is the head of a family and kinsfolk.
The time-to-climb of the power of the w AGC unit of virtual generating clan VGT1, consistent update is as follows:
Meanwhile, for ensure VGT power-balance, head G1 time-to-climb should update as follows:
Wherein, μiRepresent the power error regulatory factor of clan i, μi> 0.In the present embodiment, μiTake 0.001.
4. power of the assembling unit Δ P is calculatediw。
Power of the assembling unit Δ PiwCan be calculated by following formula:
5. power of the assembling unit Δ P is judgediwThe most out-of-limit.If power of the assembling unit Δ PiwOut-of-limit, calculate power of the assembling unit Δ the most respectively
PiwWith t time-to-climb of the power of the assembling unitiw, and update row stochastic matrix element;Otherwise, next step is turned.
If power of the assembling unit Δ PiwOut-of-limit, the power Δ P of unitiwAnd time-to-climb tiwCan recalculate as follows:
Wherein,Represent VGT respectivelyiThe minimum and maximum spare capacity of power of w unit.At this
Unit in embodimentNumerical value be shown in Table 1.
If power of the assembling unit Δ PiwOut-of-limit, with connection weight all vanishing of unit w, it may be assumed that awj=0, j=1,2, L, mi, with
Shi Jinhang corresponding row stochastic matrix element updates.
6. VGT is updatediPower offset value Δ Perror-i;In the present embodiment, VGT1 power offset value is carried out more by definition
New:
7. precision is judged.If | Δ Perror-i| < εi, then power of the assembling unit Δ P is obtainediw, otherwise return step 3..In this reality
Execute in example, take peak power deviation | Δ Perror-1| < 0.1MW is as the condition of convergence of VGT1.
Table 1 Guangdong Power Grid AGC unit parameter statistical table
Above-described embodiment is the present invention preferably embodiment, but embodiments of the present invention are not by above-described embodiment
Limit, the change made under other any spirit without departing from the present invention and principle, modify, substitute, combine, simplify,
All should be the substitute mode of equivalence, within being included in protection scope of the present invention.
Claims (8)
1. an AGC power dynamic allocation method based on virtual generating clan, it is characterised in that comprise the following steps:
(1) build Automation generation control power distribution frame based on virtual generating clan VGT, determine the leader of clan;
(2) state discrete collection S and action discrete set A is determinedQ;
(3) real-time total generated output instruction Δ P of current control period regional power grid is gathered∑For current state, and according to
Currently optimize task, utilize transfer learning to initialize value function matrix and the selection action of each virtual generating clan;
(4) according to leader issue policy selection action and calculate clan regulation power Δ Pi;
(5) power Δ P is regulated by claniReward function value R immediatelyiBetween mapping calculation, it is thus achieved that each virtual Power Generation Section
Reward function value R immediately falleni;
(6) concordance weight and the self study weight of each virtual generating clan under current state are updated;
(7) according under the renewal current control period of award value immediately of each virtual generating clan of current control period each virtual
The value function matrix Q of electricity clani;
(8) the value function matrix Q of leader's+1 iteration of kth in current control period is judgedi k+1Value function with kth time iteration
Matrix Qi kTwo norms of difference whether less than an infinitesimal positive number ε, i.e. | | Qi k+1‐Qi k||2<ε;If being unsatisfactory for, then return
Step (5), if meeting, turns next step;
(9) power Δ P is regulated according to clani, by time-to-climb consistency algorithm try to achieve power of the assembling unit Δ Piw, and at the next one
When the control cycle arrives, return step (3).
AGC power dynamic allocation method based on virtual generating clan the most according to claim 1, it is characterised in that step
(1) described based on virtual generating clan VGT the Automation generation control power distribution frame in, is by traditional area electrical network
Being divided into several manors electrical network VGT according to geographical distribution, manor electrical network VGT is joined by large power plant PLC, active in the electrical network of manor
The generating set group composition that net AGC, microgrid AGC and load control system are constituted;Between the electrical network VGT of manor use leader with
Communicating cooperation with person's pattern, leader is the control centre of regional power grid, is responsible for power disturbance balance;Follower is common
The dispatching terminal of VGT, main being responsible for interacts collaborative with leader.
AGC power dynamic allocation method based on virtual generating clan the most according to claim 1, it is characterised in that described
State discrete collection S in step (2) is to be instructed Δ P by each total generated output∑Constituted as a state.
AGC power dynamic allocation method based on virtual generating clan the most according to claim 1, it is characterised in that described
In step (3) according to currently optimizing task, utilize transfer learning initialize each virtual generating clan value function matrix and
Before selection action, need to set up value function source matrix;Δ P is instructed with total generated output∑The degree of closeness of size is as difference
The relativity evaluation index of optimization task;First, total generated output is instructed Δ P∑Size carries out interval range division, interval point
It is not:
Wherein,Represent η total generated output instruction Δ P respectively∑Interval left and right end points;Z is total generating
Power instruction Δ P∑Interval number;
Gather total generated output instruction Δ P of current control period∑, it is assumed that it falls in the η load disturbance interval, the most currently
The dependency that optimization task optimizes task with left and right end points is respectively as follows:
Wherein, rleft、rrightRepresent respectively and currently optimize the corresponding dependency optimizing task of task end points left and right with interval, 0≤
rleft≤ 1,0≤rright≤ 1, rleft+rright=1;
The initial value function matrix of the optimization task under current state is:
Qi=rleftQi,left+rrightQi,rightI=1,2 ..., n
In formula: Qi,left、Qi,rightRepresent that virtual generating clan i, can at interval left and right end points correspondence optimal value function matrix respectively
By learning acquisition in advance.
AGC power dynamic allocation method based on virtual generating clan the most according to claim 1, it is characterised in that described
Initial selected action in step (3) is the action selected after the most pre-study before CTQ algorithm carries out transfer learning.
AGC power dynamic allocation method based on virtual generating clan the most according to claim 1, it is characterised in that step
(5) described reward function value R immediately ini(sk,ak,sk+1) it is designed as:
Wherein, μ1、μ2Be respectively regulation the goal of cost and time-to-climb target weight coefficient, and μ1>=0, μ2>=0, regulate expense
Target refer to minimize whole control area electrical network in the adjustment cost sum of all AGC units, time-to-climb target be minimum
The time-to-climb of changing all units maximum;CiwFor the adjustment cost coefficient of w unit in clan i;ΔPiwRepresent in clan i
The generated output instruction of w unit;Represent the regulations speed of w unit in clan i;N is the total number of clan;miFor
The total number of unit of clan i;ΩiRepresent the set of the clan adjacent with clan i;Use method of least square to clan power Δ Pi
And RiCarry out mapping relations matching, to accelerate the calculating of reward function value.
AGC power dynamic allocation method based on virtual generating clan the most according to claim 1, it is characterised in that described
The concordance weight beta of each the virtual generating clan in step (6)k(sk,ak) and self study weight αk(sk,ak) be updated as
Under:
Wherein, NO(sk,ak) represent that state action is to (sk,ak) algorithm explore optimizing occurrence number;o1、o2、τ1、τ2Respectively
For normal number, and τ to be met1∈(1/2,1)、0<τ2<τ1‐1/(2+ε1) constraint, wherein ε1It it is an infinitesimal positive number.
AGC power dynamic allocation method based on virtual generating clan the most according to claim 1, it is characterised in that described
The value function matrix Q of each virtual generating clan under current control period in step (7)iUpdate as follows:
Wherein, CQ(sk,ak) represent that virtual generating clan i is at skExecution action a under statekTime and adjacent virtual generating clan between
Harmonious property update item;IQ(sk,ak) represent that virtual generating clan i is in state skLower execution action akTime self study update
, as single virtual generating clan Q study update mechanism;CQ(sk,ak)、IQ(sk,ak) iteration update respectively as follows:
Wherein, sk、sk+1Represent the state of kth time and k+1 iteration respectively;It is that virtual generating clan i is at skState
Lower execution action akQ-value;ΩiK () is the set when kth time iteration with the virtual i adjacent virtual generating clan of clan that generates electricity.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610479264.9A CN106026084B (en) | 2016-06-24 | 2016-06-24 | A kind of AGC power dynamic allocation methods based on virtual power generation clan |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610479264.9A CN106026084B (en) | 2016-06-24 | 2016-06-24 | A kind of AGC power dynamic allocation methods based on virtual power generation clan |
Publications (2)
Publication Number | Publication Date |
---|---|
CN106026084A true CN106026084A (en) | 2016-10-12 |
CN106026084B CN106026084B (en) | 2018-10-09 |
Family
ID=57083887
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610479264.9A Active CN106026084B (en) | 2016-06-24 | 2016-06-24 | A kind of AGC power dynamic allocation methods based on virtual power generation clan |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106026084B (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106779248A (en) * | 2017-01-11 | 2017-05-31 | 南方电网科学研究院有限责任公司 | A kind of Economic Dispatch dispersion Q methods based on limit transfer learning |
CN107045655A (en) * | 2016-12-07 | 2017-08-15 | 三峡大学 | Wolf pack clan strategy process based on the random consistent game of multiple agent and virtual generating clan |
CN107069835A (en) * | 2017-01-25 | 2017-08-18 | 国网冀北电力有限公司电力科学研究院 | New energy power station distribution method active in real time and distributor |
CN107094321A (en) * | 2017-03-31 | 2017-08-25 | 南京邮电大学 | A kind of vehicle-carrying communication MAC layer channel access method learnt based on multiple agent Q |
CN107591847A (en) * | 2017-08-04 | 2018-01-16 | 西安五常电力技术有限公司 | A kind of method that mode using variable element adjusts Hydropower Unit AGC |
CN108320080A (en) * | 2018-01-05 | 2018-07-24 | 上海电力学院 | The real-time dynamic power allocation method in energy internet based on two layers of consistency algorithm |
CN114239439A (en) * | 2021-12-24 | 2022-03-25 | 浙江金乙昌科技股份有限公司 | Automatic filter design method based on tribal algorithm |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104037761A (en) * | 2014-06-25 | 2014-09-10 | 南方电网科学研究院有限责任公司 | AGC power multi-objective random optimization distribution method |
-
2016
- 2016-06-24 CN CN201610479264.9A patent/CN106026084B/en active Active
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104037761A (en) * | 2014-06-25 | 2014-09-10 | 南方电网科学研究院有限责任公司 | AGC power multi-objective random optimization distribution method |
Non-Patent Citations (2)
Title |
---|
SOUMMYA KAR等: "QD-Learning:A Collaborative Distributed Strategy for Multi-agent Reinforcement Learning Through Consensus+Innovations", 《IEEE TRANSACTIONS ON SIGNAL PROCESSING》 * |
张孝顺等: "互联电网AGC功率动态分配的虚拟发电部落协同一致性算法", 《中国电机工程学报》 * |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107045655A (en) * | 2016-12-07 | 2017-08-15 | 三峡大学 | Wolf pack clan strategy process based on the random consistent game of multiple agent and virtual generating clan |
CN106779248A (en) * | 2017-01-11 | 2017-05-31 | 南方电网科学研究院有限责任公司 | A kind of Economic Dispatch dispersion Q methods based on limit transfer learning |
CN107069835A (en) * | 2017-01-25 | 2017-08-18 | 国网冀北电力有限公司电力科学研究院 | New energy power station distribution method active in real time and distributor |
CN107069835B (en) * | 2017-01-25 | 2020-02-18 | 国网冀北电力有限公司电力科学研究院 | Real-time active distribution method and device for new energy power station |
CN107094321A (en) * | 2017-03-31 | 2017-08-25 | 南京邮电大学 | A kind of vehicle-carrying communication MAC layer channel access method learnt based on multiple agent Q |
CN107094321B (en) * | 2017-03-31 | 2020-04-28 | 南京邮电大学 | Multi-agent Q learning-based vehicle-mounted communication MAC layer channel access method |
CN107591847A (en) * | 2017-08-04 | 2018-01-16 | 西安五常电力技术有限公司 | A kind of method that mode using variable element adjusts Hydropower Unit AGC |
CN107591847B (en) * | 2017-08-04 | 2020-05-01 | 西安五常电力技术有限公司 | Method for adjusting Automatic Gain Control (AGC) of hydroelectric generating set by using variable parameter mode |
CN108320080A (en) * | 2018-01-05 | 2018-07-24 | 上海电力学院 | The real-time dynamic power allocation method in energy internet based on two layers of consistency algorithm |
CN108320080B (en) * | 2018-01-05 | 2021-09-07 | 上海电力学院 | Energy internet real-time dynamic power distribution method based on two-layer consistency algorithm |
CN114239439A (en) * | 2021-12-24 | 2022-03-25 | 浙江金乙昌科技股份有限公司 | Automatic filter design method based on tribal algorithm |
Also Published As
Publication number | Publication date |
---|---|
CN106026084B (en) | 2018-10-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112615379B (en) | Power grid multi-section power control method based on distributed multi-agent reinforcement learning | |
CN106026084A (en) | AGC power dynamic distribution method based on virtual generation tribe | |
CN111242443B (en) | Deep reinforcement learning-based economic dispatching method for virtual power plant in energy internet | |
Xi et al. | A wolf pack hunting strategy based virtual tribes control for automatic generation control of smart grid | |
CN103683337B (en) | A kind of interconnected network CPS instruction dynamic assignment optimization method | |
CN104537428B (en) | One kind meter and the probabilistic economical operation appraisal procedure of wind power integration | |
CN105023056B (en) | The optimal carbon energy combined-flow acquisition methods of power grid based on colony intelligence intensified learning | |
CN106682808A (en) | Online rolling optimization scheduling model | |
CN105391090B (en) | A kind of intelligent grid multiple agent multiple target uniformity optimization method | |
CN105389633A (en) | Optimization planning method of substation considering distributed power supplies | |
CN108494022B (en) | Accurate scheduling method based on distributed power supply in micro-grid | |
CN104333047B (en) | Real-time rolling planning method applied to wind power integration of power system | |
CN109149648A (en) | A kind of adaptive width Dynamic Programming intelligent power generation control method | |
CN107092991A (en) | A kind of adaptive economic load dispatching distribution method of intelligent grid | |
CN104536304A (en) | Electric system load multi-agent control method based on Matlab and Netlogo | |
CN107612045A (en) | A kind of Power Plant generated energy intelligent dispensing system and method | |
CN116207739B (en) | Optimal scheduling method and device for power distribution network, computer equipment and storage medium | |
CN105912767B (en) | Multistage power grid distributed collaborative combined calculation method based on B/S framework | |
CN105787650A (en) | Simulation calculation method for Nash equilibrium point of electricity market including multiple load agents | |
CN110474353A (en) | Layer-stepping energy-storage system and its power grid frequency modulation control method for coordinating of participation | |
Liu et al. | Spatial assessment of China’s green governance efficiency in the period of high-quality development | |
CN107392350B (en) | Comprehensive optimization method for power distribution network extension planning containing distributed energy and charging stations | |
CN111259315B (en) | Decentralized scheduling method of multi-subject coordinated pricing mode | |
CN107357988A (en) | Distributed photovoltaic cluster dynamic modelling method based on IEC61850 | |
CN106651136A (en) | Day-ahead power generation plan compilation method of bilateral transaction and apparatus thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |