CN106026084A - AGC power dynamic distribution method based on virtual generation tribe - Google Patents

AGC power dynamic distribution method based on virtual generation tribe Download PDF

Info

Publication number
CN106026084A
CN106026084A CN201610479264.9A CN201610479264A CN106026084A CN 106026084 A CN106026084 A CN 106026084A CN 201610479264 A CN201610479264 A CN 201610479264A CN 106026084 A CN106026084 A CN 106026084A
Authority
CN
China
Prior art keywords
clan
power
delta
virtual
agc
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610479264.9A
Other languages
Chinese (zh)
Other versions
CN106026084B (en
Inventor
张孝顺
李清
徐豪
余涛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
South China University of Technology SCUT
Original Assignee
South China University of Technology SCUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China University of Technology SCUT filed Critical South China University of Technology SCUT
Priority to CN201610479264.9A priority Critical patent/CN106026084B/en
Publication of CN106026084A publication Critical patent/CN106026084A/en
Application granted granted Critical
Publication of CN106026084B publication Critical patent/CN106026084B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H02GENERATION; CONVERSION OR DISTRIBUTION OF ELECTRIC POWER
    • H02JCIRCUIT ARRANGEMENTS OR SYSTEMS FOR SUPPLYING OR DISTRIBUTING ELECTRIC POWER; SYSTEMS FOR STORING ELECTRIC ENERGY
    • H02J3/00Circuit arrangements for ac mains or ac distribution networks
    • HELECTRICITY
    • H02GENERATION; CONVERSION OR DISTRIBUTION OF ELECTRIC POWER
    • H02JCIRCUIT ARRANGEMENTS OR SYSTEMS FOR SUPPLYING OR DISTRIBUTING ELECTRIC POWER; SYSTEMS FOR STORING ELECTRIC ENERGY
    • H02J3/00Circuit arrangements for ac mains or ac distribution networks
    • H02J3/04Circuit arrangements for ac mains or ac distribution networks for connecting networks of the same frequency but supplied from different sources
    • HELECTRICITY
    • H02GENERATION; CONVERSION OR DISTRIBUTION OF ELECTRIC POWER
    • H02JCIRCUIT ARRANGEMENTS OR SYSTEMS FOR SUPPLYING OR DISTRIBUTING ELECTRIC POWER; SYSTEMS FOR STORING ELECTRIC ENERGY
    • H02J3/00Circuit arrangements for ac mains or ac distribution networks
    • H02J3/38Arrangements for parallely feeding a single network by two or more generators, converters or transformers
    • H02J3/46Controlling of the sharing of output between the generators, converters, or transformers
    • HELECTRICITY
    • H02GENERATION; CONVERSION OR DISTRIBUTION OF ELECTRIC POWER
    • H02JCIRCUIT ARRANGEMENTS OR SYSTEMS FOR SUPPLYING OR DISTRIBUTING ELECTRIC POWER; SYSTEMS FOR STORING ELECTRIC ENERGY
    • H02J2203/00Indexing scheme relating to details of circuit arrangements for AC mains or AC distribution networks
    • H02J2203/20Simulating, e g planning, reliability check, modelling or computer assisted design [CAD]

Abstract

The invention discloses an AGC power dynamic distribution method based on a virtual generation tribe. According to the AGC power dynamic distribution method based on the virtual generation tribe, an automatic generation control (AGC) power distribution framework based on the virtual generation tribe (VGT) is firstly established, and a state discrete set and a motion discrete set are determined; a total real-time power grid generation power instruction of a present control period area is acquired, a value function matrix and selection motion of each virtual generation tribe are initialized by utilizing migration learning; motion selection is carried out according to leader strategy, and a tribe power is calculated; an instant reward function value of each virtual generation tribe is acquired; consistency weight, self-learning weight and a value function matrix of each virtual generation tribe in the present state are updated; whether an algorithm satisfies a convergence condition is determined. Through the method, under the AGC power distribution framework based on the VGT, in combination with QD learning and migration learning, AGC control time requirements can be satisfied, and the method is more suitable for solving an AGC dynamic power distribution problem of a large-scale complex power grid having properties of relatively strong randomness and nondeterminacy.

Description

A kind of AGC power dynamic allocation method based on virtual generating clan
Technical field
The present invention relates to power system Automatic Generation Control technical field, particularly relate to migrate Q based on harmonious property The virtual generating clan AGC power dynamic allocation method of study, the method is applicable to interconnected network AGC scattered control system power Random optimization dynamically distribute.
Background technology
AGC (Automatic Generation Control, Automatic Generation Control) is as EMS (Energy Management System, EMS) one of important step, it is ensured that the frequency of interconnected network and interconnection are handed over Change power and be maintained at rated value.In general, AGC is broadly divided into two processes: the 1) tracking of general power instruction, actual electric network is adjusted Degree center main PI to be used controller, also has scholar to propose the intelligent control method such as fuzzy control, intensified learning;2) general power refers to The distribution of order, often carries out merit according to engineering experience or in the method for identical variable capacity ratio fixed allocation in real system Rate is distributed, for reducing regulation expense and improving CPS (Control Performance Standard, control performance standard), remaining The scholars such as great waves have employed a series of nitrification enhancement such as Q study, multistep backtracking Q (λ) method, improvement layering Q study and carry out merit Rate dynamically distributes optimization.But, above-mentioned all of allocation algorithm is all centralized optimized algorithm, when AGC unit scale increases, Its effect of optimization is consequently increased also with decline, convergence time, it is difficult to meet the time scale requirements in 4~16 seconds of AGC.Another Aspect, centralized optimized algorithm needs to be acquired the service data of each AGC unit, is also easily caused communication blocking.
To this end, applicant " the virtual generating clan harmonious property algorithm that interconnected network AGC power dynamically distributes " (in State's electrical engineering journal) to propose power based on VGT (Virtual Generation Tribe, virtual generating clan) dynamic Distribute harmonious property algorithm, respectively with regulation expense and time-to-climb for coherency state variable, efficiently solve AGC power The Decentralized Autonomous problem of distribution, but its consistency algorithm is simple single order consistency algorithm, and use the distribution of two-layer power, Stronger to the dependency of Optimized model, it is easier to be absorbed in locally optimal solution.For improving consistency algorithm at dynamic random environment Adaptability, Moura teaches at " QD Learning:a collaborative distributed strategy for multi‐agent reinforcement learning through consensus+innovations》(IEEE Transactions on Signal Processing) carry out consistency algorithm and classical intensified learning Q algorithm highly having Machine merges, it is proposed that the how virtual clan's intensified learning QD learning algorithm that generates electricity of a kind of brand-new distributing.But, QD learns As conventional machines learning algorithm, learning new task when, do not utilize learning experience in the past and result so that When algorithm carries out the new optimization task of enquiry learning every time, will take a substantial amount of time, this also cannot be applied to ultra-short term control The AGC power distribution of time scale.
In recent years, vaild act and the knowledge of task is optimized for making full use of history, to improve of nitrification enhancement Practising efficiency, many scholars expand in-depth study to transfer learning.Fachantidis A, Partalas I, The scholars such as Tsoumakas G are at " Transferring task models in reinforcement learning Agents " (Neurocomputing) point out that the legacy information that transfer learning is intended to utilize in the past history to optimize task processes newly Optimization task, wherein the quality of migration performance depends entirely on the similarity of new task and historic task, actually distinct Optimization task also tends to interrelated.
Summary of the invention
It is an object of the invention to overcome the deficiencies in the prior art, propose a kind of based on the migration Q study of harmonious property The virtual generating clan AGC power dynamic allocation method of (consensus transfer Q-learning, CTQ).The method exists Under virtual generating clan control framework, it is adjacent clan in each clan and carries out the interactive consistency calculating of value function matrix After, clan leader can work in coordination with to self-organizing the generated output of each clan, thus reaches the effect of " Decentralized Autonomous is concentrated and coordinated " Really;Effectively utilize history optimization information to carry out quick power and dynamically distribute optimization, want meeting the control time scale of AGC Ask.
The present invention migrates the virtual generating clan AGC power dynamic allocation method of Q study based on harmonious property, including two Individual process: 1) the general power instruction of regional power grid is assigned to each virtual generating clan VGT;2) by each virtual generating clan The general power instruction of VGT is assigned to its generating set group.CTQ algorithm is applied to first process and carries out power distribution, second Assigning process is in large scale due to unit, for improve optimize calculate speed, still use applicant " interconnected network AGC power move The virtual generating clan harmonious property algorithm of state distribution " consistent the time-to-climb of simple unit in (Proceedings of the CSEE) Property algorithm carries out power distribution.Based on the transfer of behavior in transfer learning, the present invention proposes a kind of CTQ algorithm, and uses merit The linear similarity factor of rate distribution optimization task, the AGC power being organically applied to virtual generating clan dynamically distributes, effectively Solve the distributing optimization problem that complicated large scale electric network AGC power dynamically distributes, reduce the same of AGC unit regulation expense Time, the control performance standard of regional power grid can be improved.
The purpose of the present invention is achieved through the following technical solutions:
Virtual generating clan AGC power dynamic allocation method based on CTQ, comprises the following steps:
(1) build AGC power distribution frame based on virtual generating clan VGT, determine the leader of virtual generating clan VGT Person;
(2) state set S and action discrete set A is determinedQ
(3) real-time total generated output instruction Δ P of current control period regional power grid is gathered, it is current state, Transfer learning is utilized to initialize value function matrix and the selection action of each virtual generating clan;
(4) the policy selection action issued according to leader calculates i-th virtual generating clan VGTiRegulation power Δ Pi
(5) by Δ Pi→RiCarry out reward function value mapping calculation, it is thus achieved that the reward function immediately of each virtual generating clan Value Ri, wherein Δ PiIt is i-th virtual generating clan VGTiRegulation power, RiIt is i-th virtual generating clan VGTiAward Function;
(6) concordance weight and the self study weight of each virtual generating clan under current state are updated;
(7) each void under current control period is updated according to the award value immediately of each virtual generating clan of current control period Send out the value function matrix Q of electricity clani
(8) the value function matrix Q of leader's+1 iteration of kth in current control period is judgedi k+1Value with kth time iteration Jacobian matrix Qi kTwo norms of difference whether less than an infinitesimal positive number ε, i.e. | | Qi k+1‐Qi k||2<ε;If being unsatisfactory for, then Returning step (5), if meeting, turning next step;
(9) solve the selection action under current state, and then obtain i-th virtual generating clan VGTiRegulation power ΔPi
(10) according to i-th virtual generating clan VGTiRegulation power Δ Pi, by time-to-climb consistency algorithm try to achieve I-th virtual generating clan VGTiW unit unit regulation power Δ Piw, and when the next one controls cycle arrival, Return step (3).
AGC power distribution frame based on virtual generating clan in described step 1, under this framework, traditional area electricity Net is divided into several manor electrical networks according to geographical distribution, and the most virtual generating clan VGT, actually in AGC and power plant Power controls to increase between (Plant Controller, PLC) a new power generation dispatching and key-course, is by the electrical network of manor The generating set group that large power plant PLC, actively distribution AGC, microgrid AGC and load control system are constituted.Leader is used between VGT Person's follower's pattern communicates cooperation, the leader i.e. control centre of regional power grid, is responsible for power disturbance balance;Follower The dispatching terminal of the most common VGT, main being responsible for interacts collaborative with leader.
The state discrete collection S of described step 2 is to be instructed Δ P by each total generated outputConstituted as a state 's.
Action discrete set A in described step 2QIt is made up of several combination of actions.When there is y action policy, n During individual virtual generating clan, action discrete set AQIt is the matrix of a y × n, can be expressed as follows:
AQ=[a1 a2 … ay]=[(λ1112,…,λ1n),
2122,…,λ2n),…,(λy1y2,…,λyn)]
Wherein, ayFor action discrete set AQY-th action;λynFor taking y-th action policy is distributed to virtual The distribution power factor of electricity clan n.
Initial selected action in described step 3 is action discrete set AQMiddle any action.
According to currently optimizing task in described step 3, transfer learning is utilized to initialize the value letter of each virtual generating clan Before matrix number, need to set up value function source matrix.The inventive method instructs Δ P with total generated outputDegree of closeness conduct The relativity evaluation index of Different Optimization task.First, total generated output is instructed Δ PCarry out interval range division, interval point It is not:
{ &lsqb; &Delta;P l e f t 1 , &Delta;P r i g h t 1 ) , ... , &lsqb; &Delta;P l e f t &eta; , &Delta;P r i g h t &eta; ) , ... , &lsqb; &Delta;P l e f t Z , &Delta;P r i g h t Z &rsqb; }
Wherein,Represent η total generated output instruction Δ P respectivelyLeft and right end points;Z is total generating Power instruction Δ PInterval number.
Gather total generated output instruction Δ P of current control period, it is assumed that it falls in the η load disturbance interval, then The dependency that current optimization task and left and right end points optimize task is respectively as follows:
r l e f t = &Delta;P r i g h t &eta; - &Delta;P &Sigma; &Delta;P r i g h t &eta; - &Delta;P l e f t &eta; r r i g h t = &Delta;P &Sigma; - &Delta;P l e f t &eta; &Delta;P r i g h t &eta; - &Delta;P l e f t &eta;
Wherein, rleft、rrightRepresent current respectively and optimize the relevant of task end points left and right to interval corresponding optimization task Property, 0≤rleft≤ 1,0≤rright≤ 1, rleft+rright=1.
Therefore the initial value function matrix of the optimization task under current state is:
Qi=rleftQi,left+rrightQi,rightI=1,2 ..., n
In formula: Qi,left、Qi,rightRepresent that virtual generating clan i is at interval left and right end points correspondence optimal value function square respectively Battle array, can be by learning acquisition (see flow chart 2) in advance.
The policy selection action that leader in described step 4 issues, is greedy strategy π*, as follows:
&pi; * ( s k ) = arg m a x a &prime; &Element; A Q Q l e a d e r k ( s k , a &prime; )
Wherein, QleaderRepresent the state action value function matrix of leader, skRepresent the state of kth time iteration;A' is Refer to motion space AQAny one interior action.
I-th virtual generating clan VGT in described step 4iRegulation power Δ PiCan be calculated by equation below:
ΔPiyiΔPΣ
Wherein, λyiFor taking y-th action policy is distributed to the distribution power factor of clan of virtual generating clan i, and MeetConstraint.
Reward function value R immediately in described step 5i(sk,ak,sk+1) can design as follows:
R i ( s k , a k , s k + 1 ) = - &mu; 1 &Sigma; w = 1 m i C i w &Delta;P i w - &mu; 2 max j = i &cup; j &Element; &Omega; i ( min max w = 1 , 2 , K , m j ( &Delta;P j w / &Delta;P j w r a t e ) )
Wherein, μ1、μ2Be respectively regulation the goal of cost and time-to-climb target weight coefficient, and μ1>=0, μ2>=0, adjust Joint the goal of cost refer to minimize whole control area electrical network in the adjustment cost sum of all AGC units, time-to-climb target The time-to-climb of being to minimize all units maximum;CiwFor the adjustment cost coefficient of w unit in clan i;ΔPiwExpressed portion Fall the generated output instruction of w unit in i;Represent the regulations speed of w unit in clan i;N is that clan is the most individual Number;miThe total number of unit for clan i;ΩiRepresent the set of the clan adjacent with clan i.
Owing to the calculating of reward function value needs first to determine the regulation power of clan's each unit internal, and time-to-climb Under consistency algorithm, the unit regulation power under different clan's power situations determines that.Therefore, method of least square can be used Regulation power Δ P to clan iiAnd RiCarry out mapping relations matching, to accelerate the calculating of reward function value.
The concordance weight beta of each the virtual generating clan in described step 6k(sk,ak) and self study weight αk(sk,ak) It is updated as follows:
&beta; k ( s k , a k ) = o 2 ( N O ( s k , a k ) + 1 ) &tau; 2
&alpha; k ( s k , a k ) = o 1 ( N O ( s k , a k ) + 1 ) &tau; 1
Wherein, NO(sk,ak) represent that state action is to (sk,ak) algorithm explore optimizing occurrence number;o1、o2、τ1、τ2 It is respectively normal number, and τ to be met1∈(1/2,1)、0<τ21‐1/(2+ε1) constraint, wherein ε1It it is an infinitesimal positive number.
The value function matrix Q of each virtual generating clan under current control period in described step 7iUpdate as follows:
Q i k + 1 ( s k , a k ) = Q i k ( s k , a k ) - C Q ( s k , a k ) + I Q ( s k , a k )
Wherein, CQ(sk,ak) represent that virtual generating clan i is at skExecution action a under statekTime with adjacent virtual generate electricity clan Between harmonious property update item, be also to characterize the maximum different qualities that CTQ learns with single virtual generating clan Q;IQ(sk, ak) represent that virtual generating clan i is in state skLower execution action akTime self study update item, learn with single virtual generating clan Q Update mechanism is the same.Wherein, CQ(sk,ak)、IQ(sk,ak) iteration update respectively as follows:
C Q ( s k , a k ) = &beta; k ( s k , a k ) &Sigma; j &Element; &Omega; i ( k ) ( Q i k ( s k , a k ) - Q j k ( s k , a k ) )
I Q ( s k , a k ) = &alpha; k ( s k , a k ) &lsqb; R i ( s k , a k , s k + 1 ) + &gamma; max a &prime; &Element; A Q Q i k ( s k + 1 , a &prime; ) - Q i k ( s k , a k ) &rsqb;
Wherein, sk、sk+1Represent the state of kth time and k+1 iteration respectively;It is that virtual generating clan i is at sk Execution action a under statekQ-value;ΩiK () is the collection when kth time iteration with the virtual i adjacent virtual generating clan of clan that generates electricity Close.
Time-to-climb consistency algorithm in described step 10 comprises the following steps, and asks for an interview " the interconnected network of the author in detail The virtual generating clan harmonious property algorithm that AGC power dynamically distributes " (Proceedings of the CSEE):
1. to i-th virtual generating clan VGTiW the power of the assembling unit time-to-climb tiwGenerating clan virtual with i-th VGTiPower offset value Δ Perror-iInitialize, wherein
2. gather the real-time running data of current control period regional power grid, instruct Δ P including total generated output, i-th Virtual generating clan VGTiRegulation power Δ PiAnd the real-time active power of output of each unit, and calculate virtual of i-th Electricity clan VGTiPower offset value Δ Perror-i, simultaneously according to i-th virtual generating clan VGTiRegulation power Δ PiDetermine Regulations speed directionWherein, regulations speed directionDetermine according to following formula:
&Delta;P i w r a t e = P i w r a t e + i f &Delta;P i > 0 P i w r a t e - i f &Delta;P i < 0
In formula:Represent i-th virtual generating clan VGTiThe w AGC unit rising regulation rate limit;Represent i-th virtual generating clan VGTiThe w AGC unit decline regulations speed limit.
3. according to current control period i-th virtual generating clan VGTiThe power of the assembling unit time-to-climb tiwEmpty with i-th Send out electricity clan VGTiPower offset value Δ Perror-iCarry out concordance calculating;
I-th virtual generating clan VGTiThe power of the w AGC unit time-to-climb consistent update as follows:
t i w &lsqb; k + 1 &rsqb; = &Sigma; j = 1 m i d j w &lsqb; k &rsqb; t i j &lsqb; k &rsqb;
Meanwhile, for ensure VGT power-balance, head time-to-climb should update as follows:
t i w &lsqb; k + 1 &rsqb; = &Sigma; j = 1 m i d j w &lsqb; k &rsqb; t i j &lsqb; k &rsqb; + &mu; i &Delta;P e r r o r - i , i f &Delta;P i > 0 &Sigma; m i m i d j w &lsqb; k &rsqb; t i j &lsqb; k &rsqb; - &mu; i &Delta;P e r r o r - i , i f &Delta;P i < 0
Wherein, μiRepresent the power error regulatory factor of clan i, μi> 0.
4. power of the assembling unit Δ P is calculatediw
Power of the assembling unit Δ PiwCan be calculated by following formula:
t i w = &Delta;P i w / &Delta;P i w r a t e
5. power of the assembling unit Δ P is judgediwThe most out-of-limit.If power of the assembling unit Δ PiwOut-of-limit, calculate power of the assembling unit Δ the most respectively PiwWith t time-to-climb of the power of the assembling unitiw, and update row stochastic matrix element;Otherwise, next step is turned;
If power of the assembling unit Δ PiwOut-of-limit, the power Δ P of unitiwAnd time-to-climb tiwCan recalculate as follows:
&Delta;P i w = &Delta;P i w max i f &Delta;P i w > &Delta;P i w max &Delta;P i w min i f &Delta;P i w < &Delta;P i w min
t i w = t i w max = &Delta;P i w max / P i w r a t e + i f &Delta;P i w > &Delta;P i w max &Delta;P i w min / P i w r a t e - i f &Delta;P i w < &Delta;P i w min
Wherein,Represent VGT respectivelyiThe minimum and maximum spare capacity of power of w unit.
If power of the assembling unit Δ PiwOut-of-limit, with connection weight all vanishing of unit w, it may be assumed that awj=0, j=1,2, L, mi, with Shi Jinhang corresponding row stochastic matrix element updates.
6. VGT is updatediPower offset value Δ Perror-i
7. precision is judged.If | Δ Perror-i| < εi, wherein εiIt is an infinitesimal positive number, then obtains power of the assembling unit Δ Pim, Otherwise return step 3..
The present invention has such advantages as relative to prior art and effect:
(1) to be that one has the how virtual generating clan of " Decentralized Autonomous, concentrate coordinate " characteristic strong for the CTQ method of the present invention Change learning algorithm, the control framework of virtual generating clan can be fused to well, fill for information network from now on and energy networks Intensive energy the Internet AGC decentralized coordinated control is divided to provide new thinking.
(2) the CTQ method of the present invention is after the transfer learning introducing task dependencies, can be quickly carried out the most excellent Change, ensure the stability of global convergence simultaneously, while the control time scale meeting AGC requires, the control of CPS can be improved Performance processed, reduces the regulation expense of AGC.
(3) how virtual the CTQ method of the present invention is strong as a kind of generating clan with self study and Cooperative Study ability Change learning algorithm, be more applicable for solving having stronger randomness, probabilistic large-scale complex power grid AGC power dynamic Assignment problem.
Accompanying drawing explanation
Fig. 1 is the flow chart of present invention AGC power dynamic allocation method based on virtual generating clan.
Fig. 2 is Guangdong Power Grid VGT communication network topology figure in the present embodiment.
Detailed description of the invention
For being more fully understood that the present invention, below in conjunction with embodiment and accompanying drawing, the invention will be further described, but this Bright embodiment is not limited to this.
Embodiment
In the present embodiment based on Guangdong Power Grid LOAD FREQUENCY Controlling model, 93 the machine components participating in AGC frequency modulation Being 6 VGT, wherein the communication network topology of VGT is as in figure 2 it is shown, the relevant parameter of 93 units is referring specifically to table 1.In example The connection weight a having communication for informationijIt is set to 1.For the void migrating Q study in model based on harmonious property in the present embodiment Send out electricity clan AGC power dynamic allocation method to comprise the following steps (Fig. 2):
(1) based on Guangdong Power Grid LOAD FREQUENCY Controlling model, according to the location distribution situation in Guangdong, with region Electrical network inner high voltage interconnection is electrical network border, manor, 93 units participating in AGC frequency modulation is divided into 6 VGT, builds based on VGT AGC power distribution frame.Wherein, VGT1~VGT6Represent North Guangdong, south, west of Guangdong Province Yuexi, west of Guangdong Province Yuexi, Pearl River Delta, south, East Guangdong and Guangdong respectively 6 manor electrical networks such as east, and with the VGT of hair electricity core the most4As leader.
(2) state discrete collection S and action discrete set A is determinedQ
The state discrete collection S wherein determined in the present embodiment is to be instructed Δ P by each total generated outputAs a shape State is constituted, unit MW.
The action discrete set A determined in the present embodimentQFor:
A=[(0,0,0,0,0,1), (0.1,0,0,0,0.9) ..., (1,0,0,0,0,0)];
A total of 2568 discrete movement.(action discrete set is exactly the situation of exhaustive all distribution in fact, due to amount of calculation Problem, the precision of distribution factor gets 0.1)
(3) real-time total generated output that the AGC general power PI controller of current control period regional power grid sends is gathered Instruction Δ P, it is current state, initializes selection action, and according to currently optimizing task, utilize transfer learning to initialize The value function matrix of each virtual generating clan.
In the present embodiment, initial selected action is action discrete set AQMiddle any action.
According to currently optimizing task, before utilizing transfer learning to initialize the value function matrix of each virtual generating clan, Need to set up sufficient value function source matrix.The inventive method is appointed using the degree of closeness of general power instruction size as Different Optimization The relativity evaluation index of business.Interval is respectively as follows:
{ &lsqb; &Delta;P l e f t 1 , &Delta;P r i g h t 1 ) , ... , &lsqb; &Delta;P l e f t &eta; , &Delta;P r i g h t &eta; ) , ... , &lsqb; &Delta;P l e f t Z , &Delta;P r i g h t Z &rsqb; }
In the present embodiment, general power instruction size is carried out interval range division, interval be respectively as follows: [1500, 1250), [1250,1000), [1000,750) ..., (750,1000], (1000,1250], (1250,1500] } MW, always Totally 12 migrate interval, optimize originating task 12 altogether.Wherein, 1500MW is corresponding Guangdong Power Grid maximum single failure (direct current One pole locking) under load disturbance size.
Gather total generated output instruction Δ P of current control period, it is assumed that it falls in the η load disturbance interval, then The dependency that current optimization task and left and right end points optimize task is respectively as follows:
r l e f t = &Delta;P r i g h t &eta; - &Delta;P &Sigma; &Delta;P r i g h t &eta; - &Delta;P l e f t &eta; r r i g h t = &Delta;P &Sigma; - &Delta;P l e f t &eta; &Delta;P r i g h t &eta; - &Delta;P l e f t &eta;
Wherein, rleft、rrightRepresent current respectively and optimize the relevant of task end points left and right to interval corresponding optimization task Property, 0≤rleft≤ 1,0≤rright≤ 1, rleft+rright=1.In the present embodiment, the control cycle is 8s.
Therefore the initial value function matrix of the optimization task under current state is:
Qi=rleftQi,left+rrightQi,rightI=1,2 ..., n
Wherein, Qi,left、Qi,rightRepresent that virtual generating clan i is at interval left and right end points correspondence optimal value function square respectively Battle array, can be by learning acquisition in advance.(see flow chart 2)
(4) the policy selection action issued according to leader calculates clan regulation power Δ Pi
In the present embodiment, the policy selection action that leader issues, it is greedy strategy π*, as follows:
&pi; * ( s k ) = arg m a x a &prime; &Element; A Q Q l e a d e r k ( s k , a &prime; )
Wherein, QleaderRepresent the state action value function matrix of leader, skRepresent the state of kth time iteration;A' is Refer to motion space AQAny one interior action.
In the present embodiment, clan regulation power Δ PiGreedy strategy π issued according to leader*Obtain selection action, and then Can be calculated by equation below:
ΔPiyiΔPΣ
Wherein, λyiFor taking y-th action policy is distributed to the distribution power factor of virtual generating clan i, and to expire FootConstraint.
Nitrification enhancement is exactly the probability i.e. action policy of action discrete set of exhaustive all of distribution, each action One group of scale factor is constituted in fact.
(5) by Δ Pi→RiCarry out reward function value mapping calculation, it is thus achieved that the reward function immediately of each virtual generating clan Value Ri(sk,ak,sk+1)。
In the present embodiment, reward function value R immediatelyi(sk,ak,sk+1) can design as follows:
R i ( s k , a k , s k + 1 ) = - &mu; 1 &Sigma; w = 1 m i C i w &Delta;P i w - &mu; 2 max j = i &cup; j &Element; &Omega; i ( min max w = 1 , 2 , K , m j ( &Delta;P j w / &Delta;P j w r a t e ) )
Wherein, μ1、μ2Be respectively regulation the goal of cost and time-to-climb target weight coefficient, and μ1>=0, μ2>=0, adjust Joint the goal of cost refer to minimize whole control area electrical network in the adjustment cost sum of all AGC units, time-to-climb target The time-to-climb of being to minimize all units maximum.In the present embodiment, to regulation the goal of cost and time-to-climb target preference Unanimously, thus μ1、μ2All it is set to 0.5;CiwFor the adjustment cost coefficient of w unit in clan i;ΔPiwRepresent that in clan i, w is individual The generated output instruction of unit;Represent the regulations speed of w unit in clan i;N is the total number of clan, the present embodiment In, n is 6;miThe total number of unit for clan i;ΩiRepresent the set of the clan adjacent with clan i.
Owing to the calculating of reward function value needs first to determine the regulation power of clan's each unit internal, and time-to-climb Under consistency algorithm, the unit regulation power under different clan's power situations determines that.Therefore, method of least square can be used To clan power Δ PiAnd RiCarry out mapping relations matching, to accelerate the calculating of reward function value.
(6) concordance weight and the self study weight of each virtual generating clan under current state are updated.
The concordance weight beta of each virtual generating clank(sk,ak) and self study weight αk(sk,ak) be updated as follows:
&beta; k ( s k , a k ) = o 2 ( N O ( s k , a k ) + 1 ) &tau; 2
&alpha; k ( s k , a k ) = o 1 ( N O ( s k , a k ) + 1 ) &tau; 1
Wherein, NO(sk,ak) represent that state action is to (sk,ak) algorithm explore optimizing occurrence number;o1、o2、τ1、τ2 It is respectively normal number, and τ to be met1∈(1/2,1)、0<τ21‐1/(2+ε1) constraint, wherein ε1It it is an infinitesimal positive number. In the present embodiment, through substantial amounts of simulating, verifying, o1Take 0.2, o2Take 0.8, τ1Take 0.55, τ2Take 0.005.
(7) each void under current control period is updated according to the award value immediately of each virtual generating clan of current control period Send out the value function matrix Q of electricity clani
In the present embodiment, the value function matrix Q of each virtual generating clan under current control periodiUpdate as follows:
Q i k + 1 ( s k , a k ) = Q i k ( s k , a k ) - C Q ( s k , a k ) + I Q ( s k , a k )
Wherein, CQRepresent the harmonious property renewal item that virtual generating clan i and adjacent virtual generate electricity between clan, be also Characterize the maximum different qualities of CTQ and single virtual generating clan Q study;IQRepresent that the self study of virtual generating clan i updates item, As single virtual generating clan Q study update mechanism.Wherein, CQ、IQIteration update respectively as follows:
C Q ( s k , a k ) = &beta; k ( s k , a k ) &Sigma; j &Element; &Omega; i ( k ) ( Q i k ( s k , a k ) - Q j k ( s k , a k ) )
I Q ( s k , a k ) = &alpha; k ( s k , a k ) &lsqb; R i ( s k , a k , s k + 1 ) + &gamma; max a &prime; &Element; A Q Q i k ( s k + 1 , a &prime; ) - Q i k ( s k , a k ) &rsqb;
Wherein, sk、sk+1Represent the state of kth time and k+1 iteration respectively;It is that virtual generating clan i is at sk Execution action a under statekQ-value;ΩiK () is the collection when kth time iteration with the virtual i adjacent virtual generating clan of clan that generates electricity Close.
(8) the value function matrix Q of leader's+1 iteration of kth in current control period is judgedi k+1Value with kth time iteration Jacobian matrix Qi kTwo norms of difference whether less than an infinitesimal positive number ε, i.e. | | Qi k+1‐Qi k||2<ε.At the present embodiment In, ε takes 0.001.
(9) solve the selection action under current state, and then obtain clan regulation power Δ Pi
(10) power Δ P is regulated according to clani, by time-to-climb consistency algorithm try to achieve power of the assembling unit Δ Piw, and When the next control cycle arrives, return step (3).Time-to-climb the specifically comprising the following steps that of consistency algorithm
1. to i-th virtual generating clan VGTiW the power of the assembling unit time-to-climb tiwGenerating clan virtual with i-th VGTiPower offset value Δ Perror-iInitialize, wherein
2. gather the real-time running data of current control period regional power grid, instruct Δ P including total generated output, i-th Virtual generating clan VGTiRegulation power Δ PiAnd the real-time active power of output of each unit, and calculate virtual of i-th Electricity clan VGTiPower offset value Δ Perror-i, simultaneously according to i-th virtual generating clan VGTiRegulation power Δ PiDetermine Regulations speed direction;
Regulations speed directionDetermine according to following formula:
&Delta;P i w r a t e = P i w r a t e + i f &Delta;P i > 0 P i w r a t e - i f &Delta;P i < 0
In formula:Represent i-th virtual generating clan VGTiThe w AGC unit rising regulation rate limit;Represent i-th virtual generating clan VGTiThe w AGC unit decline regulations speed limit, machine in the present embodiment GroupNumerical value be shown in Table 1.
3. according to current control period i-th virtual generating clan VGTiThe power of the assembling unit time-to-climb tiwEmpty with i-th Send out electricity clan VGTiPower offset value Δ Perror-iCarry out concordance calculating;
In the present embodiment, it is similar to therewith using VGT1 as object of study, the analysis of other VGT.As shown in table 1, it is assumed that G1 is the head of VGT1, and other unit is the head of a family and kinsfolk.
The time-to-climb of the power of the w AGC unit of virtual generating clan VGT1, consistent update is as follows:
t i w &lsqb; k + 1 &rsqb; = &Sigma; j = 1 m i d j w &lsqb; k &rsqb; t i j &lsqb; k &rsqb;
Meanwhile, for ensure VGT power-balance, head G1 time-to-climb should update as follows:
t i w &lsqb; k + 1 &rsqb; = &Sigma; j = 1 m i d j w &lsqb; k &rsqb; t i j &lsqb; k &rsqb; + &mu; i &Delta;P e r r o r - i , i f &Delta;P i > 0 &Sigma; m i m i d j w &lsqb; k &rsqb; t i j &lsqb; k &rsqb; - &mu; i &Delta;P e r r o r - i , i f &Delta;P i < 0
Wherein, μiRepresent the power error regulatory factor of clan i, μi> 0.In the present embodiment, μiTake 0.001.
4. power of the assembling unit Δ P is calculatediw
Power of the assembling unit Δ PiwCan be calculated by following formula:
t i w = &Delta;P i w / &Delta;P i w r a t e
5. power of the assembling unit Δ P is judgediwThe most out-of-limit.If power of the assembling unit Δ PiwOut-of-limit, calculate power of the assembling unit Δ the most respectively PiwWith t time-to-climb of the power of the assembling unitiw, and update row stochastic matrix element;Otherwise, next step is turned.
If power of the assembling unit Δ PiwOut-of-limit, the power Δ P of unitiwAnd time-to-climb tiwCan recalculate as follows:
&Delta;P i w = &Delta;P i w max i f &Delta;P i w > &Delta;P i w max &Delta;P i w min i f &Delta;P i w < &Delta;P i w min
t i w = t i w max = &Delta;P i w max / P i w r a t e + i f &Delta;P i w > &Delta;P i w max &Delta;P i w min / P i w r a t e - i f &Delta;P i w < &Delta;P i w min
Wherein,Represent VGT respectivelyiThe minimum and maximum spare capacity of power of w unit.At this Unit in embodimentNumerical value be shown in Table 1.
If power of the assembling unit Δ PiwOut-of-limit, with connection weight all vanishing of unit w, it may be assumed that awj=0, j=1,2, L, mi, with Shi Jinhang corresponding row stochastic matrix element updates.
6. VGT is updatediPower offset value Δ Perror-i;In the present embodiment, VGT1 power offset value is carried out more by definition New:
&Delta;P e r r o r - i = &Delta;P i - &Sigma; w = 1 m i &Delta;P i w
7. precision is judged.If | Δ Perror-i| < εi, then power of the assembling unit Δ P is obtainediw, otherwise return step 3..In this reality Execute in example, take peak power deviation | Δ Perror-1| < 0.1MW is as the condition of convergence of VGT1.
Table 1 Guangdong Power Grid AGC unit parameter statistical table
Above-described embodiment is the present invention preferably embodiment, but embodiments of the present invention are not by above-described embodiment Limit, the change made under other any spirit without departing from the present invention and principle, modify, substitute, combine, simplify, All should be the substitute mode of equivalence, within being included in protection scope of the present invention.

Claims (8)

1. an AGC power dynamic allocation method based on virtual generating clan, it is characterised in that comprise the following steps:
(1) build Automation generation control power distribution frame based on virtual generating clan VGT, determine the leader of clan;
(2) state discrete collection S and action discrete set A is determinedQ
(3) real-time total generated output instruction Δ P of current control period regional power grid is gatheredFor current state, and according to Currently optimize task, utilize transfer learning to initialize value function matrix and the selection action of each virtual generating clan;
(4) according to leader issue policy selection action and calculate clan regulation power Δ Pi
(5) power Δ P is regulated by claniReward function value R immediatelyiBetween mapping calculation, it is thus achieved that each virtual Power Generation Section Reward function value R immediately falleni
(6) concordance weight and the self study weight of each virtual generating clan under current state are updated;
(7) according under the renewal current control period of award value immediately of each virtual generating clan of current control period each virtual The value function matrix Q of electricity clani
(8) the value function matrix Q of leader's+1 iteration of kth in current control period is judgedi k+1Value function with kth time iteration Matrix Qi kTwo norms of difference whether less than an infinitesimal positive number ε, i.e. | | Qi k+1‐Qi k||2<ε;If being unsatisfactory for, then return Step (5), if meeting, turns next step;
(9) power Δ P is regulated according to clani, by time-to-climb consistency algorithm try to achieve power of the assembling unit Δ Piw, and at the next one When the control cycle arrives, return step (3).
AGC power dynamic allocation method based on virtual generating clan the most according to claim 1, it is characterised in that step (1) described based on virtual generating clan VGT the Automation generation control power distribution frame in, is by traditional area electrical network Being divided into several manors electrical network VGT according to geographical distribution, manor electrical network VGT is joined by large power plant PLC, active in the electrical network of manor The generating set group composition that net AGC, microgrid AGC and load control system are constituted;Between the electrical network VGT of manor use leader with Communicating cooperation with person's pattern, leader is the control centre of regional power grid, is responsible for power disturbance balance;Follower is common The dispatching terminal of VGT, main being responsible for interacts collaborative with leader.
AGC power dynamic allocation method based on virtual generating clan the most according to claim 1, it is characterised in that described State discrete collection S in step (2) is to be instructed Δ P by each total generated outputConstituted as a state.
AGC power dynamic allocation method based on virtual generating clan the most according to claim 1, it is characterised in that described In step (3) according to currently optimizing task, utilize transfer learning initialize each virtual generating clan value function matrix and Before selection action, need to set up value function source matrix;Δ P is instructed with total generated outputThe degree of closeness of size is as difference The relativity evaluation index of optimization task;First, total generated output is instructed Δ PSize carries out interval range division, interval point It is not:
{ &lsqb; &Delta;P l e f t 1 , &Delta;P r i g h t 1 ) , ... , &lsqb; &Delta;P l e f t &eta; , &Delta;P r i g h t &eta; ) , ... , &lsqb; &Delta;P l e f t Z , &Delta;P r i g h t Z &rsqb; }
Wherein,Represent η total generated output instruction Δ P respectivelyInterval left and right end points;Z is total generating Power instruction Δ PInterval number;
Gather total generated output instruction Δ P of current control period, it is assumed that it falls in the η load disturbance interval, the most currently The dependency that optimization task optimizes task with left and right end points is respectively as follows:
r l e f t = &Delta;P r i g h t &eta; - &Delta;P &Sigma; &Delta;P r i g h t &eta; - &Delta;P l e f t &eta; r r i g h t = &Delta;P &Sigma; - &Delta;P l e f t &eta; &Delta;P r i g h t &eta; - &Delta;P l e f t &eta;
Wherein, rleft、rrightRepresent respectively and currently optimize the corresponding dependency optimizing task of task end points left and right with interval, 0≤ rleft≤ 1,0≤rright≤ 1, rleft+rright=1;
The initial value function matrix of the optimization task under current state is:
Qi=rleftQi,left+rrightQi,rightI=1,2 ..., n
In formula: Qi,left、Qi,rightRepresent that virtual generating clan i, can at interval left and right end points correspondence optimal value function matrix respectively By learning acquisition in advance.
AGC power dynamic allocation method based on virtual generating clan the most according to claim 1, it is characterised in that described Initial selected action in step (3) is the action selected after the most pre-study before CTQ algorithm carries out transfer learning.
AGC power dynamic allocation method based on virtual generating clan the most according to claim 1, it is characterised in that step (5) described reward function value R immediately ini(sk,ak,sk+1) it is designed as:
R i ( s k , a k , s k + 1 ) = - &mu; 1 &Sigma; w = 1 m i C i w &Delta;P i w - &mu; 2 max j = i &cup; j &Element; &Omega; i ( min max w = 1 , 2 , K , m j ( &Delta;P j w / &Delta;P j w r a t e ) )
Wherein, μ1、μ2Be respectively regulation the goal of cost and time-to-climb target weight coefficient, and μ1>=0, μ2>=0, regulate expense Target refer to minimize whole control area electrical network in the adjustment cost sum of all AGC units, time-to-climb target be minimum The time-to-climb of changing all units maximum;CiwFor the adjustment cost coefficient of w unit in clan i;ΔPiwRepresent in clan i The generated output instruction of w unit;Represent the regulations speed of w unit in clan i;N is the total number of clan;miFor The total number of unit of clan i;ΩiRepresent the set of the clan adjacent with clan i;Use method of least square to clan power Δ Pi And RiCarry out mapping relations matching, to accelerate the calculating of reward function value.
AGC power dynamic allocation method based on virtual generating clan the most according to claim 1, it is characterised in that described The concordance weight beta of each the virtual generating clan in step (6)k(sk,ak) and self study weight αk(sk,ak) be updated as Under:
&beta; k ( s k , a k ) = o 2 ( N O ( s k , a k ) + 1 ) &tau; 2
&alpha; k ( s k , a k ) = o 1 ( N O ( s k , a k ) + 1 ) &tau; 1
Wherein, NO(sk,ak) represent that state action is to (sk,ak) algorithm explore optimizing occurrence number;o1、o2、τ1、τ2Respectively For normal number, and τ to be met1∈(1/2,1)、0<τ21‐1/(2+ε1) constraint, wherein ε1It it is an infinitesimal positive number.
AGC power dynamic allocation method based on virtual generating clan the most according to claim 1, it is characterised in that described The value function matrix Q of each virtual generating clan under current control period in step (7)iUpdate as follows:
Q i k + 1 ( s k , a k ) = Q i k ( s k , a k ) - C Q ( s k , a k ) + I Q ( s k , a k )
Wherein, CQ(sk,ak) represent that virtual generating clan i is at skExecution action a under statekTime and adjacent virtual generating clan between Harmonious property update item;IQ(sk,ak) represent that virtual generating clan i is in state skLower execution action akTime self study update , as single virtual generating clan Q study update mechanism;CQ(sk,ak)、IQ(sk,ak) iteration update respectively as follows:
C Q ( s k , a k ) = &beta; k ( s k , a k ) &Sigma; j &Element; &Omega; i ( k ) ( Q i k ( s k , a k ) - Q j k ( s k , a k ) )
I Q ( s k , a k ) = &alpha; k ( s k , a k ) &lsqb; R i ( s k , a k , s k + 1 ) + &gamma; max a &prime; &Element; A Q Q i k ( s k + 1 , a &prime; ) - Q i k ( s k , a k ) &rsqb;
Wherein, sk、sk+1Represent the state of kth time and k+1 iteration respectively;It is that virtual generating clan i is at skState Lower execution action akQ-value;ΩiK () is the set when kth time iteration with the virtual i adjacent virtual generating clan of clan that generates electricity.
CN201610479264.9A 2016-06-24 2016-06-24 A kind of AGC power dynamic allocation methods based on virtual power generation clan Active CN106026084B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610479264.9A CN106026084B (en) 2016-06-24 2016-06-24 A kind of AGC power dynamic allocation methods based on virtual power generation clan

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610479264.9A CN106026084B (en) 2016-06-24 2016-06-24 A kind of AGC power dynamic allocation methods based on virtual power generation clan

Publications (2)

Publication Number Publication Date
CN106026084A true CN106026084A (en) 2016-10-12
CN106026084B CN106026084B (en) 2018-10-09

Family

ID=57083887

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610479264.9A Active CN106026084B (en) 2016-06-24 2016-06-24 A kind of AGC power dynamic allocation methods based on virtual power generation clan

Country Status (1)

Country Link
CN (1) CN106026084B (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106779248A (en) * 2017-01-11 2017-05-31 南方电网科学研究院有限责任公司 A kind of Economic Dispatch dispersion Q methods based on limit transfer learning
CN107045655A (en) * 2016-12-07 2017-08-15 三峡大学 Wolf pack clan strategy process based on the random consistent game of multiple agent and virtual generating clan
CN107069835A (en) * 2017-01-25 2017-08-18 国网冀北电力有限公司电力科学研究院 New energy power station distribution method active in real time and distributor
CN107094321A (en) * 2017-03-31 2017-08-25 南京邮电大学 A kind of vehicle-carrying communication MAC layer channel access method learnt based on multiple agent Q
CN107591847A (en) * 2017-08-04 2018-01-16 西安五常电力技术有限公司 A kind of method that mode using variable element adjusts Hydropower Unit AGC
CN108320080A (en) * 2018-01-05 2018-07-24 上海电力学院 The real-time dynamic power allocation method in energy internet based on two layers of consistency algorithm
CN114239439A (en) * 2021-12-24 2022-03-25 浙江金乙昌科技股份有限公司 Automatic filter design method based on tribal algorithm

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104037761A (en) * 2014-06-25 2014-09-10 南方电网科学研究院有限责任公司 AGC power multi-objective random optimization distribution method

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104037761A (en) * 2014-06-25 2014-09-10 南方电网科学研究院有限责任公司 AGC power multi-objective random optimization distribution method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
SOUMMYA KAR等: "QD-Learning:A Collaborative Distributed Strategy for Multi-agent Reinforcement Learning Through Consensus+Innovations", 《IEEE TRANSACTIONS ON SIGNAL PROCESSING》 *
张孝顺等: "互联电网AGC功率动态分配的虚拟发电部落协同一致性算法", 《中国电机工程学报》 *

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107045655A (en) * 2016-12-07 2017-08-15 三峡大学 Wolf pack clan strategy process based on the random consistent game of multiple agent and virtual generating clan
CN106779248A (en) * 2017-01-11 2017-05-31 南方电网科学研究院有限责任公司 A kind of Economic Dispatch dispersion Q methods based on limit transfer learning
CN107069835A (en) * 2017-01-25 2017-08-18 国网冀北电力有限公司电力科学研究院 New energy power station distribution method active in real time and distributor
CN107069835B (en) * 2017-01-25 2020-02-18 国网冀北电力有限公司电力科学研究院 Real-time active distribution method and device for new energy power station
CN107094321A (en) * 2017-03-31 2017-08-25 南京邮电大学 A kind of vehicle-carrying communication MAC layer channel access method learnt based on multiple agent Q
CN107094321B (en) * 2017-03-31 2020-04-28 南京邮电大学 Multi-agent Q learning-based vehicle-mounted communication MAC layer channel access method
CN107591847A (en) * 2017-08-04 2018-01-16 西安五常电力技术有限公司 A kind of method that mode using variable element adjusts Hydropower Unit AGC
CN107591847B (en) * 2017-08-04 2020-05-01 西安五常电力技术有限公司 Method for adjusting Automatic Gain Control (AGC) of hydroelectric generating set by using variable parameter mode
CN108320080A (en) * 2018-01-05 2018-07-24 上海电力学院 The real-time dynamic power allocation method in energy internet based on two layers of consistency algorithm
CN108320080B (en) * 2018-01-05 2021-09-07 上海电力学院 Energy internet real-time dynamic power distribution method based on two-layer consistency algorithm
CN114239439A (en) * 2021-12-24 2022-03-25 浙江金乙昌科技股份有限公司 Automatic filter design method based on tribal algorithm

Also Published As

Publication number Publication date
CN106026084B (en) 2018-10-09

Similar Documents

Publication Publication Date Title
CN112615379B (en) Power grid multi-section power control method based on distributed multi-agent reinforcement learning
CN106026084A (en) AGC power dynamic distribution method based on virtual generation tribe
CN111242443B (en) Deep reinforcement learning-based economic dispatching method for virtual power plant in energy internet
Xi et al. A wolf pack hunting strategy based virtual tribes control for automatic generation control of smart grid
CN103683337B (en) A kind of interconnected network CPS instruction dynamic assignment optimization method
CN104537428B (en) One kind meter and the probabilistic economical operation appraisal procedure of wind power integration
CN105023056B (en) The optimal carbon energy combined-flow acquisition methods of power grid based on colony intelligence intensified learning
CN106682808A (en) Online rolling optimization scheduling model
CN105391090B (en) A kind of intelligent grid multiple agent multiple target uniformity optimization method
CN105389633A (en) Optimization planning method of substation considering distributed power supplies
CN108494022B (en) Accurate scheduling method based on distributed power supply in micro-grid
CN104333047B (en) Real-time rolling planning method applied to wind power integration of power system
CN109149648A (en) A kind of adaptive width Dynamic Programming intelligent power generation control method
CN107092991A (en) A kind of adaptive economic load dispatching distribution method of intelligent grid
CN104536304A (en) Electric system load multi-agent control method based on Matlab and Netlogo
CN107612045A (en) A kind of Power Plant generated energy intelligent dispensing system and method
CN116207739B (en) Optimal scheduling method and device for power distribution network, computer equipment and storage medium
CN105912767B (en) Multistage power grid distributed collaborative combined calculation method based on B/S framework
CN105787650A (en) Simulation calculation method for Nash equilibrium point of electricity market including multiple load agents
CN110474353A (en) Layer-stepping energy-storage system and its power grid frequency modulation control method for coordinating of participation
Liu et al. Spatial assessment of China’s green governance efficiency in the period of high-quality development
CN107392350B (en) Comprehensive optimization method for power distribution network extension planning containing distributed energy and charging stations
CN111259315B (en) Decentralized scheduling method of multi-subject coordinated pricing mode
CN107357988A (en) Distributed photovoltaic cluster dynamic modelling method based on IEC61850
CN106651136A (en) Day-ahead power generation plan compilation method of bilateral transaction and apparatus thereof

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant