CN111275572B - Unit scheduling system and method based on particle swarm and deep reinforcement learning - Google Patents

Unit scheduling system and method based on particle swarm and deep reinforcement learning Download PDF

Info

Publication number
CN111275572B
CN111275572B CN202010043546.0A CN202010043546A CN111275572B CN 111275572 B CN111275572 B CN 111275572B CN 202010043546 A CN202010043546 A CN 202010043546A CN 111275572 B CN111275572 B CN 111275572B
Authority
CN
China
Prior art keywords
target
coal
input
particle swarm
unit
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010043546.0A
Other languages
Chinese (zh)
Other versions
CN111275572A (en
Inventor
于长军
林志赟
韩志敏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Dianzi University
Original Assignee
Hangzhou Dianzi University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Dianzi University filed Critical Hangzhou Dianzi University
Priority to CN202010043546.0A priority Critical patent/CN111275572B/en
Publication of CN111275572A publication Critical patent/CN111275572A/en
Application granted granted Critical
Publication of CN111275572B publication Critical patent/CN111275572B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
    • G06Q50/06Electricity, gas or water supply
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/004Artificial life, i.e. computing arrangements simulating life
    • G06N3/006Artificial life, i.e. computing arrangements simulating life based on simulated virtual individual or collective life forms, e.g. social simulations or particle swarm optimisation [PSO]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y04INFORMATION OR COMMUNICATION TECHNOLOGIES HAVING AN IMPACT ON OTHER TECHNOLOGY AREAS
    • Y04SSYSTEMS INTEGRATING TECHNOLOGIES RELATED TO POWER NETWORK OPERATION, COMMUNICATION OR INFORMATION TECHNOLOGIES FOR IMPROVING THE ELECTRICAL POWER GENERATION, TRANSMISSION, DISTRIBUTION, MANAGEMENT OR USAGE, i.e. SMART GRIDS
    • Y04S10/00Systems supporting electrical power generation, transmission or distribution
    • Y04S10/50Systems or methods supporting the power network operation or management, involving a certain degree of interaction with the load-side end user applications

Abstract

The invention discloses a unit scheduling system and a unit scheduling method based on particle swarm and deep reinforcement learning, wherein the system comprises a particle swarm module and a deep reinforcement learning model, wherein the deep reinforcement learning model comprises an evaluation network, an experience playback pool, a target network and a loss function, wherein the input of the particle swarm module is a load demand, the output of the particle swarm module is connected with the evaluation network, and the evaluation network outputs a Q estimated value and the experience playback pool; and the experience playback pool output is connected with the target network, the target network outputs a Q target value, both the Q target value and the Q estimated value are input into the loss function, and the output of the loss function is fed back to the evaluation network. The invention can optimize the dispatching of the unit, and can meet the load requirement, save at least 0.1g of coal per degree of electricity, and realize the control optimization of the dispatching control integration of the bottom equipment and the unit.

Description

Unit scheduling system and method based on particle swarm and deep reinforcement learning
Technical Field
The invention belongs to the field of information control, and relates to a unit scheduling system and method based on particle swarm and deep reinforcement learning.
Background
The economic dispatch of the electric power unit is an important link in the operation of an electric power system, and is always a research object of academic scholars due to the characteristics of multiple constraints, nonlinearity and high dimensionality. The significance of unit economic dispatch optimization is that the power system not only improves the working and running efficiency, greatly improves the comprehensive benefit of power enterprises, reduces the environmental impact, but also realizes the automation and the intellectualization of the system by the application of artificial intelligence.
The power economy schedule can be understood as: on the premise of ensuring that the power production is met, the power generation production of each unit is safely and fully scheduled, so that the power generation cost is minimum. There are many studies on economic optimization problems of the unit, such as genetic algorithm, ant colony algorithm, particle swarm algorithm, neural network, reinforcement learning and various algorithms generated by fusion. With the development of the power system, the economic dispatch complexity of the unit is increased, and different constraints such as the starting and stopping time cost of the unit, the climbing consumption cost of the unit and the like are added in the original economic optimization problem of the unit. However, in all researches, unit scheduling is only optimized in the original system, but control optimization of the parameter integration of the bottom equipment cannot be realized, so that the required coal burning amount cannot be further optimized.
Disclosure of Invention
In order to solve the problems, the invention aims to provide a set scheduling system based on particle swarm and deep reinforcement learning, which comprises a particle swarm module and a deep reinforcement learning model, wherein the deep reinforcement learning model comprises an evaluation network, an experience playback pool, a target network and a loss function,
the input of the particle swarm module is load demand, the output of the particle swarm module is connected with the evaluation network, and the evaluation network outputs a Q estimated value and the experience playback pool; and the experience playback pool output is connected with the target network, the target network outputs a Q target value, both the Q target value and the Q estimated value are input into the loss function, and the output of the loss function is fed back to the evaluation network.
Preferably, the particle swarm module outputs a target coal amount and bottom layer controllable equipment parameters, the target coal amount is used as an input state, and the bottom layer controllable equipment parameters are used as input actions.
Preferably, the evaluation network outputs to the experience playback pool a target coal amount, an underlying controllable device parameter, a predicted reward, and a next state target coal amount.
Preferably, the experience playback pool outputs a target coal amount for a next state to the target network.
Preferably, the particle number in the particle swarm module is 80, the inertia weight w=1, and the learning factor c 1 =c 2 =2.01, maximum velocity of particles is 1, number of iterations is 1500, fitness function is:
Figure GDA0004170829640000021
Figure GDA0004170829640000022
wherein a is i ,b i ,c i The energy consumption coefficient of each unit is calculated; the particle position and velocity update formula is:
Figure GDA0004170829640000023
where k represents the number of iteration steps, alpha is the contraction factor,
Figure GDA0004170829640000024
pbest is the optimal position in the particle history, gbest is the optimal position in all particles, rand () is the random function value range of [0,1 ]]。
Based on the above purpose, the invention also provides a method for the unit scheduling system based on particle swarm and deep reinforcement learning, which comprises the following steps:
s10, optimizing and obtaining the output of all units by utilizing a particle swarm module according to the load demand: with the demand meeting instruction as a target, under the constraint of the force of each unit, reasonably distributing all units by using the particle swarm module, wherein the obtained result is the production value of each unit;
s20, calculating the coal burning amount according to the unit output: calculating according to a conversion formula of the coal-fired quantity and the unit output to obtain an average coal-fired quantity, namely the coal-fired quantity of each degree of electricity, wherein the target coal-fired quantity is the average coal-fired quantity, and the average coal-fired quantity is reduced by at least 1g coal-fired quantity per degree of electricity;
s30, taking the target coal quantity as an input state, taking bottom adjustable equipment parameters such as wind, water, coal and the like as input actions, and inputting the input actions into a deep reinforcement learning model;
s40, taking the target coal quantity as a target, regulating and controlling bottom layer controllable equipment parameters, and obtaining optimal parameters of all bottom layer equipment controllers on the premise of the target coal quantity by the obtained result;
s50, obtaining new coal quantity and cost according to the target coal quantity and unit load demand; and (3) re-utilizing the particle swarm module to plan the output of the unit according to new cost and load requirements, repeating S10-S40, obtaining the optimal parameters of bottom equipment according to the target coal burning quantity, and finally re-planning the output of the unit according to a new cost function.
Preferably, in S40, the target coal amount is used as an input state S, the bottom controllable equipment parameter is used as an input action a, the input state S is input into an evaluation network of the deep reinforcement learning model, the evaluation network autonomously learns to obtain an estimated reward reaching the next state, the Q estimated value is input into a loss function, the input state S, the input action a, the estimated reward r and the next state S 'are combined and input into an experience playback pool, the next state S' is used as an input state to be input into the target network to obtain an actually available reward, namely, the Q target value, and the difference between the Q target value and the Q estimated value is used as feedback of the deep reinforcement learning model to be input into the evaluation network again, so that the learning performance is improved.
Compared with the prior art, the invention has at least the following beneficial effects: the method has the advantages that when the unit scheduling is optimized, from the perspective of saving the coal consumption, the load demand is met, at least 0.1 gram of coal consumption is saved on each degree of electricity, and meanwhile, the control optimization of the integration of the bottom equipment and the unit scheduling control is realized. The invention combines a particle swarm module and a deep reinforcement learning model. The particle swarm module has the characteristics of few parameters, easy realization, global optimum searching and the like, and has universal application on the problem of unit dispatching optimization. The deep reinforcement learning model is a combination of deep learning and reinforcement learning, and the deep learning has stronger perceptibility but lacks a certain decision capability; the reinforcement learning has decision making capability, combines the two to complement advantages, and provides a solution to the problem of complex system perception decision making.
Drawings
FIG. 1 is a block diagram of a group scheduling system based on particle swarm and deep reinforcement learning according to an embodiment of the present invention;
FIG. 2 is a flow chart of steps of a method for scheduling a unit based on particle swarm and deep reinforcement learning according to an embodiment of the present invention;
FIG. 3 is a graph of the variation of the unit output distribution before and after optimization based on the unit scheduling of particle swarm and deep reinforcement learning according to the embodiment of the invention;
FIG. 4 is a schematic diagram of a power plant cost change process in a unit output process of a unit dispatching optimization unit based on particle swarm and deep reinforcement learning according to an embodiment of the invention;
fig. 5 is a schematic diagram of a loss function change process based on a particle swarm and a deep reinforcement learning system according to an embodiment of the present invention.
Detailed Description
The present invention will be described in further detail with reference to the drawings and examples, in order to make the objects, technical solutions and advantages of the present invention more apparent. It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
On the contrary, the invention is intended to cover any alternatives, modifications, equivalents, and variations as may be included within the spirit and scope of the invention as defined by the appended claims. Further, in the following detailed description of the present invention, certain specific details are set forth in order to provide a better understanding of the present invention. The present invention will be fully understood by those skilled in the art without the details described herein.
System example 1
Referring to fig. 1, a particle swarm and deep reinforcement learning-based crew scheduling system according to an embodiment of the present invention is shown comprising a particle swarm module 10 and a deep reinforcement learning model 20, wherein the deep reinforcement learning model 20 comprises an evaluation network 21, an empirical playback pool 22, a target network 23, and a loss function 24, wherein,
the input of the particle swarm module 10 is the load demand, the output is connected with the evaluation network 21, and the evaluation network 21 outputs the Q estimated value and the experience playback pool 22; the output of the experience playback pool 22 is connected with a target network 23, the target network 23 outputs a Q target value, the Q target value and the Q estimated value are both input into a loss function 24, and the output of the loss function 24 is fed back to the evaluation network 21.
System example 2
The particle swarm module 10 outputs a target coal amount and bottom layer controllable equipment parameters, wherein the target coal amount is used as an input state, and the bottom layer controllable equipment parameters are used as input actions.
The evaluation network 21 outputs to the experience playback pool 22 a target coal amount, bottom layer controllable device parameters, estimated rewards, and next state target coal amount.
The experience playback pool 22 outputs a target coal amount for the next state to the target network 23.
The particle swarm module 10 has a particle number of 80, an inertia weight w=1, and a learning factor c 1 =c 2 =2.01, maximum velocity of particles is 1, number of iterations is 1500, fitness function is:
Figure GDA0004170829640000051
Figure GDA0004170829640000052
wherein a is i ,b i ,c i The energy consumption coefficient of each unit is calculated; the particle position and velocity update formula is:
Figure GDA0004170829640000053
where k represents the number of iteration steps, alpha is the contraction factor,
Figure GDA0004170829640000054
pbest is the optimal position in the particle history, gbest is the optimal position in all particles, rand () is the random function value range of [0,1 ]]。
Parameter settings in the deep reinforcement learning model 20: deep reinforcement learning is implemented using a fixed Q-network and an experience playback pool 22. Wherein the evaluation network 21 and the target network 23 in the fixed Q-network are 5 hidden layers, each layer is 20 neurons, each 5 steps replace the parameters of the target network, and the activation function is that
Figure GDA0004170829640000055
The learning rate is 0.01, epsilon-greedy is set to be 0.9, the reward attenuation value gamma is 0.9, the memory storage is 500, the reward rule is +1 if the reward rule is higher than the load requirement, is-1 if the reward rule is lower than the load requirement, is 0 if the reward rule meets the load requirement, the iteration number is 300, and the deep reinforcement learning reward function is designed based on a cost function. The input layer is an observed value (Observation), and the actions (Action) are the opening of a secondary air baffle, the speed b of a coal mill, the rotating speed c of a belt, the opening of a water supply valve d and the power e of a water supply pump. The observation table is shown in Table 1, wherein C isThe average coal burning amount, C-0.1, C-0.2, C-0.3, C-0.4 and C-0.5 are the target average coal burning amount, namely the minimum saving of 0.1g, 0.2g, 0.3g, 0.4g and 0.5g of coal burning amount per degree of electricity.
TABLE 1 watch list
Figure GDA0004170829640000056
Method embodiment
Referring to fig. 2, a method of the set scheduling system based on particle swarm and deep reinforcement learning includes the following steps:
s10, optimizing and obtaining the output of all units by utilizing a particle swarm module according to the load demand: with the demand meeting instruction as a target, under the constraint of the force of each unit, reasonably distributing all units by using the particle swarm module, wherein the obtained result is the production value of each unit;
s20, calculating the coal burning amount according to the unit output: calculating according to a conversion formula of the coal-fired quantity and the unit output to obtain an average coal-fired quantity, namely the coal-fired quantity of each degree of electricity, wherein the target coal-fired quantity is the average coal-fired quantity, and the average coal-fired quantity is reduced by at least 1g coal-fired quantity per degree of electricity;
s30, taking the target coal quantity as an input state, taking bottom adjustable equipment parameters such as wind, water, coal and the like as input actions, and inputting the input actions into a deep reinforcement learning model;
s40, taking the target coal quantity as a target, regulating and controlling bottom layer controllable equipment parameters, and obtaining optimal parameters of all bottom layer equipment controllers on the premise of the target coal quantity by the obtained result;
s50, obtaining new coal quantity and cost according to the target coal quantity and unit load demand; and (3) re-utilizing the particle swarm module to plan the output of the unit according to new cost and load requirements, repeating S10-S40, obtaining the optimal parameters of bottom equipment according to the target coal burning quantity, and finally re-planning the output of the unit according to a new cost function.
In the specific embodiment, in S40, the target coal amount is used as an input state S, the bottom controllable equipment parameter is used as an input action a, the input state S is input into an evaluation network of the deep reinforcement learning model, the evaluation network learns autonomously to obtain an estimated reward reaching the next state, the Q estimated value is input into a loss function, the input state S, the input action a, the estimated reward r and the next state S 'are combined and stored in an experience playback pool, the next state S' is used as the input state to be input into the target network to obtain an actually available reward, namely, the Q target value, and the difference between the Q target value and the Q estimated value is used as feedback of the deep reinforcement learning model to be input into the evaluation network again, so that the learning performance is improved.
In the specific embodiment
The output of each unit is optimized according to the load demand by utilizing the particle swarm module, and the obtained simulation effect is shown in fig. 3 and 4.
Fig. 3 is a graph showing changes before and after optimizing the distribution of the machine set output by using the shrinkage factor particle swarm module 10, and the abscissa is the number of the machine sets, and the total number of the machine sets is 40. The ordinate is the output condition of each unit. The black bar graph is the initial output of the unit before optimization, and the white bar graph is the output of the unit after optimization.
FIG. 4 is a graph of the change in plant cost during optimization of unit output. The abscissa is the iteration step number of the method, and the ordinate is the power plant cost. From the figure, it can be seen that as the method iterates, the curve always shows a decreasing trend, and the cost of the power plant is also continuously decreasing.
And calculating the coal burning amount by utilizing the optimized output of each unit according to the following formula:
Figure GDA0004170829640000071
Figure GDA0004170829640000072
wherein B is the combustion quantity (t/h) of the boiler, N is the unit output power (MW), Q net,ar The generated heat of the coal is lower than the generated heat of the standard coal (kJ/kg), 29271 is lower than the generated heat of the standard coal (kJ/kg), f b The standard coal consumption (g/kWh) is used for power generation.
Dividing the coal consumption by the load requirement to obtain the coal consumption C of each degree of electricity, dividing 5 target coal consumption states (states) which are C-0.1, C-0.2, C-0.3, C-0.4 and C-0.5 respectively, inputting observed values (observation) into a deep reinforcement learning model, wherein the obtained effect is shown in figure 5,
fig. 5 shows a change in the loss function of the deep reinforcement learning model 20, wherein the abscissa indicates the number of learning steps of the deep reinforcement learning, and the ordinate indicates the change in the prediction error in the deep reinforcement learning. Because the input is a constantly exploring process and the input data is obtained from learning, the curve is not a smooth curve.
And (3) the particle swarm module is reused to plan the unit output according to the new coal cost and the original load requirement, so that the optimization of the whole unit scheduling is completed.
The foregoing description of the preferred embodiments of the invention is not intended to be limiting, but rather is intended to cover all modifications, equivalents, and alternatives falling within the spirit and principles of the invention.

Claims (2)

1. A unit scheduling method based on particle swarm and deep reinforcement learning is characterized in that the unit scheduling system based on the particle swarm and the deep reinforcement learning comprises a particle swarm module and a deep reinforcement learning model, wherein the deep reinforcement learning model comprises an evaluation network, an experience playback pool, a target network and a loss function,
the input of the particle swarm module is load demand, the output of the particle swarm module is connected with the evaluation network, and the evaluation network outputs a Q estimated value and the experience playback pool; the experience playback pool output is connected with the target network, the target network outputs a Q target value, the Q target value and the Q estimated value are both input into the loss function, and the output of the loss function is fed back to the evaluation network;
the particle swarm module outputs target coal quantity and bottom layer controllable equipment parameters, wherein the target coal quantity is used as an input state, and the bottom layer controllable equipment parameters are used as input actions;
the evaluation network outputs target coal burning quantity, bottom controllable equipment parameters, estimated rewards and target coal burning quantity in the next state to the experience playback pool;
the experience playback pool outputs a target coal amount in the next state to the target network;
the particle number in the particle swarm module is 80, the inertia weight w=1, and the learning factor c 1 =c 2 =2.01, maximum velocity of particles is 1, number of iterations is 1500, fitness function is:
Figure FDA0004170829630000011
Figure FDA0004170829630000012
wherein a is i ,b i ,c i The energy consumption coefficient of each unit is calculated; the particle position and velocity update formula is:
Figure FDA0004170829630000013
where k represents the number of iteration steps, alpha is the contraction factor,
Figure FDA0004170829630000014
pbest is the optimal position in the particle history, gbest is the optimal position in all particles, rand () is the random function value range of [0,1 ]];
The method comprises the following steps:
s10, optimizing and obtaining the output of all units by utilizing a particle swarm module according to the load demand: with the demand meeting instruction as a target, under the constraint of the force of each unit, reasonably distributing all units by using the particle swarm module, wherein the obtained result is the production value of each unit;
s20, calculating the coal burning amount according to the unit output: calculating according to a conversion formula of the coal-fired quantity and the unit output to obtain an average coal-fired quantity, namely the coal-fired quantity of each degree of electricity, wherein the target coal-fired quantity is the average coal-fired quantity, and the average coal-fired quantity is reduced by at least 1g coal-fired quantity per degree of electricity;
s30, taking the target coal quantity as an input state, taking bottom adjustable equipment parameters such as wind, water, coal and the like as input actions, and inputting the input actions into a deep reinforcement learning model;
s40, taking the target coal quantity as a target, regulating and controlling bottom layer controllable equipment parameters, and obtaining optimal parameters of all bottom layer equipment controllers on the premise of the target coal quantity by the obtained result;
s50, obtaining new coal quantity and cost according to the target coal quantity and unit load demand; and (3) re-utilizing the particle swarm module to plan the output of the unit according to new cost and load requirements, repeating S10-S40, obtaining the optimal parameters of bottom equipment according to the target coal burning quantity, and finally re-planning the output of the unit according to a new cost function.
2. The method according to claim 1, wherein in S40, the target coal amount is used as an input state S, the bottom controllable device parameter is used as an input action a, the input state S is input into an evaluation network of the deep reinforcement learning model, the evaluation network autonomously learns to obtain an estimated reward reaching the next state, the Q estimated value is input into a loss function, the input state S, the input action a, the estimated reward r and the next state S 'are combined and stored in an experience playback pool, the next state S' is used as the input state and is input into the target network to obtain an actually available reward, namely, the Q target value, and the difference between the Q target value and the Q estimated value is used as feedback of the deep reinforcement learning model to be input into the evaluation network again, so as to improve learning performance.
CN202010043546.0A 2020-01-15 2020-01-15 Unit scheduling system and method based on particle swarm and deep reinforcement learning Active CN111275572B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010043546.0A CN111275572B (en) 2020-01-15 2020-01-15 Unit scheduling system and method based on particle swarm and deep reinforcement learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010043546.0A CN111275572B (en) 2020-01-15 2020-01-15 Unit scheduling system and method based on particle swarm and deep reinforcement learning

Publications (2)

Publication Number Publication Date
CN111275572A CN111275572A (en) 2020-06-12
CN111275572B true CN111275572B (en) 2023-07-11

Family

ID=71001640

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010043546.0A Active CN111275572B (en) 2020-01-15 2020-01-15 Unit scheduling system and method based on particle swarm and deep reinforcement learning

Country Status (1)

Country Link
CN (1) CN111275572B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112016811A (en) * 2020-08-04 2020-12-01 四叶草(苏州)智能科技有限公司 AGV intelligent scheduling system and method based on reinforcement learning

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2012143424A1 (en) * 2011-04-19 2012-10-26 Ge Energy Products France Snc System and method for controlling an electrical energy production installation
CN108108532A (en) * 2017-12-06 2018-06-01 华南理工大学 With the method for particle cluster algorithm optimization power electronic circuit
CN108539784A (en) * 2018-04-13 2018-09-14 华南理工大学 The optimization method of the optimal unit of micro-capacitance sensor and tou power price based on Demand Side Response

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9373960B2 (en) * 2013-03-13 2016-06-21 Oracle International Corporation Computerized system and method for distributed energy resource scheduling
CN104682405B (en) * 2015-03-31 2018-01-12 福州大学 A kind of var Optimization Method in Network Distribution based on taboo particle cluster algorithm
CN108390369A (en) * 2017-12-31 2018-08-10 天津求实智源科技有限公司 Electric load general power real-time decomposition method and system
US11416739B2 (en) * 2018-01-29 2022-08-16 Lawrence Livermore National Security, Llc Optimization control technology for building energy conservation
CN109347149B (en) * 2018-09-20 2022-04-22 国网河南省电力公司电力科学研究院 Micro-grid energy storage scheduling method and device based on deep Q-value network reinforcement learning
CN110414725B (en) * 2019-07-11 2021-02-19 山东大学 Wind power plant energy storage system scheduling method and device integrating prediction and decision
CN110518580B (en) * 2019-08-15 2023-04-28 上海电力大学 Active power distribution network operation optimization method considering micro-grid active optimization
CN110535146B (en) * 2019-08-27 2022-09-23 哈尔滨工业大学 Electric power system reactive power optimization method based on depth determination strategy gradient reinforcement learning
CN110515303B (en) * 2019-09-17 2022-09-09 余姚市浙江大学机器人研究中心 DDQN-based self-adaptive dynamic path planning method

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2012143424A1 (en) * 2011-04-19 2012-10-26 Ge Energy Products France Snc System and method for controlling an electrical energy production installation
CN108108532A (en) * 2017-12-06 2018-06-01 华南理工大学 With the method for particle cluster algorithm optimization power electronic circuit
CN108539784A (en) * 2018-04-13 2018-09-14 华南理工大学 The optimization method of the optimal unit of micro-capacitance sensor and tou power price based on Demand Side Response

Also Published As

Publication number Publication date
CN111275572A (en) 2020-06-12

Similar Documents

Publication Publication Date Title
Zhang et al. Lifelong learning for complementary generation control of interconnected power grids with high-penetration renewables and EVs
Li et al. Coordinated load frequency control of multi-area integrated energy system using multi-agent deep reinforcement learning
CN105870976B (en) A kind of low-carbon dispatching method and device based on energy environment efficiency
CN111340274A (en) Virtual power plant participation-based comprehensive energy system optimization method and system
CN111242443B (en) Deep reinforcement learning-based economic dispatching method for virtual power plant in energy internet
Xi et al. Automatic generation control based on multiple-step greedy attribute and multiple-level allocation strategy
CN108736509A (en) A kind of active distribution network multi-source coordinating and optimizing control method and system
CN104166877A (en) Microgrid optimization operation method based on improved binary system particle swarm optimization algorithm
Zhu et al. A parallel meta-heuristic method for solving large scale unit commitment considering the integration of new energy sectors
CN106712075A (en) Peaking strategy optimization method considering safety constraints of wind power integration system
CN111934360A (en) Virtual power plant-energy storage system energy collaborative optimization regulation and control method based on model predictive control
Yin et al. Fuzzy vector reinforcement learning algorithm for generation control of power systems considering flywheel energy storage
CN109901389A (en) A kind of new energy consumption method based on deep learning
CN114330869A (en) Multi-energy day-ahead scheduling optimization method for iron and steel enterprise considering multi-level indexes
CN115544899A (en) Water plant water intake pump station energy-saving scheduling method based on multi-agent deep reinforcement learning
CN111275572B (en) Unit scheduling system and method based on particle swarm and deep reinforcement learning
CN115147245A (en) Virtual power plant optimal scheduling method with industrial load participating in peak shaving auxiliary service
Hu et al. Optimal dispatch of combined heat and power units based on particle swarm optimization with genetic algorithm
CN116451947A (en) Optimal scheduling method for electric heating gas comprehensive energy system, terminal equipment and storage medium
CN109617052B (en) Intelligent layered control method for large-scale electric heat storage units
CN112132379B (en) Economic-considered new energy cross-region consumption evaluation method and storage medium
Luan et al. Load-following control of nuclear reactors based on Takagi-Sugeno fuzzy model
CN109066769B (en) Virtual power plant internal resource scheduling control method under wind power complete consumption
CN111541279A (en) Wind power plant power automatic control system and method considering unit output state
CN114285093B (en) Source network charge storage interactive scheduling method and system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant