CN114444802A - Electric vehicle charging guide optimization method based on graph neural network reinforcement learning - Google Patents
Electric vehicle charging guide optimization method based on graph neural network reinforcement learning Download PDFInfo
- Publication number
- CN114444802A CN114444802A CN202210109887.2A CN202210109887A CN114444802A CN 114444802 A CN114444802 A CN 114444802A CN 202210109887 A CN202210109887 A CN 202210109887A CN 114444802 A CN114444802 A CN 114444802A
- Authority
- CN
- China
- Prior art keywords
- node
- charging
- electric vehicle
- neural network
- reinforcement learning
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 93
- 230000002787 reinforcement Effects 0.000 title claims abstract description 75
- 238000005457 optimization Methods 0.000 title claims abstract description 53
- 238000000034 method Methods 0.000 title claims abstract description 30
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 52
- 230000006870 function Effects 0.000 claims abstract description 30
- 230000006399 behavior Effects 0.000 claims abstract description 26
- 230000004927 fusion Effects 0.000 claims abstract description 24
- 230000005611 electricity Effects 0.000 claims description 13
- 230000009471 action Effects 0.000 claims description 12
- 230000009977 dual effect Effects 0.000 claims description 11
- 238000004364 calculation method Methods 0.000 claims description 10
- 238000011160 research Methods 0.000 claims description 10
- 238000011478 gradient descent method Methods 0.000 claims description 9
- 239000003795 chemical substances by application Substances 0.000 claims description 8
- 239000011159 matrix material Substances 0.000 claims description 7
- 239000003245 coal Substances 0.000 claims description 6
- 108010074864 Factor XI Proteins 0.000 claims description 3
- OAICVXFJPJFONN-UHFFFAOYSA-N Phosphorus Chemical compound [P] OAICVXFJPJFONN-UHFFFAOYSA-N 0.000 claims description 3
- 238000006243 chemical reaction Methods 0.000 claims description 3
- 238000009795 derivation Methods 0.000 claims description 3
- 238000000605 extraction Methods 0.000 claims description 3
- 238000010586 diagram Methods 0.000 claims 1
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 230000003993 interaction Effects 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 3
- 230000008901 benefit Effects 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 230000001788 irregular Effects 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- YCKRFDGAMUMZLT-UHFFFAOYSA-N Fluorine atom Chemical compound [F] YCKRFDGAMUMZLT-UHFFFAOYSA-N 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 150000001875 compounds Chemical class 0.000 description 1
- 229910052731 fluorine Inorganic materials 0.000 description 1
- 239000011737 fluorine Substances 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
- G06Q10/04—Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
- G06Q30/0283—Price estimation or determination
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q50/00—Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
- G06Q50/06—Energy or water supply
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Business, Economics & Management (AREA)
- Engineering & Computer Science (AREA)
- Economics (AREA)
- Theoretical Computer Science (AREA)
- Strategic Management (AREA)
- Physics & Mathematics (AREA)
- Development Economics (AREA)
- General Physics & Mathematics (AREA)
- Marketing (AREA)
- Health & Medical Sciences (AREA)
- General Business, Economics & Management (AREA)
- Human Resources & Organizations (AREA)
- General Health & Medical Sciences (AREA)
- Tourism & Hospitality (AREA)
- Entrepreneurship & Innovation (AREA)
- Game Theory and Decision Science (AREA)
- Accounting & Taxation (AREA)
- Finance (AREA)
- Operations Research (AREA)
- Computational Linguistics (AREA)
- Water Supply & Treatment (AREA)
- Quality & Reliability (AREA)
- Primary Health Care (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Public Health (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Charge And Discharge Circuits For Batteries Or The Like (AREA)
- Electric Propulsion And Braking For Vehicles (AREA)
Abstract
The invention provides an electric vehicle charging guidance optimization method based on graph neural network reinforcement learning, which comprises the following steps: step S1: initializing a power-traffic fusion network collaborative optimization model; step S2: updating the charging load of the electric automobile; step S3: generating according to epsilon-Greedy algorithm and graph neural network reinforcement learning algorithma i,t (ii) a Step S4: enforcing charging boot behavior policiesa i,t (ii) a Step S5: calculating a reward function of the graph neural network reinforcement learning algorithm; step S6: partially observing the state of the Markov decision processx i,t Updating; step S7: information of the current step: (x i,t , a i,t ,r i,t ,x i,t ) Is stored in a memory cellDPerforming the following steps; step S8: judging whether the preset time is reachedT end(ii) a If not, executing (2) to (7); if yes, outputting the parameters of the graph neural network reinforcement learning algorithm and corresponding output results. By applying the technical scheme, the total charging cost of the electric automobile can be effectively reduced, and the orderly charging of the electric automobile and the collaborative optimization scheduling of the power system are realized.
Description
Technical Field
The invention relates to the technical field of collaborative optimization of an electric power-traffic fusion network, in particular to an electric vehicle charging guidance optimization method based on graph neural network reinforcement learning.
Background
With the large-scale operation of electric automobiles, a power system and a traffic system have a lot of interactive fusion to form a power-traffic fusion network. The fusion network relates to a plurality of main bodies such as electric automobiles, electric power systems, traffic systems and the like, and comprises a plurality of random uncertain factors. The interaction of a plurality of subjects, the influence of a plurality of random factors and the coupling relationship of the plurality of random factors make the understanding of the interaction influence mechanism of the electric power and traffic system and the solving of the cooperative optimization of the electric power-traffic fusion network more difficult. For example, traveling, psychological behaviors and driving behaviors of electric vehicle users have certain randomness, which affects flow distribution of a traffic system, so that the traffic flow also has certain uncertainty, further affects the time of the electric vehicle reaching a charging station, and causes charging time, queuing time and charging duration of the electric vehicle to have strong uncertainty. Unlike the conventional electric load, the electric vehicle is a movable load, and its randomness is stronger and more difficult to predict than the conventional electric load.
The current research on the electric power-traffic fusion network can be divided into three research directions: 1) from the perspective of an electric power system, the electric vehicle is guided to be charged at the lowest cost by calculating the marginal cost electricity price of the node or optimizing the service pricing of the charging station; 2) the charging cost is minimized by considering the charging path optimization from the perspective of a traffic system; 3) the benefits of the electric automobile, the electric power and the traffic system are comprehensively considered, and the comprehensive benefit maximization is realized by optimizing the charging strategy of the electric automobile and the scheduling decision of the electric power system. However, most of the existing research belongs to a static optimization problem, and the coupling relation of main bodies such as an electric vehicle, a charging station and an electric power system on a continuous time scale is not considered; meanwhile, most of the existing researches do not consider the influence of various uncertain factors and relevant coupling on the collaborative optimization of the power-traffic fusion network. More importantly, the influence of interaction between electric vehicles on the collaborative optimization of the electric power-traffic fusion network is not considered in the existing research.
Disclosure of Invention
In view of this, the present invention provides an electric vehicle charging guidance optimization method based on graph neural network reinforcement learning, which can effectively reduce the total charging cost of an electric vehicle and realize ordered charging of the electric vehicle and collaborative optimization scheduling of an electric power system under the condition of considering multiple uncertainty factors of an electric power-traffic fusion network.
In order to achieve the purpose, the invention adopts the following technical scheme: the electric vehicle charging guidance optimization method based on graph neural network reinforcement learning comprises the following steps:
step S1: initializing a power-traffic fusion network collaborative optimization model;
step S2: updating the charging load of the electric automobile, and carrying out optimization calculation on the marginal cost electricity price of the node where the electric automobile charging station is located based on second-order cone relaxation optimization and dual theory;
step S3: generating an electric vehicle charging guide behavior strategy a according to epsilon-Greedy algorithm and graph neural network reinforcement learning algorithmi,t;
Step S4: enforcing a charging bootstrapping behavior policy ai,tJudging and updating the state of the electric automobile;
step S5: calculating a reward function of a neural network reinforcement learning algorithm according to the electric power-traffic fusion environment;
step S6: partially observing state x of Markov decision processi,tUpdating;
step S7: information (x) of the current stepi,t,ai,t,ri,t,xi,t') storing in a memory unit D, and updating the graph neural network reinforcement learning algorithm weight based on a random gradient descent method; wherein x isi,tShowing the current state of the graph neural network reinforcement learning; a isi,tRepresenting an electric vehicle behavior strategy;ri,ta reward function value representing graph neural network reinforcement learning; x is a radical of a fluorine atomi,t' represents the state of the next step of the graph neural network reinforcement learning;
step S8: judging whether the preset time T is reachedend(ii) a If not, executing (2) - (7); if yes, outputting parameters of the graph neural network reinforcement learning algorithm and corresponding output results.
In a preferred embodiment, initializing the collaborative optimization model of the power-traffic fusion network comprises the following steps:
step 21: determining topological structures and parameters of a power network and a traffic network, wherein the topological structures and the parameters comprise power system nodes, lines, initial voltage and optimized upper and lower limit values, and the traffic network comprises traffic nodes, road parameters, capacity and maximum running speed;
step 22: neural network parameter initialization, including neural network weight initialization and hyper-parameter setting, such as learning rate alpha, discount factor gamma, batch size B and memory unit D capacity;
step 23: regarding each electric automobile in the research area as an agent, regarding the agent as a node N ∈ N, regarding the connection between the electric automobiles as an edge E ∈ E, and forming a graph network structure G ═ N, E, and regarding each electric automobile i in the current state x ∈ N, Ei,tAnd the adjacency matrix a.
In a preferred embodiment, the updating of the charging load of the electric vehicle and the optimization calculation of the marginal cost price of the node where the electric vehicle charging station is located based on the second order cone relaxation optimization and the dual theory comprise:
step 31: updating the charging load of the electric automobile: calculating the charging load of each charging station according to the number of electric vehicles and the charging power in the charging stations, and adding the basic load of the node to the charging load of each station to obtain the final electric load of the node;
step 32: establishing a power distribution network optimal power flow model based on a branch power flow model:
min f(p,q,P,Q,V,I) (1)
in the formula, ENAnd ELRespectively representing a power distribution network node and a line set; pijAnd QijRepresenting the active power and the reactive power of a branch flowing from the node i to the node j; pjkRepresenting the branch active power flowing from the node j to the node k;andrepresenting active and reactive power of the generator, i.e. active injected into node jPower and reactive power;andrepresenting the active power and the reactive power injected into the node j by the fan; qjsRepresenting the branch reactive power flowing from node j to node s; r isijAnd xijRepresents the branch resistance and reactance from node i to node j; i isijRepresents the branch current from node i to node j; π (j) represents the set of branches connected to node j;andrepresenting the active and reactive loads connected at node j; viRepresents the voltage amplitude of the node i; vjRepresents the voltage magnitude of node j; z is a radical ofijRepresents the impedance of the branch connecting node i and node j, and satisfies zij=rij+jxij;Representing the maximum value of the branch current connecting the node i and the node j;V jandrepresents the minimum and maximum voltages of node j;representing the maximum active output of the fan connected to node j;represents a power factor of the fan connected to node j;
load of distribution network node jIncluding base loadAnd charging load of electric vehicleNamely, it is
According to the actual demand of the distribution network, the objective function min f (P, Q, V, I) can be finally defined as:
in the formula (I), the compound is shown in the specification,representing the active output of the generator injected into node i; a isiAnd biRespectively representing the secondary coal consumption and the primary coal consumption coefficient of the generator;andrespectively purchasing the electricity price and the active power of the electric quantity from the main network;
step 33: converting the optimal power flow model of the nonlinear power distribution network into a second-order cone relaxation planning model:
the BFM-OPF is a nonlinear programming model, so that the branch current amplitude valueAnd branch voltage amplitudeAnd performing a second order taper relaxation (SOCR) conversion on the equationAlternatively, the following model can be obtained:
in the formula | · | non-conducting phosphor2Representing a second order cone operation; the formula constitutes the basic form of the optimal power flow of the power distribution network after relaxation;
step 34: solving the original problem and the dual variable of the model by adopting a Gurobi solver to obtain the marginal cost electrovalence lambda of the node where the charging station is locatedk。
In a preferred embodiment, the epsilon-Greedy algorithm comprises the steps of:
step 41: generating a random number u, and judging the size of the random number u and a decay factor xi of the epsilon-Greedy algorithm;
step 42: if u<Xi, then adopting random mode to treat each vehicle in current stateThe electric automobile generates a behavior ai,tThis behavior is expressed in the patent as an electric vehicle charging path strategy;
ai,t=randint(Naction) (19)
in the formula, NactionRepresenting the number of electric vehicle behavior decisions;
step 43: if u is larger than or equal to xi, each electric automobile i is in the current state x according to experience of the graph neural network reinforcement learning algorithmi,tAnd generating a behavior a under the adjacency matrix Ai,tI.e. by
In the formula, thetatParameters representing a graph neural network reinforcement learning algorithm; argmax () represents a parameter operation corresponding to the maximum value; x is the number ofi,tThe state of the ith electric vehicle at the time t is shown, which is mainly composed of the state x of the ith electric vehicle at the time ti,tFrom the electric vehicle state EVi,tAnd adjacent traffic road information Roi,tState Ne of electric vehicle in close proximityi,tAnd each charging station information CStIs composed of, i.e.
xi,t=[EVi,t,Roi,t,Nei,t,CSt] (21)
In the formula, i-th electric vehicle state EVi,tIncluding the next node when the electric vehicle goes to the charging stationRoad numberingElectric automobile driving speed vi,tAnd remaining capacity SOCi,t(ii) a Neighbor traffic road information state Roi,tComprises the next node of the electric automobile iStarting node of the next road connectedEnd nodeRoad lengthAnd number of electric vehicles on roadNearest neighbor electric vehicle state Nei,tIncluding the status of each neighboring electric vehicle k, e.g. the next node of the kth electric vehicle adjacent to the ith electric vehicleThe road number of the roadElectric automobile driving speed vi,k,tAnd remaining capacity SOCi,k,t(ii) a Charging station information CStCharging tariff p comprising charging stationsc,tAnd number of electric vehicles
The graph neural network reinforcement learning algorithm comprises a neural network structure comprising an input layer, a full connection layer and an input state xi,tPerforming feature extraction xi,t', then the characteristic x to be proposedi,tInputting the two layers of the adjacent matrixes A into a two-layer graph neural network together, extracting features, and finally connecting a full-connection layer to carry out a strategy a of charging paths of the electric automobilei,tOutputting; wherein, the graph neural network adopts a graph attention network.
In a preferred embodiment, the reward function r of the graph neural network reinforcement learning algorithmi,tAs shown in the formula:
in the formula, a nodecurAnd a nodetarIndicating the current node of the electric automobile and any charging station node to which the electric automobile is going, and step indicating the number of steps the electric automobile has run; penalty represents a large penalty factor; w is aiIs shown asiThe cost per unit time of the electric vehicle;andrespectively representing the driving time, the charging waiting time and the charging required time when the ith electric vehicle goes to the kth charging station at the time t; lambda [ alpha ]k,tRepresenting the marginal cost price of the node where the charging station k is located at the time t; SOCi,k,tRepresents the remaining capacity SOC when the ith electric vehicle reaches the charging station k at the time ti,k,t;The battery rated capacity of the ith electric automobile is represented;
the reward function can be seen from the formulari,tIs a piecewise function; if the ith electric vehicle does not reach the charging station nodecur≠nodetarAnd the step number of the current electric vehicle going to the charging station is step < N within the given maximum charging step numberstepAt this time, its reward function ri,t0; if the number of steps of the ith electric vehicle going to the charging station is more than or equal to the given maximum charging step number step which is more than or equal to NstepWhen the charging behavior is failed to be explored, a larger negative reward r is given to the charging behaviori,t-penalty; if the ith electric vehicle arrives at the charging station nodecur=nodetarAnd the step number of the current electric vehicle going to the charging station is step < N within the given maximum charging step numberstepAt the moment, the reward function is based on the running time of the electric automobileAnd charging timeAnd calculating the electricity charge during charging;
the passing time t of the ith electric automobile on the road section aa,tCalculated according to the U.S. Federal road administration function (BPR), i.e. the buseau of public roads
In the formula, na,tRepresenting the number of electric vehicles on the road section a at the moment t; c. CaAndrespectively representing the upper limit of the capacity of the road section a and the free passing time of the electric automobile at the time t; therefore, the time required for the ith electric vehicle to go to the charging station k can be obtainedNamely that
in the formula, SOCtRepresenting the residual electric quantity of the electric automobile;the rated capacity of the battery of the electric automobile is represented; eta represents a charging power factor, PchargingIndicating the rated power for charging the electric vehicle.
In a preferred embodiment, the updating weights of the graph neural network reinforcement learning algorithm based on the stochastic gradient descent method comprises:
step 61: randomly extracting a certain number of samples from the memory unit D;
step 62: constructing a loss function as shown in the formula, and updating the weight of the graph neural network reinforcement learning algorithm according to a random gradient descent method under the extracted Sample as shown in the formula;
wherein x, a, x 'and a' are the current state, action and the state and action at the next moment, respectively; r represents the immediate reward of graph neural network reinforcement learning; thetatA graph neural network reinforcement learning algorithm parameter representing the current time t; gamma is greater than or equal to 0 and less than or equal to 1, representing a discount factor that reflects the impact of future Q values on the current action;representing the neural network reinforcement learning calculation in the target graphMethod parameter theta'tState-action value of;
in the formula, thetatA graph neural network reinforcement learning algorithm parameter representing the current time t;is expressed in the pair thetatCarrying out derivation operation; α represents a learning rate;
and step 63: the neural network reinforcement learning parameter theta is obtained according to the current graph after a certain number of stepstReinforcing learning parameter theta 'for neural network of target map'tAnd (6) updating.
Compared with the prior art, the invention has the following beneficial effects:
the invention provides an electric vehicle charging guidance optimization method based on graph neural network reinforcement learning, which converts the mutual influence relationship among electric vehicles into a dynamic network graph structure based on a graph theory, and provides an attention mechanism-based graph neural network reinforcement learning method for processing irregular non-European structural data so as to research the communication and cooperation among multiple intelligent agents and discuss the mutual influence among the electric vehicles. On the basis of considering the output of renewable energy sources of an active power distribution network, the optimal power flow of the power distribution network is solved through a second-order cone optimization and dual optimization theory, and the marginal cost price of the power distribution network node is obtained, so that the power-traffic fusion network collaborative optimization is researched. The electric vehicle charging guiding optimization method based on graph neural network reinforcement learning can effectively reduce the total charging cost of the electric vehicle and realize ordered charging of the electric vehicle and cooperative optimization scheduling of a power system under the condition that multiple uncertainty factors of a power-traffic fusion network are considered.
Drawings
Fig. 1 is a flowchart of an electric vehicle charging guidance optimization method based on graph neural network reinforcement learning according to a preferred embodiment of the present invention.
Detailed Description
The invention is further explained below with reference to the drawings and the embodiments.
It should be noted that the following detailed description is exemplary and is intended to provide further explanation of the disclosure. Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs.
It is noted that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of example embodiments according to the present application; as used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, and it should be understood that when the terms "comprises" and/or "comprising" are used in this specification, they specify the presence of stated features, steps, operations, devices, components, and/or combinations thereof, unless the context clearly indicates otherwise.
As shown in fig. 1, the invention relates to an electric vehicle charging guidance optimization method based on graph neural network reinforcement learning, which includes the following steps:
s11: initializing a power-traffic fusion network collaborative optimization model;
s12: updating the charging load of the electric automobile, and carrying out optimization calculation on the marginal cost electricity price of the node where the electric automobile charging station is located based on second-order cone relaxation optimization and dual theory;
s13: generating an electric vehicle charging guide behavior strategy a according to epsilon-Greedy algorithm and graph neural network reinforcement learning algorithmi,t;
S14: enforcing a charging bootstrapping behavior policy ai,tJudging and updating the state of the electric automobile;
s15: calculating a reward function of a neural network reinforcement learning algorithm according to the electric power-traffic fusion environment;
s16: partially observing state x of Markov decision processi,tUpdating;
s17: information (x) of the current stepi,t,ai,t,ri,t,xi,t') stored in memory cells D and based on randomUpdating the weight of the graph neural network reinforcement learning algorithm by a gradient descent method;
s18: judging whether the preset time T is reachedend. If not, executing (2) - (7); if yes, outputting the parameters of the graph neural network reinforcement learning algorithm and corresponding output results.
Specifically, the method comprises the following steps:
firstly, initializing a power-traffic fusion network collaborative optimization model. The method mainly comprises the steps of determining topological structures and parameters of the power network and the traffic network, wherein the topological structures and the parameters comprise power system nodes, lines, initial voltage and optimized upper and lower limit values, and the traffic network comprises traffic nodes, road parameters, capacity, maximum driving speed and the like.
Neural network parameter initialization, including neural network weight initialization and hyper-parameter setting, such as learning rate alpha, discount factor gamma, batch size B and memory unit capacity D;
regarding each electric automobile in the research area as an agent, regarding the agent as a node N ∈ N, regarding the connection between the electric automobiles as an edge E ∈ E, and forming a graph network structure G ═ N, E, and regarding each electric automobile i in the current state x ∈ N, Ei,tAnd the adjacency matrix a.
And secondly, updating the charging load of the electric automobile, and performing optimization calculation on the marginal cost electricity price of the node where the electric automobile charging station is located based on second-order cone relaxation optimization and dual theory. The method mainly comprises the following steps:
step 21: updating the charging load of the electric automobile: calculating the charging load of each charging station according to the number of electric vehicles and the charging power in the charging stations, and adding the basic load of the node to the charging load of each station to obtain the final electric load of the node;
step 22: establishing a power distribution network optimal power flow model based on a branch power flow model:
min f(p,q,P,Q,V,I) (1)
in the formula, ENAnd ELRespectively representing a power distribution network node and a line set; pijAnd QijRepresenting the active power and the reactive power of a branch flowing from the node i to the node j;andrepresenting the active and reactive outputs of the generator, i.e. the active and reactive power injected into node j;andrepresenting active power and reactive power injected into the node j by the fan; r isijAnd xijRepresents the branch resistance and reactance from node i to node j; I.C. AijRepresents the branch current from node i to node j; π (j) represents the set of branches connected to node j;andrepresenting the active and reactive loads connected at node j; viRepresents the voltage amplitude of the node i; z is a radical ofijRepresents the impedance of the branch connecting node i and node j, and satisfies zij=rij+jxij;Representing the maximum value of the branch current connecting the node i and the node j;V jandrepresents the minimum and maximum voltages of node j;representing the maximum active output of the fan connected to node j;representing the power factor of the fan connected to node j.
Load of distribution network node jIncluding base loadAnd charging load of electric vehicleNamely, it is
According to the actual demand of the distribution network, the objective function min f (P, Q, V, I) can be finally defined as:
in the formula, aiAnd biRespectively representing the secondary coal consumption and the primary coal consumption coefficient of the generator;andthe electricity price and the active power of the electric quantity are respectively purchased from the main network.
Step 23, converting the optimal power flow model of the nonlinear power distribution network into a second-order cone relaxation planning model:
since BFM-OPF is a nonlinear programming model, letAndand performing second-order cone relaxation (SOCR) conversion on the formula to obtain the following model:
in the formula | · | non-conducting phosphor2Representing a second order cone operation; the above formula constitutes the basic form of the optimal power flow of the power distribution network after relaxation.
Step 24, solving the original problem and the dual variable of the model by adopting a Gurobi solver, and obtaining the marginal cost electrovalence lambda of the node where the charging station is locatedk。
Thirdly, generating an electric vehicle charging guide behavior strategy a according to the epsilon-Greedy algorithm and the graph neural network reinforcement learning algorithmi,t. The method mainly comprises the following steps:
step 31: and generating a random number u, and judging the size of the random number u and a decay factor xi of the epsilon-Greedy algorithm.
Step 32: if u<Xi, generating a behavior a for each electric automobile in the current state in a random modei,tThis behavior is expressed in the patent as an electric vehicle charging path strategy;
ai,t=randint(Naction) (19)
in the formula, NactionRepresenting the number of electric vehicle behavior decisions.
Step 33: if u is larger than or equal to xi, each electric automobile i is subjected to experience of the graph neural network reinforcement learning algorithmAt the current state xi,tAnd generating a behavior a under the adjacency matrix Ai,tI.e. by
In the formula, thetatParameters representing a graph neural network reinforcement learning algorithm; argmax () represents a parameter operation corresponding to the maximum value; x is the number ofi,tThe state of the ith electric vehicle at the time t is shown, which is mainly composed of the state x of the ith electric vehicle at the time ti,tFrom the electric vehicle state EVi,tAnd adjacent traffic road information Roi,tState Ne of neighboring electric vehiclei,tAnd each charging station information CStIs composed of, i.e.
xi,t=[EVi,t,Roi,t,Nei,t,CSt] (21)
In the formula, i-th electric vehicle state EVi,tIncluding the next node when the electric vehicle goes to the charging stationRoad numberingElectric steamVehicle running speed vi,tAnd remaining capacity SOCi,t(ii) a Neighbor traffic road information state Roi,tComprises the next node of the electric automobile iStarting node of the next road connectedEnd nodeRoad lengthAnd number of electric vehicles on roadNearest neighbor electric vehicle state Nei,tIncluding the status of each neighboring electric vehicle k, e.g. the next node of the kth electric vehicle adjacent to the ith electric vehicleThe road number of the roadElectric automobile driving speed vi,k,tAnd remaining capacity SOCi,k,t(ii) a Charging station information CStCharging tariff p comprising charging stationsc,tAnd number of electric vehicles
The graph neural network reinforcement learning algorithm comprises a neural network structure comprising an input layer, a full connection layer and an input state xi,tPerforming feature extraction xi,t', then the characteristic x to be proposedi,tInputting the two layers of the adjacent matrixes A into a two-layer graph neural network together, extracting features, and finally connecting a layer of full-connection layer to control the charging path of the electric automobileA little bit lessi,tAnd outputting the data. The graph neural network described in this patent employs a graph attention network.
Fourthly, executing a charging guide action strategy ai,tAnd the state of the electric automobile is judged and updated. The state of the electric vehicle is classified into three types: decision state, running state and charging state. If the electric automobile arrives at the intersection nodecur=nodenextAnd the intersection is not a charging station nodecur≠nodetarWhen the electric vehicle is in a decision-making state, the electric vehicle executes a charging guide action strategy ai,tUpdating road states such as the number of electric vehicles and ideal driving speed, and updating information such as the positions of the roads where the electric vehicles are located, driving speed, distance and the like; if the electric automobile does not reach the intersection nodecur≠nodenextWhen the electric vehicle is in the running state, namely the electric vehicle follows the charging guide strategy a of the previous stepi,t-1Continuing to drive forwards along the current road, and updating the position information, the speed information and the SOC state of the electric vehicle at the moment; if node on the charging station node at the node position of the electric automobilecur=nodetarAnd at the moment, the electric automobile is in a charging state, if the number of the current electric automobiles is larger than the number of the charging piles in the charging station, the electric automobiles need to wait in a queue for charging, if available charging piles in the charging station are used, the electric automobiles immediately charge, and the charging waiting time, the charging time and the SOC state of the electric automobiles are updated.
And fifthly, calculating a reward function of the neural network reinforcement learning algorithm according to the power-traffic fusion environment. In particular, the reward function ri,tIs a piecewise function: if the ith electric vehicle does not reach the charging station nodecur≠nodetarAnd the step number of the current electric vehicle going to the charging station is step < N within the given maximum charging step numberstepAt this time, its reward function ri,t0; if the number of steps of the ith electric vehicle going to the charging station is more than or equal to the given maximum charging step number step which is more than or equal to NstepWhen the charging action fails to be explored, a larger negative prize is given to the charging actionExciteri,t-penalty; if the ith electric vehicle arrives at the charging station nodecur=nodetarAnd the step number of the current electric vehicle going to the charging station is step < N within the given maximum charging step numberstepWhen the reward function is according to the driving time of the electric automobileCharging waiting timeCharging timeAnd calculating the electricity charge during charging, wherein the specific calculation expression is shown.
The passing time of the ith electric vehicle on the road section a is calculated according to the U.S. Federal road administration function (BPR), namely
In the formula, na,tRepresenting the number of electric vehicles on the road section a at the moment t; c. CaAndand respectively representing the upper capacity limit of the road section a and the free passing time of the electric automobile at the time t. Therefore, the time required for the ith electric vehicle to go to the charging station k can be obtainedNamely, it is
In the formula, SOCtRepresenting the residual capacity of the electric automobile;the rated capacity of the battery of the electric automobile is represented; eta represents a charging power factor, PchargingIndicating the rated power for charging the electric vehicle.
Sixthly, partially observing state x of Markov decision processi,tUpdating, including updating electric vehicle state EVi,tAnd adjacent traffic road information Roi,tState Ne of electric vehicle in close proximityi,tAnd each charging station information CSt。
Seventhly, information (x) of the current step is obtainedi,t,ai,t,ri,t,xi,t') is stored in the memory unit D, and the weights of the graph neural network reinforcement learning algorithm are updated based on the stochastic gradient descent method. The method mainly comprises the following steps:
step 71: randomly extracting a certain number of samples from the memory unit D;
step 72: constructing a loss function as shown in the formula, and updating the weight of the graph neural network reinforcement learning algorithm according to a random gradient descent method under the extracted Sample as shown in the formula;
wherein x, a, x 'and a' are the current state, action and the state and action at the next moment, respectively; thetatA graph neural network reinforcement learning algorithm parameter representing the current time t; gamma is greater than or equal to 0 and less than or equal to 1, representing a discount factor that reflects the impact of future Q values on the current action;representing parameter theta 'of neural network reinforcement learning algorithm in target graph'tState of state-action value.
In the formula, thetatA graph neural network reinforcement learning algorithm parameter representing the current time t;is expressed in the pair thetatCarrying out derivation operation; α represents a learning rate.
Step 73: the neural network reinforcement learning parameter theta is obtained according to the current graph after a certain number of stepstReinforcing learning parameter theta 'for neural network of target map'tAnd (6) updating.
Eighthly, judging whether the preset time T is reachedend. If not, executing (2) - (7); if yes, outputting the parameters of the graph neural network reinforcement learning algorithm and corresponding output results.
The invention discloses an electric vehicle charging guiding optimization method based on graph neural network reinforcement learning, which converts the mutual influence relation among electric vehicles into a dynamic network graph structure based on a graph theory, provides an attention mechanism-based graph neural network reinforcement learning method for processing irregular non-European structural data, and researches the communication and cooperation among multiple intelligent agents to discuss the mutual influence among the electric vehicles. On the basis of considering the active power distribution network of renewable energy output, the optimal power flow of the power distribution network is solved through a second-order cone optimization theory and a dual optimization theory, and the marginal cost electricity price of the power distribution network node is obtained, so that the power-traffic fusion network collaborative optimization is researched. The electric vehicle charging guiding optimization method based on graph neural network reinforcement learning can effectively reduce the total charging cost of the electric vehicle and realize ordered charging of the electric vehicle and cooperative optimization scheduling of a power system under the condition that multiple uncertainty factors of a power-traffic fusion network are considered.
The above-mentioned embodiments only express several embodiments of the present invention, and the description thereof is more specific and detailed, but not construed as limiting the scope of the present invention. It should be noted that, for a person skilled in the art, several variations and modifications can be made without departing from the inventive concept, which falls within the scope of the present invention. Therefore, the protection scope of the present patent shall be subject to the appended claims.
Claims (6)
1. The electric vehicle charging guidance optimization method based on graph neural network reinforcement learning is characterized by comprising the following steps:
step S1: initializing a power-traffic fusion network collaborative optimization model;
step S2: updating the charging load of the electric automobile, and carrying out optimization calculation on the marginal cost electricity price of the node where the electric automobile charging station is located based on second-order cone relaxation optimization and dual theory;
step S3: generating an electric vehicle charging guide behavior strategy a according to epsilon-Greedy algorithm and graph neural network reinforcement learning algorithmi,t;
Step S4: enforcing a charging bootstrapping behavior policy ai,tJudging and updating the state of the electric automobile;
step S5: reward function r of neural network reinforcement learning algorithm according to electric power-traffic fusion environment calculation diagrami,t;
Step S6: partially observing state x of Markov decision processi,tUpdating;
step S7: information (x) of the current stepi,t,ai,t,ri,t,xi,t') storing in a memory unit D, and updating the graph neural network reinforcement learning algorithm weight based on a random gradient descent method; wherein x isi,tShowing the current state of the graph neural network reinforcement learning; a isi,tRepresenting an electric vehicle behavior strategy; r isi,tThe reward function value represents the graph neural network reinforcement learning; x is the number ofi,t' represents the state of the next step of the graph neural network reinforcement learning;
step S8: judging whether the preset time T is reachedend(ii) a If not, executing (2) - (7); if yes, outputting the parameters of the graph neural network reinforcement learning algorithm and corresponding output results.
2. The electric vehicle charging guidance optimization method based on graph neural network reinforcement learning according to claim 1, wherein initializing a power-traffic fusion network collaborative optimization model comprises the following steps:
step 21: determining topological structures and parameters of a power network and a traffic network, wherein the topological structures and the parameters comprise power system nodes, lines, initial voltage and optimized upper and lower limit values, and the traffic network comprises traffic nodes, road parameters, capacity and maximum running speed;
step 22: neural network parameter initialization, including neural network weight initialization and hyper-parameter setting, such as learning rate alpha, discount factor gamma, batch size B and memory unit D capacity;
step 23: regarding each electric automobile in the research area as an agent, regarding the agent as a node N belongs to N, regarding the connection between the electric automobiles as an edge E belongs to E, forming a graph network structure G-N, E, and regarding each electric automobile i in the current state xi,tAnd the adjacency matrix a.
3. The electric vehicle charging guidance optimization method based on graph neural network reinforcement learning of claim 1, wherein the steps of updating the electric vehicle charging load and performing optimization calculation on the marginal cost electricity price of the node where the electric vehicle charging station is located based on the second-order cone relaxation optimization and the dual theory comprise:
step 31: updating the charging load of the electric automobile: calculating the charging load of each charging station according to the number of electric vehicles and the charging power in the charging stations, and adding the basic load of the node to the charging load of each station to obtain the final electric load of the node;
step 32: establishing a power distribution network optimal power flow model based on a branch power flow model:
minf(p,q,P,Q,V,I) (1)
s.t.
in the formula, ENAnd ELRespectively representing a power distribution network node and a line set; pijAnd QijRepresenting the active power and the reactive power of a branch flowing from the node i to the node j; pjkRepresenting the branch active power flowing from the node j to the node k;andrepresenting the active and reactive outputs of the generator, i.e. the active and reactive power injected into node j;andrepresenting the active power and the reactive power injected into the node j by the fan; qjsRepresenting the branch reactive power flowing from node j to node s; r isijAnd xijRepresents the branch resistance and reactance from node i to node j; i isijRepresents the branch current from node i to node j; π (j) represents the set of branches connected to node j;andrepresenting the active and reactive loads connected at node j; viRepresents the voltage amplitude of the node i; vjRepresents the voltage magnitude of node j; z is a radical ofijShow the connecting jointThe branch impedance of the point i and the node j satisfies zij=rij+jxij;Representing the maximum value of the branch current connecting the node i and the node j;V jandrepresents the minimum and maximum voltages of node j;representing the maximum active output of the fan connected to node j;represents a power factor of the fan connected to node j;
load of distribution network node jIncluding base loadAnd charging load of electric vehicleNamely, it is
According to the actual demand of the distribution network, the objective function minf (P, Q, V, I) can be finally defined as:
in the formula, aiAnd biRespectively representThe secondary coal consumption and the primary coal consumption coefficient of the generator;representing the active output of the generator injected into node i;andrespectively purchasing the electricity price and the active power of the electric quantity from the main network;
step 33: converting the optimal power flow model of the nonlinear power distribution network into a second-order cone relaxation planning model:
the BFM-OPF is a nonlinear programming model, so that the branch current amplitude valueAnd branch voltage amplitudeAnd performing second-order cone relaxation (SOCR) conversion on the formula to obtain the following model:
s.t.
in the formula | · | non-conducting phosphor2Representing a second order cone operation; the formula constitutes the basic form of the optimal power flow of the power distribution network after relaxation;
step 34: solving the original problem and the dual variable of the model by adopting a Gurobi solver to obtain the marginal cost electrovalence lambda of the node where the charging station is locatedk。
4. The electric vehicle charging guidance optimization method based on graph neural network reinforcement learning according to claim 3, characterized in that the epsilon-Greedy algorithm comprises the following steps:
step 41: generating a random number u, and judging the size of the random number u and a decay factor xi of the epsilon-Greedy algorithm;
step 42: if u<Xi, generating a behavior a for each electric automobile in the current state in a random modei,tThis behavior is expressed in the patent as an electric vehicle charging path strategy;
ai,t=randint(Naction) (19)
in the formula, NactionRepresenting the number of electric vehicle behavior decisions;
step 43: if u is larger than or equal to xi, each electric automobile i is in the current state x according to experience of the graph neural network reinforcement learning algorithmi,tAnd generating a behavior a under the adjacency matrix Ai,tI.e. by
In the formula, thetatParameters representing a graph neural network reinforcement learning algorithm; argmax () represents a parameter operation corresponding to the maximum value; x is the number ofi,tThe state of the ith electric vehicle at the time t is shown, which is mainly composed of the state x of the ith electric vehicle at the time ti,tFrom the electric vehicle state EVi,tAnd adjacent traffic road information Roi,tState Ne of electric vehicle in close proximityi,tAnd each charging station information CStIs composed of, i.e.
xi,t=[EVi,t,Roi,t,Nei,t,CSt] (21)
In the formula, i-th electric vehicle state EVi,tIncluding the next node when the electric vehicle goes to the charging stationRoad numberingElectric automobile driving speed vi,tAnd remainderElectric quantity SOCi,t(ii) a Neighbor traffic road information state Roi,tComprises the next node of the electric automobile iStarting node of the next road connectedEnd nodeRoad lengthAnd number of electric vehicles on roadNearest neighbor electric vehicle state Nei,tIncluding the status of each neighboring electric vehicle k, e.g. the next node of the kth electric vehicle adjacent to the ith electric vehicleThe road number of the roadElectric automobile driving speed vi,k,tAnd remaining capacity SOCi,k,t(ii) a Charging station information CStCharging tariff p comprising charging stationsc,tAnd number of electric vehicles
The graph neural network reinforcement learning algorithm comprises a neural network structure comprising an input layer, a full connection layer and an input state xi,tPerforming feature extraction xi,t', then the characteristic x to be proposedi,t' inputting the two-layer graph neural network together with the adjacency matrix A and then carrying out feature extractionTaking and finally connecting a full connecting layer to charge the electric automobilei,tOutputting is carried out; wherein, the graph neural network adopts a graph attention network.
5. The electric vehicle charging guidance optimization method based on graph neural network reinforcement learning of claim 2, characterized in that the reward function r of the graph neural network reinforcement learning algorithmi,tAs shown in the formula:
in the formula, a nodecurAnd a nodetarIndicating the current node of the electric automobile and any charging station node to which the electric automobile is going, and step indicating the number of steps the electric automobile has run; penalty represents a large penalty factor; w is aiIs shown asiCost per unit time of the electric vehicle;andrespectively representing the driving time, the charging waiting time and the charging required time when the ith electric vehicle goes to the kth charging station at the time t; lambdak,tRepresenting the marginal cost price of the node where the charging station k is located at the time t; SOCi,k,tRepresents the remaining capacity SOC when the ith electric vehicle reaches the charging station k at the time ti,k,t;The battery rated capacity of the ith electric automobile is represented;
slave typeThe reward function r can be seeni,tIs a piecewise function; if the ith electric vehicle does not reach the charging station nodecur≠nodetarAnd the step number of the current electric vehicle going to the charging station is step < N within the given maximum charging step numberstepAt this time, its reward function ri,t0; if the step number of the ith electric vehicle going to the charging station is more than or equal to the given maximum charging step number step which is more than or equal to NstepWhen the charging behavior is failed to be explored, a larger negative reward r is given to the charging behaviori,t-penalty; if the ith electric vehicle arrives at the charging station nodecur=nodetarAnd the step number of the current electric vehicle going to the charging station is step < N within the given maximum charging step numberstepWhen the reward function is according to the driving time of the electric automobileAnd charging timeAnd calculating the electricity charge during charging;
the passing time t of the ith electric automobile on the road section aa,tCalculated according to the U.S. Federal road administration function (BPR), i.e. the buseau of public roads
In the formula, na,tRepresenting the number of electric vehicles on the road section a at the moment t; c. CaAndrespectively representing the upper limit of the capacity of the road section a and the free passing time of the electric automobile at the time t; therefore, the time required for the ith electric vehicle to go to the charging station k can be obtainedNamely that
6. The electric vehicle charging guidance optimization method based on graph neural network reinforcement learning of claim 1, wherein the updating of the graph neural network reinforcement learning algorithm weight based on the stochastic gradient descent method comprises:
step 61: randomly extracting a certain number of samples from the memory unit D;
step 62: constructing a loss function as shown in the formula, and updating the weight of the graph neural network reinforcement learning algorithm according to a random gradient descent method under the extracted Sample as shown in the formula;
wherein x, a, x 'and a' are the current state, action and the state and action at the next moment, respectively; r represents the immediate reward of graph neural network reinforcement learning; thetatIndicates the currentA graph neural network reinforcement learning algorithm parameter at the moment t; gamma is greater than or equal to 0 and less than or equal to 1, representing a discount factor that reflects the impact of future Q values on the current action;representing parameter theta 'of neural network reinforcement learning algorithm in target graph'tState-action value of;
in the formula, thetatA graph neural network reinforcement learning algorithm parameter representing the current time t;is expressed in the pair thetatCarrying out derivation operation; α represents a learning rate;
and step 63: the neural network reinforcement learning parameter theta is obtained according to the current graph after a certain number of stepstReinforcing learning parameter theta 'for neural network of target map'tAnd (6) updating.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210109887.2A CN114444802B (en) | 2022-01-29 | 2022-01-29 | Electric vehicle charging guide optimization method based on graph neural network reinforcement learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210109887.2A CN114444802B (en) | 2022-01-29 | 2022-01-29 | Electric vehicle charging guide optimization method based on graph neural network reinforcement learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN114444802A true CN114444802A (en) | 2022-05-06 |
CN114444802B CN114444802B (en) | 2024-06-04 |
Family
ID=81372174
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210109887.2A Active CN114444802B (en) | 2022-01-29 | 2022-01-29 | Electric vehicle charging guide optimization method based on graph neural network reinforcement learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114444802B (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116436019A (en) * | 2023-04-12 | 2023-07-14 | 国网江苏省电力有限公司电力科学研究院 | Multi-resource coordination optimization method, device and storage medium |
CN118098000A (en) * | 2024-04-24 | 2024-05-28 | 哈尔滨华鲤跃腾科技有限公司 | Urban comprehensive management method based on artificial intelligence |
CN118438918A (en) * | 2024-05-09 | 2024-08-06 | 烟台开发区德联软件有限责任公司 | Mobile energy storage device charging regulation and control method and device |
WO2024165229A1 (en) * | 2023-02-08 | 2024-08-15 | E.On Se | Edge computing with ai at mains supply network interconnection points |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110570050A (en) * | 2019-09-25 | 2019-12-13 | 国网浙江省电力有限公司经济技术研究院 | Road-network-vehicle-related electric vehicle charging guiding method |
TWI687785B (en) * | 2019-02-25 | 2020-03-11 | 華碩電腦股份有限公司 | Method of returning to charging station |
CN111934335A (en) * | 2020-08-18 | 2020-11-13 | 华北电力大学 | Cluster electric vehicle charging behavior optimization method based on deep reinforcement learning |
WO2021143075A1 (en) * | 2020-01-17 | 2021-07-22 | 南京东博智慧能源研究院有限公司 | Demand response method taking space-time distribution of electric vehicle charging loads into consideration |
CN113159578A (en) * | 2021-04-22 | 2021-07-23 | 杭州电子科技大学 | Charging optimization scheduling method of large-scale electric vehicle charging station based on reinforcement learning |
CN113515884A (en) * | 2021-04-19 | 2021-10-19 | 国网上海市电力公司 | Distributed electric vehicle real-time optimization scheduling method, system, terminal and medium |
-
2022
- 2022-01-29 CN CN202210109887.2A patent/CN114444802B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TWI687785B (en) * | 2019-02-25 | 2020-03-11 | 華碩電腦股份有限公司 | Method of returning to charging station |
CN110570050A (en) * | 2019-09-25 | 2019-12-13 | 国网浙江省电力有限公司经济技术研究院 | Road-network-vehicle-related electric vehicle charging guiding method |
WO2021143075A1 (en) * | 2020-01-17 | 2021-07-22 | 南京东博智慧能源研究院有限公司 | Demand response method taking space-time distribution of electric vehicle charging loads into consideration |
CN111934335A (en) * | 2020-08-18 | 2020-11-13 | 华北电力大学 | Cluster electric vehicle charging behavior optimization method based on deep reinforcement learning |
CN113515884A (en) * | 2021-04-19 | 2021-10-19 | 国网上海市电力公司 | Distributed electric vehicle real-time optimization scheduling method, system, terminal and medium |
CN113159578A (en) * | 2021-04-22 | 2021-07-23 | 杭州电子科技大学 | Charging optimization scheduling method of large-scale electric vehicle charging station based on reinforcement learning |
Non-Patent Citations (1)
Title |
---|
夏冬: "多信息融合下电动汽车充电路径规划", 电测与仪器, vol. 57, no. 22, 25 December 2019 (2019-12-25), pages 24 - 32 * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2024165229A1 (en) * | 2023-02-08 | 2024-08-15 | E.On Se | Edge computing with ai at mains supply network interconnection points |
CN116436019A (en) * | 2023-04-12 | 2023-07-14 | 国网江苏省电力有限公司电力科学研究院 | Multi-resource coordination optimization method, device and storage medium |
CN116436019B (en) * | 2023-04-12 | 2024-01-23 | 国网江苏省电力有限公司电力科学研究院 | Multi-resource coordination optimization method, device and storage medium |
CN118098000A (en) * | 2024-04-24 | 2024-05-28 | 哈尔滨华鲤跃腾科技有限公司 | Urban comprehensive management method based on artificial intelligence |
CN118438918A (en) * | 2024-05-09 | 2024-08-06 | 烟台开发区德联软件有限责任公司 | Mobile energy storage device charging regulation and control method and device |
Also Published As
Publication number | Publication date |
---|---|
CN114444802B (en) | 2024-06-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN114444802A (en) | Electric vehicle charging guide optimization method based on graph neural network reinforcement learning | |
CN110570050B (en) | Road-network-vehicle-related electric vehicle charging guiding method | |
Sun et al. | A customized voltage control strategy for electric vehicles in distribution networks with reinforcement learning method | |
Luo et al. | Optimal charging scheduling for large-scale EV (electric vehicle) deployment based on the interaction of the smart-grid and intelligent-transport systems | |
Xu et al. | A soft actor-critic-based energy management strategy for electric vehicles with hybrid energy storage systems | |
Wang et al. | Optimising route choices for the travelling and charging of battery electric vehicles by considering multiple objectives | |
Luo et al. | Joint deployment of charging stations and photovoltaic power plants for electric vehicles | |
Panahi et al. | Forecasting plug-in electric vehicles load profile using artificial neural networks | |
CN109523051A (en) | A kind of electric car charging Real time optimal dispatch method | |
CN103241130A (en) | Energy management method and system for electric bus charging and swap station | |
CN106875075A (en) | A kind of electric automobile charging station points distributing method based on travel behaviour | |
CN113887858A (en) | Charging station micro-grid system optimal scheduling method based on CNN-LSTM load prediction | |
CN103078152B (en) | Intelligent charging method for centralized charging station | |
CN112330025B (en) | Prediction method of space-time charging load for urban electric vehicle | |
Qian et al. | Shadow-price DRL: A framework for online scheduling of shared autonomous EVs fleets | |
CN110276517A (en) | A kind of electric automobile charging station site selecting method based on MOPSO algorithm | |
Kiaee | Integration of electric vehicles in smart grid using deep reinforcement learning | |
Rahmani-Andebili et al. | Optimal incentive plans for plug-in electric vehicles | |
CN111626492B (en) | Fuzzy multi-target opportunity constraint planning method for electric vehicle charging network | |
Zhang et al. | A safe reinforcement learning-based charging strategy for electric vehicles in residential microgrid | |
CN112381359A (en) | Multi-critic reinforcement learning power economy scheduling method based on data mining | |
CN110197310B (en) | Charging and battery replacing station optimal scheduling method based on load margin domain | |
Fan et al. | Optimal shared mobility planning for electric vehicles in the distribution network | |
Pourvaziri et al. | Planning of electric vehicle charging stations: An integrated deep learning and queueing theory approach | |
Chen et al. | A Deep Reinforcement Learning-Based Charging Scheduling Approach with Augmented Lagrangian for Electric Vehicle |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |