CN117032134A - Multi-AGV scheduling strategy evaluation and optimization method and system based on approximate synchronization estimation - Google Patents
Multi-AGV scheduling strategy evaluation and optimization method and system based on approximate synchronization estimation Download PDFInfo
- Publication number
- CN117032134A CN117032134A CN202311099265.7A CN202311099265A CN117032134A CN 117032134 A CN117032134 A CN 117032134A CN 202311099265 A CN202311099265 A CN 202311099265A CN 117032134 A CN117032134 A CN 117032134A
- Authority
- CN
- China
- Prior art keywords
- strategy
- agv
- agvs
- model
- optimization
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 71
- 238000005457 optimization Methods 0.000 title claims abstract description 59
- 238000011156 evaluation Methods 0.000 title claims abstract description 42
- 230000006870 function Effects 0.000 claims abstract description 74
- 238000012549 training Methods 0.000 claims abstract description 29
- 238000013210 evaluation model Methods 0.000 claims abstract description 26
- 230000008901 benefit Effects 0.000 claims abstract description 21
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 16
- 238000005520 cutting process Methods 0.000 claims abstract description 16
- 230000001360 synchronised effect Effects 0.000 claims abstract description 14
- 230000007613 environmental effect Effects 0.000 claims abstract description 8
- 230000009471 action Effects 0.000 claims description 49
- 238000009826 distribution Methods 0.000 claims description 22
- 230000000875 corresponding effect Effects 0.000 claims description 20
- 230000015654 memory Effects 0.000 claims description 20
- 230000008569 process Effects 0.000 claims description 17
- 238000004590 computer program Methods 0.000 claims description 13
- 238000003860 storage Methods 0.000 claims description 13
- 230000003993 interaction Effects 0.000 claims description 12
- 230000007704 transition Effects 0.000 claims description 11
- 238000004364 calculation method Methods 0.000 claims description 9
- 230000002787 reinforcement Effects 0.000 claims description 8
- 238000005070 sampling Methods 0.000 claims description 7
- 238000010276 construction Methods 0.000 claims description 6
- 238000013528 artificial neural network Methods 0.000 claims description 5
- 230000003213 activating effect Effects 0.000 claims description 4
- 238000012935 Averaging Methods 0.000 claims description 3
- 230000001276 controlling effect Effects 0.000 claims description 3
- 125000004122 cyclic group Chemical group 0.000 claims description 3
- 230000009916 joint effect Effects 0.000 claims description 3
- 239000003795 chemical substances by application Substances 0.000 description 21
- 238000010586 diagram Methods 0.000 description 9
- 230000006872 improvement Effects 0.000 description 6
- 238000004891 communication Methods 0.000 description 4
- 238000012545 processing Methods 0.000 description 4
- 230000006399 behavior Effects 0.000 description 3
- 239000012723 sample buffer Substances 0.000 description 3
- 238000004088 simulation Methods 0.000 description 3
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000000306 recurrent effect Effects 0.000 description 2
- 239000000523 sample Substances 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 238000000342 Monte Carlo simulation Methods 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000010399 physical interaction Effects 0.000 description 1
- 230000001953 sensory effect Effects 0.000 description 1
- 239000004984 smart glass Substances 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B19/00—Programme-control systems
- G05B19/02—Programme-control systems electric
- G05B19/418—Total factory control, i.e. centrally controlling a plurality of machines, e.g. direct or distributed numerical control [DNC], flexible manufacturing systems [FMS], integrated manufacturing systems [IMS] or computer integrated manufacturing [CIM]
- G05B19/4189—Total factory control, i.e. centrally controlling a plurality of machines, e.g. direct or distributed numerical control [DNC], flexible manufacturing systems [FMS], integrated manufacturing systems [IMS] or computer integrated manufacturing [CIM] characterised by the transport system
- G05B19/41895—Total factory control, i.e. centrally controlling a plurality of machines, e.g. direct or distributed numerical control [DNC], flexible manufacturing systems [FMS], integrated manufacturing systems [IMS] or computer integrated manufacturing [CIM] characterised by the transport system using automatic guided vehicles [AGV]
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/31—From computer integrated manufacturing till monitoring
- G05B2219/31002—Computer controlled agv conveys workpieces between buffer and cell
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Manufacturing & Machinery (AREA)
- Quality & Reliability (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Automation & Control Theory (AREA)
- Information Transfer Between Computers (AREA)
Abstract
The application discloses a multi-AGV scheduling strategy evaluation and optimization method and system based on approximate synchronous estimation, wherein the method comprises the following steps: constructing a strategy model shared by multiple AGVs and a combined strategy evaluation model; based on a strategy model shared by multiple AGVs and a combined strategy evaluation model, enabling the AGVs to interact with the environment to generate training samples, and calculating edge dominance functions; calculating an update amplitude cut-off according to the uncertainty of the AGV strategy; and under any environmental state, replacing a joint advantage function according to the edge advantage function, and combining a near-end strategy optimization algorithm, taking the gradient cut-off quantity as a strategy ratio cutting interval, and performing independent optimization on the strategy of each AGV until the multi-AGV scheduling strategy is optimized. The method can realize stable and accurate strategy evaluation in the AGV scheduling task, and improves the training stability and scheduling efficiency.
Description
Technical Field
The application belongs to the field of AGV scheduling, and particularly relates to a multi-AGV scheduling strategy evaluation and optimization method and system based on approximate synchronous estimation.
Background
Decision tasks in many real-world scenarios, such as traffic scheduling, multi-sensor collaboration, robot cluster collaboration, etc., can be modeled as multi-agent collaborative decision-making problems. In such problems, in order to ensure independence between agents and completeness of information during training, a training framework of centralized evaluation-decentralized execution is generally adopted, and the framework needs to perform differential evaluation on agent policies according to a shared reward function.
In the policy-based deep reinforcement learning method, the policy of each agent is generally evaluated through a centralized reviewer network. The criticizing network takes the environmental state and the past counter experience and the strategy of each agent as input, and the evaluation result of the current agent strategy depends on the past specific behaviors of other agents. However, for the current agent, other agents are considered part of the environment, whose behavior is subject to uncertainty. Due to this uncertainty, the same behavior of the current agent may be evaluated differently depending on the counterfacts experience, thereby affecting the stability of the training process. In addition, because other agent strategies are continuously updated, the new strategies deviate from the past experience distribution, and the experience and the strategies are not synchronous, so that a certain error exists in the evaluation result.
How to reduce the variance from uncertainty in multi-unmanned truck (Automated Guided Vehicle, AGV) strategy evaluation is of great importance to the stability of the training process; at the same time, achieving synchronization of experience and strategy in strategy evaluation is critical to reducing evaluation errors.
Disclosure of Invention
The application aims to overcome the defects and provide a multi-AGV scheduling strategy evaluation and optimization method and system based on approximate synchronous estimation.
In order to achieve the above purpose, the application adopts the following technical scheme:
the first aspect of the present application provides a multi-AGV scheduling policy evaluation and optimization method based on approximate synchronization estimation, comprising:
constructing a strategy model shared by multiple AGVs and a combined strategy evaluation model;
based on a strategy model shared by multiple AGVs and a combined strategy evaluation model, enabling the AGVs to interact with the environment to generate training samples, and calculating edge dominance functions;
calculating an update amplitude cut-off according to the uncertainty of the AGV strategy;
and under any environmental state, replacing a joint advantage function according to the edge advantage function, and combining a near-end strategy optimization algorithm, taking the gradient cut-off quantity as a strategy ratio cutting interval, and performing independent optimization on the strategy of each AGV until the multi-AGV scheduling strategy is optimized.
As a further improvement of the application, the construction of the multi-AGV shared strategy model and the combined strategy evaluation model comprises the following steps:
the method comprises the steps of taking historical track information, current observation information and AGV marks of each AGV as input, inputting the probability of the action of the AGVs, and constructing a decision model shared by a plurality of AGVs through a cyclic neural network;
and taking the actions and the environmental states of all AGVs as inputs, inputting a joint action-state value function, and constructing a joint strategy evaluation model based on the critics network.
As a further improvement of the application, the strategy model and the combined strategy evaluation model based on the sharing of multiple AGVs, which enables the AGVs to interact with the environment to generate training samples, calculate edge dominance functions, comprise:
based on a strategy model shared by multiple AGVs and a combined strategy evaluation model, fixing the actions of the AGVs to be evaluated, and sampling multiple groups of actions according to the combined strategies of all other AGVs except the AGVs;
combining the sampled multiple groups of actions with the action of the AGV to be evaluated, and simulating an interaction process to generate a group of state transition data;
inputting the state transition data into a joint strategy evaluation model, and calculating a joint action-state value function corresponding to each transition data according to a calculation method of the reinforcement learning median function;
averaging the combined action-state value functions to obtain an edge dominance function of the AGV to be evaluated;
and repeating the steps for each AGV to obtain the edge dominance function of each AGV.
As a further improvement of the application, the simulation interaction process needs to construct a state transition model and a rewarding function model, wherein the state transition model and the rewarding function model take global states and joint actions as inputs.
As a further improvement of the present application, the calculating the update amplitude cut-off according to the uncertainty of the AGV policy includes:
and obtaining the action distribution of each AGV through a strategy network, calculating the variance of the action distribution, and taking the calculation result as the updated gradient cut-off quantity of the corresponding AGV.
As a further improvement of the present application, the obtaining, by the policy network, the motion profile of each AGV includes:
generating corresponding action masks according to the executable action list of the AGVs, and multiplying the output of each AGV strategy network by the corresponding action masks to obtain the value function estimation of legal actions;
and then activating a function through Geng Beier to obtain differentiable probability distribution, introducing a variable parameter t, and dynamically adjusting the probability distribution of different legal action value functions to obtain the action distribution of each AGV.
As a further improvement of the present application, the individual optimization of the policy of each AGV includes:
and using an edge dominance function as an update baseline, combining a near-end strategy optimization algorithm, adding an update gradient cut-off quantity as an update constraint, cutting and updating an AGV strategy network through a strategy ratio, controlling the update amplitude of the AGVs through a cutting interval, and performing independent optimization on the strategy of each AGV.
A second aspect of the present application provides a system for evaluating and optimizing a multi-AGV scheduling policy based on approximate synchronization estimation, including:
the construction module is used for constructing a strategy model shared by a plurality of AGVs and a combined strategy evaluation model;
the interaction module is used for enabling the AGVs to interact with the environment to generate training samples based on the strategy model shared by the AGVs and the combined strategy evaluation model, and calculating edge dominance functions;
the uncertainty module is used for calculating an update amplitude cut-off according to the uncertainty of the AGV strategy;
and the optimization module is used for replacing the joint advantage function according to the edge advantage function under any environment state, combining a near-end strategy optimization algorithm, taking the gradient cut-off quantity as a strategy ratio cutting interval, and performing independent optimization on the strategy of each AGV until the multi-AGV scheduling strategy is optimized.
A third aspect of the present application provides an electronic device, including a memory, a processor, and a computer program stored in the memory and executable on the processor, where the processor implements the multiple AGV scheduling policy evaluation and optimization method based on approximate synchronization estimation when executing the computer program.
A fourth aspect of the present application is to provide a computer readable storage medium storing a computer program, which when executed by a processor implements the multi-AGV scheduling policy evaluation and optimization method based on approximate synchronization estimation.
Compared with the prior art, the application has the following technical effects:
according to the AGV scheduling strategy evaluation and optimization method based on approximate synchronous estimation, the variance of strategy estimation is reduced by improving the number of samples, so that the stability of the training process is improved. Meanwhile, by means of uncertainty estimation on the policies of the agents, the policy updating amplitude of each agent is dynamically restrained, approximate synchronous policy estimation is achieved, and errors of policy estimation are reduced. In the multi-AGV scheduling task sharing the reward function, the edge advantage function is obtained by integrating the joint value function according to the strategies of different AGVs, and the differential evaluation of the different AGVs under the shared reward function is realized, so that the centralized scheduling of the multi-AGV system is completed.
Drawings
FIG. 1 is a flow chart of a multi-AGV scheduling strategy evaluation and optimization method based on approximate synchronization estimation;
FIG. 2 is a method framework diagram presented by an embodiment of the present application;
FIG. 3 is a schematic diagram of an edge dominance estimation method according to the present application;
fig. 4 is a schematic diagram of the approximate synchronization estimation method of the present application.
FIG. 5 is a schematic illustration of a system for evaluating and optimizing multiple AGV scheduling strategies based on approximate synchronization estimation in accordance with the present application;
fig. 6 is a schematic diagram of an electronic device according to the present application.
Detailed Description
Exemplary embodiments of the present application will now be described with reference to the accompanying drawings, in which various details of the embodiments of the present application are included to facilitate understanding, and are to be considered merely exemplary. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the application. Also, descriptions of well-known functions and constructions are omitted in the following description for clarity and conciseness.
It will be apparent that the described embodiments are some, but not all, embodiments of the application. All other embodiments, which can be made by those skilled in the art based on the embodiments of the application without making any inventive effort, are intended to be within the scope of the application.
It should be noted that, the terminal according to the embodiment of the present application may include, but is not limited to, a mobile phone, a personal digital assistant (Personal Digital Assistant, PDA), a wireless handheld device, a Tablet Computer (Tablet Computer), a personal Computer (Personal Computer, PC), an MP3 player, an MP4 player, a wearable device (e.g., smart glasses, smart watches, smart bracelets, etc.), a smart home device, and other smart devices.
In addition, the term "and/or" herein is merely an association relationship describing an association object, and means that three relationships may exist, for example, a and/or B may mean: a exists alone, A and B exist together, and B exists alone. In addition, the character "/" herein generally indicates that the front and rear associated objects are an "or" relationship.
As shown in fig. 1, a first object of the present application is to provide a multi-AGV scheduling policy evaluation and optimization method based on approximate synchronization estimation, which includes the following steps:
s1, constructing a strategy model shared by multiple AGVs and a combined strategy evaluation model;
s2, based on a strategy model shared by multiple AGVs and a combined strategy evaluation model, enabling the AGVs to interact with the environment to generate training samples, and calculating edge dominance functions;
s3, calculating an update amplitude cut-off quantity according to the uncertainty of the AGV strategy;
s4, under any environment state, replacing a joint advantage function according to the edge advantage function, combining a near-end strategy optimization algorithm, taking the gradient cut-off quantity as a strategy ratio cutting interval, and performing independent optimization on each AGV strategy until the multi-AGV scheduling strategy is optimized.
The multi-AGV scheduling strategy evaluation and optimization method based on the approximate synchronous estimation is based on edge advantage estimation, and the variance of strategy estimation is reduced by improving the number of samples, so that the stability of the training process is improved. Meanwhile, by means of uncertainty estimation on the policies of the agents, the policy updating amplitude of each agent is dynamically restrained, approximate synchronous policy estimation is achieved, and errors of policy estimation are reduced. The algorithm performance is improved considerably on a general multi-agent reinforcement learning test platform. The method can realize stable and accurate strategy evaluation in the AGV scheduling task, and improves the training stability and scheduling efficiency.
Alternatively, in the step, calculating the update amplitude cutoff amount according to the uncertainty of the AGV policy includes:
and obtaining the action distribution of each AGV through a strategy network, calculating the variance of the action distribution, and taking the calculation result as the updated gradient cut-off quantity of the corresponding AGV.
The obtaining, by the policy network, the action distribution of each AGV includes:
generating corresponding action masks according to the executable action list of the AGVs, and multiplying the output of each AGV strategy network by the corresponding action masks to obtain the value function estimation of legal actions;
and then activating a function through Geng Beier to obtain differentiable probability distribution, introducing a variable parameter t, and dynamically adjusting the probability distribution of different legal action value functions to obtain the action distribution of each AGV.
In order to realize synchronous estimation of different AGV strategies under a multi-agent training framework of centralized evaluation-decentralized execution, the method takes the variance of each AGV strategy as the standard of action uncertainty of an agent, and limits the strategy change within a certain range through strategy update amplitude cutoff according to the size of the uncertainty, so that approximate synchronous estimation is realized.
Alternatively, in the above steps, the policy of each AGV is individually optimized, including:
and using an edge dominance function as an update baseline, combining a near-end strategy optimization algorithm, adding an update gradient cut-off quantity as an update constraint, cutting and updating an AGV strategy network through a strategy ratio, controlling the update amplitude of the AGVs through a cutting interval, and performing independent optimization on the strategy of each AGV.
The method is based on a deep reinforcement learning algorithm, and the variance of AGV strategy evaluation is reduced through multi-sample estimation; the update amplitude truncation is utilized to approximate and simplify the synchronous estimation, and the policy evaluation accuracy and training efficiency are ensured. The method can be applied to the AGV system scheduling task based on online policy multi-agent reinforcement learning so as to improve the accuracy of each AGV scheduling policy evaluation and the stability of the training process.
The application is described in detail below with reference to the attached drawings and specific examples:
referring to fig. 2, the agent is an AGV, and the embodiment provides a multi-AGV scheduling policy evaluation and optimization method based on approximate synchronization estimation, which includes the following steps:
step one: the method comprises the following steps of constructing a strategy model and a combined strategy evaluation model shared by a plurality of AGVs:
a strategy model shared by a plurality of AGVs is constructed through a cyclic neural network, and the historical track information (comprising historical actions and historical observation information) of each AGV, the current observation information and the AGV identification are taken as inputs, so that the probability of the action of the AGV is input; and constructing a joint strategy evaluation model based on the critic network, and inputting a joint action-state value function by taking the actions of all AGVs and the environmental states as inputs.
Step two: based on a strategy model shared by multiple AGVs and a combined strategy evaluation model, enabling the AGVs to interact with the environment to generate training samples, and calculating edge dominance functions. Referring to fig. 3, the specific process is as follows:
1) The method comprises the steps of fixing the actions of an AGV to be evaluated based on a strategy model shared by multiple AGVs, and sampling multiple groups of actions according to the combined strategies of all other AGVs except the AGV; the interaction samples may be disabled in multiple groups.
2) Combining the sampling action with the action of the AGV to be evaluated, and generating a group of state transition data by simulating the interaction process in the previous step; and maintaining the sample to obtain a sample buffer pool.
3) Calculating a joint action-state value function corresponding to each transfer data according to a calculation method of the reinforcement learning median function;
4) Averaging the set of value functions to obtain an edge dominance function of the AGV to be evaluated;
5) The above steps 1) through 4) are repeated for each AGV.
In the second step, in order to simulate the interaction process of the AGV and the environment, a state transition model and a reward function model need to be constructed, and the two models take a global state and a joint action as input.
The sample buffer pool is used for randomly sampling training samples, performing criticism network training, performing TD-residual estimation to obtain a counter fact advantage estimation, and performing a Monte Carlo method to obtain an edge advantage estimation.
Step three: and calculating the update amplitude cut-off according to the uncertainty of the AGV strategy. The specific process is as follows:
and obtaining the action distribution of each AGV through a strategy network, calculating the variance of the action distribution, and taking the calculation result as the updated gradient cut-off quantity of the corresponding AGV.
As an alternative, in the third step, as shown in fig. 4, since the AGV can only execute some specific actions in its action set, in order to improve the action sampling efficiency, a corresponding action mask is generated according to the executable action list of the AGV, and the output of each AGV policy network is multiplied by the corresponding action mask to obtain a value function estimate of legal actions. And then, a differentiable probability distribution is obtained by activating a function through Geng Beier, and the variable parameter t is introduced in the work and is used for dynamically adjusting the probability distribution of different legal action value functions, so that the distribution is gradually gentle along with the training process, and the training process is more stable.
Step four: and under any environmental state, replacing the joint advantage function according to the edge advantage function obtained in the second step, combining a near-end strategy optimization algorithm, taking the gradient cut-off quantity obtained in the third step as a strategy ratio cutting interval, and performing independent optimization on the strategy of each AGV.
Alternatively, in the fourth step, the algorithm uses the edge dominance function estimate obtained in the second step as an update baseline, and adds the variance estimate in the third step as an update constraint. Because the hessian matrix corresponding to the KL divergence is needed to be calculated for directly solving the optimization problem, the calculation cost is high, and the training efficiency is low.
According to the embodiment, with reference to the near-end policy optimization method, an AGV policy network is updated through policy ratio cutting, and the AGV updating amplitude is controlled by a cutting interval. When the dominance function estimates positive, the AGV lifts the probability of the corresponding action, but the lifting amplitude does not exceed the set threshold.
More specifically, as shown in fig. 4, the sample buffer pool performs random sampling training samples, performs strategy network training to obtain track action probabilities of 1 to n, performs gradient constraint calculation to obtain a track n strategy gradient, and further obtains a strategy network optimization result.
The multi-AGV scheduling strategy evaluation and optimization method based on the approximate synchronous estimation is based on edge advantage estimation, and the variance of strategy estimation is reduced by improving the number of samples, so that the stability of the training process is improved. Meanwhile, by means of uncertainty estimation of the AGV strategies, the strategy updating amplitude of each AGV is dynamically restrained, approximate synchronous strategy estimation is achieved, and errors of strategy estimation are reduced. The algorithm performance is improved considerably on a general multi-AGV reinforcement learning test platform.
The shared multi-agent policy network may be implemented by a multi-layer perceptron and a recurrent neural network, where the input layer of the network is the recurrent neural network and the output layer is the multi-layer perceptron. The commentator network is implemented by a multi-layer perceptron. The output layer scale of the strategy network and the critics network can be adjusted according to the object task requirement, so that the strategy fitting and evaluation capability of different degrees can be realized. The experience playback is realized by a readable memory unit.
The algorithm related by the application can be deployed on any machine with a storage unit and floating point operation functions. For a simulation task, a corresponding simulation environment needs to be deployed on a machine; for tasks in real-world scenarios, it is also necessary to deploy corresponding sensors and action actuators for physical interaction with the environment.
As a specific embodiment, as shown in fig. 5, a second object of the present application is to provide a multi-AGV scheduling policy evaluation and optimization system based on approximate synchronization estimation, including:
the construction module is used for constructing a strategy model shared by a plurality of AGVs and a combined strategy evaluation model;
the interaction module is used for enabling the AGVs to interact with the environment to generate training samples based on the strategy model shared by the AGVs and the combined strategy evaluation model, and calculating edge dominance functions;
the uncertainty module is used for calculating an update amplitude cut-off according to the uncertainty of the AGV strategy;
and the optimization module is used for replacing the joint advantage function according to the edge advantage function under any environment state, combining a near-end strategy optimization algorithm, taking the gradient cut-off quantity as a strategy ratio cutting interval, and performing independent optimization on the strategy of each AGV until the multi-AGV scheduling strategy is optimized.
According to an embodiment of the present application, there is also provided an electronic device and a non-transitory computer-readable storage medium storing computer instructions.
FIG. 6 is a schematic diagram of an electronic device for implementing a multi-AGV scheduling policy evaluation and optimization method based on near-synchrony estimation in accordance with an embodiment of the present application. Electronic devices are intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other appropriate computers. The electronic device may also represent various forms of mobile devices, such as personal digital processing, cellular telephones, smartphones, wearable devices, and other similar computing devices. The components shown herein, their connections and relationships, and their functions, are meant to be exemplary only, and are not meant to limit implementations of the applications described and/or claimed herein.
As shown in fig. 6, the electronic device includes: one or more processors 501, memory 502, and interfaces for connecting components, including high-speed interfaces and low-speed interfaces. The various components are interconnected using different buses and may be mounted on a common motherboard or in other manners as desired. The processor may process instructions executing within the electronic device, including instructions stored in or on memory to display graphical information of a GUI (graphical user interface) on an external input/output device, such as a display device coupled to the interface. In other embodiments, multiple processors and/or multiple buses may be used, if desired, along with multiple memories and multiple memories. Also, multiple electronic devices may be connected, each providing a portion of the necessary operations (e.g., as a server array, a set of blade servers, or a multiprocessor system). One processor 501 is illustrated in fig. 6.
Memory 502 is a non-transitory computer readable storage medium provided by the present application. The memory stores instructions executable by the at least one processor to cause the at least one processor to perform the multi-AGV scheduling policy evaluation and optimization method based on approximate synchronization estimation provided by the application. The non-transitory computer readable storage medium of the present application stores computer instructions for causing a computer to perform the multi-AGV scheduling policy evaluation and optimization method based on approximate synchronization estimation provided by the present application.
The memory 502 is used as a non-transitory computer readable storage medium for storing non-transitory software programs, non-transitory computer executable programs, and units, such as program instructions/units corresponding to the optimization method for multi-AGV scheduling policy evaluation based on approximate synchronization estimation in the embodiment of the application. The processor 501 executes various functional applications of the server and data processing by running non-transitory software programs, instructions and units stored in the memory 502, i.e., implements the multi-AGV scheduling policy evaluation and optimization method based on approximate synchronization estimation in the above-described method embodiment.
Memory 502 may include a storage program area that may store an operating system, at least one application program required for functionality, and a storage data area; the storage data area may store data created by the use of the electronic device according to the multi-AGV scheduling policy evaluation and optimization method based on the approximate synchronization estimation provided by the embodiment of the present application. In addition, memory 502 may include high-speed random access memory, and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid-state storage device. In some embodiments, the memory 502 may optionally include memory remotely located with respect to the processor 501, which may be connected via a network to the electronic device implementing the multiple AGV scheduling policy evaluation and optimization method based on near-synchronous estimation provided by embodiments of the present application. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof.
The electronic device of the multi-AGV scheduling policy evaluation and optimization method based on the approximate synchronization estimation may further include: an input device 503 and an output device 504. The processor 501, memory 502, input devices 503 and output devices 504 may be connected by a bus or otherwise, for example in fig. 6.
The input device 503 may receive input numeric or character information and generate key signal inputs related to user settings and function control of the electronic device implementing the multi-AGV scheduling policy evaluation and optimization method based on the near-synchrony estimation provided by the embodiments of the present application, such as input devices for a touch screen, a keypad, a mouse, a track pad, a touch pad, a pointer stick, one or more mouse buttons, a track ball, a joystick, etc. The output devices 504 may include a display device, auxiliary lighting devices (e.g., LEDs), and haptic feedback devices (e.g., vibration motors), among others. The display device may include, but is not limited to, an LCD (liquid crystal display), an LED (light emitting diode) display, and a plasma display. In some implementations, the display device may be a touch screen.
Various implementations of the systems and techniques described here can be realized in digital electronic circuitry, integrated circuitry, ASIC (application specific integrated circuits), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include: implemented in one or more computer programs, the one or more computer programs may be executed and/or interpreted on a programmable system including at least one programmable processor, which may be a special purpose or general-purpose programmable processor, that may receive data and instructions from, and transmit data and instructions to, a storage system, at least one input device, and at least one output device.
These computing programs (also referred to as programs, software applications, or code) include machine instructions for a programmable processor, and may be implemented in a high-level procedural and/or object-oriented programming language, and/or in assembly/machine language. As used herein, the terms "machine-readable medium" and "computer-readable medium" refer to any computer program product, apparatus, and/or device (e.g., magnetic discs, optical disks, memory, PLDs (programmable logic devices)) used to provide machine instructions and/or data to a programmable processor, including a machine-readable medium that receives machine instructions as a machine-readable signal. The term "machine-readable signal" refers to any signal used to provide machine instructions and/or data to a programmable processor.
To provide for interaction with a user, the systems and techniques described here can be implemented on a computer having: a display device (e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to a user; and a keyboard and pointing device (e.g., a mouse or trackball) by which a user can provide input to the computer. Other kinds of devices may also be used to provide for interaction with a user; for example, feedback provided to the user may be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and input from the user may be received in any form, including acoustic input, speech input, or tactile input.
The systems and techniques described here can be implemented in a computing system that includes a background component (e.g., as a data server), or that includes a middleware component (e.g., an application server), or that includes a front-end component (e.g., a user computer having a graphical user interface or a web browser through which a user can interact with an implementation of the systems and techniques described here), or any combination of such background, middleware, or front-end components. The components of the system can be interconnected by any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include: LAN (local area network), WAN (wide area network), internet and blockchain networks.
The computer system may include a client and a server. The client and server are typically remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
Finally, it should be noted that: the above embodiments are only for illustrating the technical aspects of the present application and not for limiting the same, and although the present application has been described in detail with reference to the above embodiments, it should be understood by those of ordinary skill in the art that: modifications and equivalents may be made to the specific embodiments of the application without departing from the spirit and scope of the application, which is intended to be covered by the claims.
Claims (10)
1. The multi-AGV scheduling strategy evaluation and optimization method based on the approximate synchronization estimation is characterized by comprising the following steps of:
constructing a strategy model shared by multiple AGVs and a combined strategy evaluation model;
based on a strategy model shared by multiple AGVs and a combined strategy evaluation model, enabling the AGVs to interact with the environment to generate training samples, and calculating edge dominance functions;
calculating an update amplitude cut-off according to the uncertainty of the AGV strategy;
and under any environmental state, replacing a joint advantage function according to the edge advantage function, and combining a near-end strategy optimization algorithm, taking the gradient cut-off quantity as a strategy ratio cutting interval, and performing independent optimization on the strategy of each AGV until the multi-AGV scheduling strategy is optimized.
2. The method for evaluating and optimizing a multi-AGV scheduling policy based on approximate synchronization estimation according to claim 1, wherein the constructing a multi-AGV shared policy model and a joint policy evaluation model comprises:
the method comprises the steps of taking historical track information, current observation information and AGV marks of each AGV as input, inputting the probability of the action of the AGVs, and constructing a decision model shared by a plurality of AGVs through a cyclic neural network;
and taking the actions and the environmental states of all AGVs as inputs, inputting a joint action-state value function, and constructing a joint strategy evaluation model based on the critics network.
3. The method for evaluating and optimizing a multi-AGV scheduling policy based on approximate synchronization estimation according to claim 1, wherein the method for evaluating the policy model and the combined policy based on sharing of the multi-AGV, for the AGV to interact with the environment to generate training samples, calculates edge dominance functions, comprises:
based on a strategy model shared by multiple AGVs and a combined strategy evaluation model, fixing the actions of the AGVs to be evaluated, and sampling multiple groups of actions according to the combined strategies of all other AGVs except the AGVs;
combining the sampled multiple groups of actions with the action of the AGV to be evaluated, and simulating an interaction process to generate a group of state transition data;
inputting the state transition data into a joint strategy evaluation model, and calculating a joint action-state value function corresponding to each transition data according to a calculation method of the reinforcement learning median function;
averaging all the combined action-state value functions to obtain an edge dominance function of the AGV to be evaluated;
and repeating the steps for each AGV to obtain the edge dominance function of each AGV.
4. The method for evaluating and optimizing multi-AGV scheduling policies based on approximately synchronous estimation according to claim 3 wherein said simulated interaction process requires the construction of a state transition model and a reward function model that take global state and joint actions as inputs.
5. The method for evaluating and optimizing a multi-AGV scheduling policy based on approximate synchronization estimation according to claim 1, wherein the calculating the update magnitude cutoff according to the uncertainty of the AGV policy comprises:
and obtaining the action distribution of each AGV through a strategy network, calculating the variance of the action distribution, and taking the calculation result as the updated gradient cut-off quantity of the corresponding AGV.
6. The method for evaluating and optimizing a multi-AGV scheduling policy based on approximate synchronization estimation according to claim 5, wherein the obtaining the motion profile of each AGV through the policy network comprises:
generating corresponding action masks according to the executable action list of the AGVs, and multiplying the output of each AGV strategy network by the corresponding action masks to obtain the value function estimation of legal actions;
and then activating a function through Geng Beier to obtain differentiable probability distribution, introducing a variable parameter t, and dynamically adjusting the probability distribution of different legal action value functions to obtain the action distribution of each AGV.
7. The method of multiple AGV scheduling policy evaluation and optimization based on approximate synchronization estimation according to claim 1, wherein the performing the individual optimization on the policy of each AGV comprises:
and using an edge dominance function as an update baseline, combining a near-end strategy optimization algorithm, adding an update gradient cut-off quantity as an update constraint, cutting and updating an AGV strategy network through a strategy ratio, controlling the update amplitude of the AGVs through a cutting interval, and performing independent optimization on the strategy of each AGV.
8. A multi-AGV scheduling policy evaluation and optimization system based on approximate synchronization estimation, comprising:
the construction module is used for constructing a strategy model shared by a plurality of AGVs and a combined strategy evaluation model;
the interaction module is used for enabling the AGVs to interact with the environment to generate training samples based on the strategy model shared by the AGVs and the combined strategy evaluation model, and calculating edge dominance functions;
the uncertainty module is used for calculating an update amplitude cut-off according to the uncertainty of the AGV strategy;
and the optimization module is used for replacing the joint advantage function according to the edge advantage function under any environment state, combining a near-end strategy optimization algorithm, taking the gradient cut-off quantity as a strategy ratio cutting interval, and performing independent optimization on the strategy of each AGV until the multi-AGV scheduling strategy is optimized.
9. An electronic device, characterized in that: comprising a memory, a processor and a computer program stored in the memory and executable on the processor, the processor implementing the multi-AGV scheduling policy evaluation and optimization method based on approximate synchronization estimation of any of claims 1-7 when executing the computer program.
10. A computer-readable storage medium, characterized by: the computer readable storage medium stores a computer program which when executed by a processor implements the multi-AGV scheduling policy evaluation and optimization method based on near-simultaneous estimation of any one of claims 1-7.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311099265.7A CN117032134A (en) | 2023-08-29 | 2023-08-29 | Multi-AGV scheduling strategy evaluation and optimization method and system based on approximate synchronization estimation |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311099265.7A CN117032134A (en) | 2023-08-29 | 2023-08-29 | Multi-AGV scheduling strategy evaluation and optimization method and system based on approximate synchronization estimation |
Publications (1)
Publication Number | Publication Date |
---|---|
CN117032134A true CN117032134A (en) | 2023-11-10 |
Family
ID=88622749
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202311099265.7A Pending CN117032134A (en) | 2023-08-29 | 2023-08-29 | Multi-AGV scheduling strategy evaluation and optimization method and system based on approximate synchronization estimation |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117032134A (en) |
-
2023
- 2023-08-29 CN CN202311099265.7A patent/CN117032134A/en active Pending
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6828121B2 (en) | Training neural networks with prioritized empirical memory | |
CN110520868A (en) | Distributed Reinforcement Learning | |
CN108885717A (en) | Asynchronous deeply study | |
CN112270403B (en) | Method, device, equipment and storage medium for constructing deep learning network model | |
CN110114784A (en) | Recurrent neural network | |
US11318938B2 (en) | Speed planning method and apparatus for self-driving, device, medium and vehicle | |
US20210232986A1 (en) | Parking lot free parking space predicting method, apparatus, electronic device and storage medium | |
CN111914994B (en) | Generation method and device of multi-layer perceptron, electronic equipment and storage medium | |
CN111966361B (en) | Method, device, equipment and storage medium for determining model to be deployed | |
EP3933712A1 (en) | Optimizer learning method and apparatus, electronic device and readable storage medium | |
CN112559721B (en) | Method, device, equipment, medium and program product for adjusting man-machine dialogue system | |
CN112100302A (en) | Map information point display method, device, equipment and readable storage medium | |
KR20230008005A (en) | Method and apparatus for determining information | |
CN114924862A (en) | Task processing method, device and medium implemented by integer programming solver | |
CN111680749A (en) | Method and device for obtaining output result of dispenser | |
CN117151390A (en) | Greedy value representation-based multi-unmanned carrier scheduling method and greedy value representation-based multi-unmanned carrier scheduling system | |
CN117933350A (en) | Multi-agent reinforcement learning system, method, electronic device and storage medium | |
CN112529162B (en) | Neural network model updating method, device, equipment and storage medium | |
CN111461306B (en) | Feature evaluation method and device | |
CN111311000B (en) | User consumption behavior prediction model training method, device, equipment and storage medium | |
CN117032134A (en) | Multi-AGV scheduling strategy evaluation and optimization method and system based on approximate synchronization estimation | |
CN111667428A (en) | Noise generation method and device based on automatic search | |
CN113159398B (en) | Power consumption prediction method and device and electronic equipment | |
CN111324747A (en) | Method and device for generating triples and electronic equipment | |
CN112308196A (en) | Method, apparatus, device and storage medium for training a model |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |