CN116205550A - Logistics garden vehicle goods taking scheduling method based on self-adaptive discrete particle swarm algorithm - Google Patents
Logistics garden vehicle goods taking scheduling method based on self-adaptive discrete particle swarm algorithm Download PDFInfo
- Publication number
- CN116205550A CN116205550A CN202310221846.7A CN202310221846A CN116205550A CN 116205550 A CN116205550 A CN 116205550A CN 202310221846 A CN202310221846 A CN 202310221846A CN 116205550 A CN116205550 A CN 116205550A
- Authority
- CN
- China
- Prior art keywords
- vehicle
- algorithm
- steps
- setting
- goods
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
- G06Q10/08—Logistics, e.g. warehousing, loading or distribution; Inventory or stock management
- G06Q10/083—Shipping
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/004—Artificial life, i.e. computing arrangements simulating life
- G06N3/006—Artificial life, i.e. computing arrangements simulating life based on simulated virtual individual or collective life forms, e.g. social simulations or particle swarm optimisation [PSO]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q10/00—Administration; Management
- G06Q10/06—Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
- G06Q10/063—Operations research, analysis or management
- G06Q10/0631—Resource planning, allocation, distributing or scheduling for enterprises or organisations
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Abstract
Logistics garden vehicle pickup scheduling method based on self-adaptive discrete particle swarm optimization. The method comprises the following specific steps: analyzing operation data related to the logistics park vehicles, and carrying out K-means clustering on target cargoes based on the park entering time and the park exiting time of each vehicle and the target cargoes corresponding to the vehicles; encoding the yard allocation section; encoding the pick-up cargo portion; setting global searching to determine an initial population; setting local search to determine an initial population; setting random search to determine an initial population; setting the maximum iteration number of the algorithm as iter; initializing a state set and an action set in reinforcement learning, selecting actions by using a greedy strategy, and calculating a report according to states and fitness values of the population chromosomes at the current moment. The invention has the beneficial effects that: the waiting time of the vehicle in the picking process can be obviously reduced, the logistics throughput is improved, and the maximum operation time of the storage yard is indirectly shortened so as to meet the work efficiency requirement of the logistics park storage yard.
Description
Technical Field
The invention relates to the technical field of vehicle dispatching, in particular to a logistics park vehicle goods taking dispatching method based on a self-adaptive discrete particle swarm algorithm.
Background
With the expansion of the logistics park scale, higher requirements are put on the efficiency of the logistics system. The efficient vehicle dispatching can obviously reduce the waiting time of the vehicle in the picking process and improve the logistics throughput.
At present, the vehicle dispatching mainly depends on manual experience, and has low efficiency on the mode that a plurality of vehicles run simultaneously in different yards, and traffic jam is easy to cause. How to utilize the relevant data to optimize the scheduling mechanism to meet the needs of vehicle and yard operations is a significant challenge.
Most of traditional vehicle round-entering queuing management adopts manual scheduling, and round-entering sequence is arranged according to a first-to-first-row criterion, so that problems of disordered scheduling, lower customer satisfaction and the like are caused due to lack of scientific optimization. The key parameters of the traditional meta-heuristic algorithm cannot be dynamically adjusted, so that the solving efficiency and the solving speed cannot reach the expected effect. The invention solves the scheduling problem of the vehicle storage yard based on the self-adaptive discrete particle swarm algorithm, and optimizes parameters by SARSA algorithm and Q learning in reinforcement learning.
Disclosure of Invention
The invention aims to provide a logistics park vehicle goods taking and scheduling method based on a self-adaptive discrete particle swarm algorithm, aiming at the defects and the shortcomings in the prior art, which can remarkably reduce the waiting time of vehicles in the goods taking process, improve logistics throughput and indirectly shorten the maximum operation time of a storage yard so as to meet the work efficiency requirement of the logistics park.
In order to achieve the above purpose, the invention adopts the following technical scheme: the logistics park vehicle goods taking and dispatching method based on the self-adaptive discrete particle swarm algorithm comprises the following specific steps: analyzing operation data related to the logistics park vehicles, and carrying out K-means clustering on target cargoes based on the park entering time and the park exiting time of each vehicle and the target cargoes corresponding to the vehicles; encoding the yard allocation section; encoding the pick-up cargo portion; setting global searching to determine an initial population; setting local search to determine an initial population; setting random search to determine an initial population; setting the maximum iteration number of the algorithm as iter; initializing a state set and an action set in reinforcement learning, selecting actions by using a greedy strategy, and calculating a report according to states and fitness values of population chromosomes at the current moment; the Sarsa algorithm updates the Q table, judges whether the current Sarsa update times are greater than the conversion times, if so, starts the QLearning algorithm to update the Q table, and selects actions by using a greedy strategy; judging the quality of the current population at every iteration moment, adaptively updating discrete particle swarm parameters omega by using the population state of the chromosome at the moment according to the trained Q table, and selecting omega which is most beneficial to the chromosome of the current population; set up [0,1]The random number is compared with omega, and the mode of goods taking variation and yard selection variation is adopted; selecting a POX chromosome crossing mode, and crossing part of chromosomes of a storage yard and the chromosomes of goods to be picked up respectively; selecting a two-point crossing mode, and respectively crossing part of chromosomes of the storage yard and the chromosomes of the goods to be picked; chromosome updating operation is carried out, the fitness value of each particle in the current population is evaluated and calculated, and the optimal P of the individual position is updated by comparison i And a global optimum position P g The method comprises the steps of carrying out a first treatment on the surface of the Judging whether the iteration times meet the conditionIf yes, stopping iteration and outputting a global optimal position P g And taking the result as an optimal scheduling result.
As a further illustration of the invention: the operation data related to the analyte logistics park vehicle are specifically as follows: according to a storage yard mixing rule, analyzing the residence time of the vehicle history in the storage yards, and giving the picking time length distribution T of the vehicles in each storage yard i 。
As a further illustration of the invention: the encoding of the yard allocation portion specifically includes: setting the chromosome length of the yard allocation part as the sum of the cargo types of all the vehicle cargo taking cargoes, wherein each genetic bit is expressed by an integer, the genetic bit is sequentially arranged according to the vehicles and the cargo taking cargoes of the vehicles, and each integer represents the yard number which is allocated by the current cargo taking cargoes of the vehicles and meets the requirement.
As a further illustration of the invention: the coding of the goods picking part is specifically as follows: the chromosome length of the goods taking part is set to be the sum of the goods types of all the goods taking of the vehicle, each gene is directly encoded by the vehicle number, and the sequence of the vehicle numbers indicates the sequence of the goods taking of the vehicle.
As a further illustration of the invention: the setting global search determines an initial population specifically as follows: setting an array with the length equal to the number of the storage yards, wherein the value on each array corresponds to the working time on the corresponding storage yard, randomly selecting a vehicle i from a vehicle picking set, starting from the first goods picking of the current vehicle i, adding the time corresponding to the storage yard of the current meeting the target goods picking to the time corresponding to the array, selecting the shortest time from the time to be used as the current target storage yard, updating the array, and so on until all the vehicles pick.
As a further illustration of the invention: the setting local search determines an initial population specifically as follows: setting an array with the length equal to the number of the storage yards, wherein the value on each array corresponds to the working time on the corresponding storage yard, randomly selecting a vehicle i from a vehicle picking set, starting from the first goods picking of the current vehicle i, adding the time corresponding to the storage yard of the current meeting the target goods picking, selecting the shortest time from the time to be used as the current target storage yard, resetting the array to 0 when the selection of the vehicle i is finished, and selecting the next vehicle until the goods picking of all vehicles is finished.
As a further illustration of the invention: the initial population is determined by setting random search specifically as follows: randomly selecting a vehicle i from a vehicle collection, randomly sequencing the target cargo taking sequence of the vehicle i, randomly selecting a storage yard from a storage yard collection meeting the cargo requirement for the sequenced target cargo, and the like until all vehicles are completely taken.
As a further illustration of the invention: the specific steps of initializing a state set and an action set in reinforcement learning, selecting actions by using a greedy strategy, and calculating a report according to state and fitness values of population chromosomes at the current moment are as follows: initializing a state set in reinforcement learning, setting the value range of the state set to be 0.5 and 1, and dividing the value range into 10 sections as state1 and state2 … … state10 on average; ) Initializing an action set in reinforcement learning, setting the value range of a state set to be [0,1], and dividing the value range into 10 intervals on average, wherein the 10 intervals are action1, action2 … … action10; and selecting actions according to states by using a greedy strategy, and calculating the reward according to states and fitness of the population chromosomes at the current moment.
As a further illustration of the invention: the Sarsa algorithm updates the Q table, judges whether the current Sarsa update times are larger than the conversion times, if so, starts the Qlearning algorithm to update the Q table, and selects the specific steps of actions by using a greedy strategy: setting the algorithm conversion times t=1, wherein the algorithm conversion times are t=25; the Sarsa algorithm is based on Q (s t ,a t )=(1-a)Q(s t ,a t )+a(r t+1 +γQ(s t+1 ,a t+1 ) Updating the Q table; judging whether the current t is larger than the conversion times, if so, starting chromosome updating operation, and if not, starting to select a POX chromosome crossing mode; QLearning algorithm is based on Q (s t ,a t )=(1-a)Q(s t ,a t )+a(r t+1 +γmax a Q(s t+1 ,a t+1 ) For a pair ofThe Q table is updated and actions are selected using a greedy strategy.
As a further illustration of the invention: the random numbers between the [0,1] are set, and compared with omega, the mode of adopting goods taking variation and yard selection variation by utilizing the size of the random numbers is specifically as follows: the method comprises the steps of respectively carrying out mutation on a chromosome part of goods to be picked, carrying out mutation on a storage yard selection part of a vehicle, and selecting one of the two types of mutation as a chromosome mutation operation by using a roulette-to-gambling mode.
After the technical scheme is adopted, the invention has the beneficial effects that: the logistics garden vehicle goods taking and dispatching method has the following advantages:
1. omega in the discrete particle swarm algorithm is intelligently updated by using the reinforcement learning algorithm, so that the solving efficiency and the solving speed are improved;
2. the chromosome state set is divided, the average fitness of the population and the diversity of the population are used as states to reflect the state of the whole population, the quality of the whole population is improved, and excellent individuals are easier to obtain;
3. the Sarsa algorithm is combined with the QLearning algorithm. The Sarsa algorithm is an on-pole algorithm, and the result obtained by the algorithm is more conservative than QLearing, the convergence speed is faster, but the learning effect is inferior to QLearing. The Qlearning algorithm is an off-policy algorithm, and the solving result of the algorithm has larger potential than that of Sarsa, has better global searching capability, and has slower convergence speed. The two algorithms are combined, the learning effect and the convergence rate of the Sarsa algorithm are inherited at the initial stage of the algorithm, and the better optimization capability of Qlearning is inherited at the later stage of the algorithm
4. The vehicle scheduling algorithm provided by the method can obviously reduce the waiting time of the vehicle in the picking process, improve the logistics throughput, indirectly shorten the maximum operation time of the storage yard and meet the work efficiency requirement of the logistics park storage yard.
Drawings
In order to more clearly illustrate the embodiments of the invention or the technical solutions of the prior art, the drawings which are used in the description of the embodiments or the prior art will be briefly described, it being obvious that the drawings in the description below are only some embodiments of the invention, and that other drawings can be obtained according to these drawings without inventive faculty for a person skilled in the art.
Fig. 1 is an algorithmic schematic of the present invention.
Fig. 2 is a schematic flow chart of the algorithm of the present invention.
Fig. 3 is a graph comparing the results of algorithm iterations of vehicle 10 in the present invention.
Fig. 4 is a graph comparing the results of algorithm iterations of vehicle 40 in the present invention.
Fig. 5 is a view showing the scheduling result of the vehicle 10 according to the present invention.
Fig. 6 is a view showing the scheduling result of the vehicle 40 in the present invention.
Detailed Description
Referring to fig. 1 to 6, the technical scheme adopted in this embodiment is as follows: the method comprises the following specific steps:
s1, analyzing operation data related to vehicles in a logistics park, carrying out K-means clustering on target cargoes based on the time of entering and exiting the park of each vehicle and the target cargoes corresponding to the vehicles, analyzing the stay time of the vehicle history in the storage yard according to the storage yard mixing rule, and giving out the cargoes taking time distribution T of the vehicles in each storage yard i 。
S2, coding a storage yard allocation part, setting the chromosome length of the storage yard allocation part as the sum of the cargo types of all the vehicle cargo taking cargoes, wherein each gene position is expressed by an integer, and sequentially arranging the cargo types according to the vehicles and the vehicles, and each integer represents the storage yard number which is allocated by the current cargo taking cargoes and meets the requirement.
S3, coding the goods taking part, setting the chromosome length of the goods taking part as the sum of the goods types of all the goods taking of the vehicle, directly coding each gene by using the vehicle number, and indicating the sequence of the goods taking of the vehicle by the sequence of the vehicle number.
S4, setting global search to determine an initial population, setting an array with the length equal to the number of storage yards, wherein the value on each array corresponds to the working time on the corresponding storage yard, randomly selecting a vehicle i from a vehicle collection, starting from the first goods taking of the current vehicle i, adding the corresponding time in the array to the goods taking time of the storage yard currently meeting the target goods taking, selecting the shortest time from the shortest time as the current target storage yard, updating the array, and so on until all vehicles are completely taken.
S5, setting local search to determine an initial population, setting an array with the length equal to the number of storage yards, wherein the value on each array corresponds to the working time on the corresponding storage yard, randomly selecting a vehicle i from a vehicle collection, starting from the first goods taking of the current vehicle i, adding the corresponding time in the array to the goods taking time of the storage yard currently meeting the target goods taking, selecting the shortest time from the time as the current target storage yard, resetting the array to 0 when the selection of the vehicle i is finished, and selecting the next vehicle until the goods taking of all vehicles is finished.
S6, setting random search to determine an initial population, randomly selecting a vehicle i from a vehicle picking set, randomly sequencing the target goods picking sequence of the vehicle i, randomly selecting a storage yard in a storage yard set meeting the goods requirement for the sequenced target goods, and the like until all vehicles are picked.
And S7, setting the maximum iteration number of the algorithm as iter, wherein the iter=50.
S8, initializing a state set and an action set in reinforcement learning, selecting actions by using a greedy strategy, and calculating reward according to states and fitness values of population chromosomes at the current moment:
1) Initializing a state set in reinforcement learning, setting the value range of the state set to be 0.5 and 1, and dividing the value range into 10 sections as state1 and state2 … … state10 on average;
2) Initializing an action set in reinforcement learning, setting the value range of a state set to be [0,1], and dividing the value range into 10 intervals on average, wherein the 10 intervals are action1, action2 … … action10;
3) And selecting actions according to states by using a greedy strategy, and calculating the reward according to states and fitness of the population chromosomes at the current moment.
S9, updating the Q table by the Sarsa algorithm, judging whether the current Sarsa updating times are larger than the conversion times, if so, starting the Qlearning algorithm to update the Q table, and selecting actions by using a greedy strategy:
1) Setting the algorithm conversion times t=1, wherein the algorithm conversion times are t=25;
2) The Sarsa algorithm is based on Q (s t ,a t )=(1-a)Q(s t ,a t )+a(r t+1 +γQ(s t+1 ,a t+1 ) Updating Q table, wherein Q(s) t ,a t ) Representing state s t By action a t And a represents the soft update weight, r t+1 Rewards representing time steps t+1, gamma represents discount rate, t=t+1
3) Judging whether the current t is larger than the conversion times, if so, starting S14, and if not, starting S12;
4) The Qkearning algorithm is based on Q (s t ,a t )=(1-a)Q(s t ,a t )+a(r t+1 +γmax a Q(s t+1 ,a t+1 ) Updating a Q table and selecting actions using a greedy strategy, where Q(s) t ,a t ) Representing state s t By action a t And a represents the soft update weight, r t+1 Representing the rewards of time steps t+1, gamma represents the discount rate, t=t+1.
And S10, judging the quality of the current population at every iteration moment, adaptively updating the discrete particle swarm parameters omega by using the current population state of the chromosome according to the trained Q table, and selecting the omega which is most beneficial to the chromosome of the current population.
S11, setting random numbers between [0,1], and using the size of the random numbers to compare with omega, adopting a mode of taking goods variation and yard selection variation to respectively perform variation on a chromosome part of the taking goods and variation on a yard selection part of a vehicle, and using a mode of roulette on a roulette to select one of the two variations as a chromosome variation operation.
S12, selecting a POX chromosome crossing mode, and crossing partial chromosomes of the storage yard and the goods taking chromosomes respectively.
S13, selecting a two-point crossing mode, and respectively crossing partial chromosomes of the storage yard and the goods taking chromosomes.
S14, performing chromosome updating operation, evaluating and calculating fitness value of each particle in the current population, and updating the optimal P of the individual position by comparing i And a global optimum position P g 。
S15, judging whether the iteration times meet the conditions, if so, stopping iteration, and outputting a global optimal position P g And taking the result as an optimal scheduling result.
The working principle of the invention is as follows: based on the combination of a Sarsa algorithm and a Qlearning algorithm in reinforcement learning, the key parameter omega of a discrete particle swarm algorithm is adaptively selected, firstly, the time of entering and exiting a round of each vehicle and the target goods for taking goods corresponding to the vehicle are analyzed, K-means clustering is carried out on the target goods, the stay time of the vehicle history in a storage yard is analyzed according to a storage yard mixing rule, the goods taking time distribution of the vehicle in each storage yard is given out, secondly, an initial population with better quality is generated by combining global searching, local searching and random searching, finally omega and omega in the discrete particle swarm algorithm are intelligently updated through reinforcement learning, the diversity of particle searching is determined, and when omega is large, the global searching capability is strong and the local searching capability is weak; and when the omega value is smaller, the local searching capability is strong, the global searching capability is weak, and the self-adaptive parameter omega is selected by using the reinforcement learning intelligent algorithm, so that the scheduling problem of the vehicle yard is solved in an efficient manner.
The foregoing is merely illustrative of the present invention and not restrictive, and other modifications and equivalents thereof may occur to those skilled in the art without departing from the spirit and scope of the present invention.
Claims (10)
1. The logistics park vehicle goods taking and dispatching method based on the self-adaptive discrete particle swarm algorithm is characterized by comprising the following steps of: the method comprises the following specific steps:
s1, analyzing operation data related to logistics garden vehicles, and carrying out K-means clustering on target cargoes based on garden entering time and garden exiting time of each vehicle and target cargo taking corresponding to the vehicles;
s2, coding a storage yard distribution part;
s3, encoding the goods taking part;
s4, setting global searching to determine an initial population;
s5, setting local search to determine an initial population;
s6, setting random search to determine an initial population;
s7, setting the maximum iteration number of the algorithm as item;
s8, initializing a state set and an action set in reinforcement learning, selecting actions by using a greedy strategy, and calculating a reward according to states and fitness values of the population chromosomes at the current moment;
s9, updating the Q table by the Sarsa algorithm, judging whether the current Sarsa updating times are larger than the conversion times, if so, starting the Qlearning algorithm to update the Q table, and selecting actions by using a greedy strategy;
s10, judging the quality of the current population at every iteration moment, adaptively updating the discrete particle swarm parameters omega by using the current population state of the chromosome according to the trained Q table, and selecting omega which is most beneficial to the chromosome of the current population;
s11, setting random numbers between [0,1], and adopting a mode of goods taking variation and yard selection variation by utilizing the size of the random numbers compared with omega;
s12, selecting a POX chromosome crossing mode, and respectively crossing part of chromosomes of a storage yard and the chromosomes of goods to be picked;
s13, selecting a two-point crossing mode, and respectively crossing part of chromosomes of the storage yard and the chromosomes of the goods to be picked;
s14, performing chromosome updating operation, evaluating and calculating fitness value of each particle in the current population, and updating the optimal P of the individual position by comparing i And a global optimum position P g ;
S15, judging whether the iteration times meet the conditions, if so, stopping iteration and outputting the wholeLocal optimum position P g And taking the result as an optimal scheduling result.
2. The logistics park vehicle pickup scheduling method based on the adaptive discrete particle swarm algorithm of claim 1, wherein the method comprises the following steps: the S1 specifically comprises the following steps: according to a storage yard mixing rule, analyzing the residence time of the vehicle history in the storage yards, and giving the picking time length distribution T of the vehicles in each storage yard i 。
3. The logistics park vehicle pickup scheduling method based on the adaptive discrete particle swarm algorithm of claim 1, wherein the method comprises the following steps: the step S2 is specifically as follows: setting the chromosome length of the yard allocation part as the sum of the cargo types of all the vehicle cargo taking cargoes, wherein each genetic bit is expressed by an integer, the genetic bit is sequentially arranged according to the vehicles and the cargo taking cargoes of the vehicles, and each integer represents the yard number which is allocated by the current cargo taking cargoes of the vehicles and meets the requirement.
4. The logistics park vehicle pickup scheduling method based on the adaptive discrete particle swarm algorithm of claim 1, wherein the method comprises the following steps: the step S3 is specifically as follows: the chromosome length of the goods taking part is set to be the sum of the goods types of all the goods taking of the vehicle, each gene is directly encoded by the vehicle number, and the sequence of the vehicle numbers indicates the sequence of the goods taking of the vehicle.
5. The logistics park vehicle pickup scheduling method based on the adaptive discrete particle swarm algorithm of claim 1, wherein the method comprises the following steps: the step S4 specifically comprises the following steps: setting an array with the length equal to the number of the storage yards, wherein the value on each array corresponds to the working time on the corresponding storage yard, randomly selecting a vehicle i from a vehicle picking set, starting from the first goods picking of the current vehicle i, adding the time corresponding to the storage yard of the current meeting the target goods picking to the time corresponding to the array, selecting the shortest time from the time to be used as the current target storage yard, updating the array, and so on until all the vehicles pick.
6. The logistics park vehicle pickup scheduling method based on the adaptive discrete particle swarm algorithm of claim 1, wherein the method comprises the following steps: the step S5 specifically comprises the following steps: setting an array with the length equal to the number of the storage yards, wherein the value on each array corresponds to the working time on the corresponding storage yard, randomly selecting a vehicle i from a vehicle picking set, starting from the first goods picking of the current vehicle i, adding the time corresponding to the storage yard of the current meeting the target goods picking, selecting the shortest time from the time to be used as the current target storage yard, resetting the array to 0 when the selection of the vehicle i is finished, and selecting the next vehicle until the goods picking of all vehicles is finished.
7. The logistics park vehicle pickup scheduling method based on the adaptive discrete particle swarm algorithm of claim 1, wherein the method comprises the following steps: the step S6 specifically comprises the following steps: randomly selecting a vehicle i from a vehicle collection, randomly sequencing the target cargo taking sequence of the vehicle i, randomly selecting a storage yard from a storage yard collection meeting the cargo requirement for the sequenced target cargo, and the like until all vehicles are completely taken.
8. The logistics park vehicle pickup scheduling method based on the adaptive discrete particle swarm algorithm of claim 1, wherein the method comprises the following steps: the specific steps of the S8 are as follows:
1) Initializing a state set in reinforcement learning, setting the value range of the state set to be 0.5 and 1, and dividing the value range into 10 sections as state1 and state2 … … state10 on average;
2) Initializing an action set in reinforcement learning, setting the value range of a state set to be [0,1], and dividing the value range into 10 intervals on average, wherein the 10 intervals are action1, action2 … … action10;
3) And selecting actions according to states by using a greedy strategy, and calculating the reward according to states and fitness of the population chromosomes at the current moment.
9. The logistics park vehicle pickup scheduling method based on the adaptive discrete particle swarm algorithm of claim 1, wherein the method comprises the following steps: the specific steps of the S9 are as follows:
1) Setting the algorithm conversion times t=1, wherein the algorithm conversion times are t=25;
2) The Sarsa algorithm is based on Q (s t ,a t )=(1-a)Q(s t ,a t )+a(r t+1 +γQ(s t+1 ,a t+1 ) Updating the Q table;
3) Judging whether the current t is larger than the conversion times, if so, starting S14, and if not, starting S12;
4) QLearning algorithm is based on Q (s t ,a t )=(1-a)Q(s t ,a t )+a(r t+1 +γmax a Q(s t+1 ,a t+1 ) Update the Q table and select actions using greedy policies.
10. The logistics park vehicle pickup scheduling method based on the adaptive discrete particle swarm algorithm of claim 1, wherein the method comprises the following steps: the step S11 specifically comprises the following steps: the method comprises the steps of respectively carrying out mutation on a chromosome part of goods to be picked, carrying out mutation on a storage yard selection part of a vehicle, and selecting one of the two types of mutation as a chromosome mutation operation by using a roulette-to-gambling mode.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310221846.7A CN116205550A (en) | 2023-03-09 | 2023-03-09 | Logistics garden vehicle goods taking scheduling method based on self-adaptive discrete particle swarm algorithm |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310221846.7A CN116205550A (en) | 2023-03-09 | 2023-03-09 | Logistics garden vehicle goods taking scheduling method based on self-adaptive discrete particle swarm algorithm |
Publications (1)
Publication Number | Publication Date |
---|---|
CN116205550A true CN116205550A (en) | 2023-06-02 |
Family
ID=86514568
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310221846.7A Pending CN116205550A (en) | 2023-03-09 | 2023-03-09 | Logistics garden vehicle goods taking scheduling method based on self-adaptive discrete particle swarm algorithm |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116205550A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117550273A (en) * | 2024-01-10 | 2024-02-13 | 成都电科星拓科技有限公司 | Multi-transfer robot cooperation method based on bee colony algorithm and transfer robot |
-
2023
- 2023-03-09 CN CN202310221846.7A patent/CN116205550A/en active Pending
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117550273A (en) * | 2024-01-10 | 2024-02-13 | 成都电科星拓科技有限公司 | Multi-transfer robot cooperation method based on bee colony algorithm and transfer robot |
CN117550273B (en) * | 2024-01-10 | 2024-04-05 | 成都电科星拓科技有限公司 | Multi-transfer robot cooperation method based on bee colony algorithm |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111860754B (en) | AGV scheduling method based on ant colony and genetic algorithm | |
CN107808215B (en) | Goods allocation optimization method applied to Flying-V type non-traditional layout warehouse | |
CN111553507B (en) | Multi-commodity-flow-based optimization method for Chinese and European container transportation scheme | |
CN111222665B (en) | Cloud manufacturing service combination optimization selection method based on preference NSGA-III algorithm | |
CN113222463B (en) | Data-driven neural network agent-assisted strip mine unmanned truck scheduling method | |
CN112053002B (en) | Cloud manufacturing multi-task scheduling method based on utility perception | |
CN112378415B (en) | Scheduling planning method, device and equipment for tools and appliances | |
CN116205550A (en) | Logistics garden vehicle goods taking scheduling method based on self-adaptive discrete particle swarm algorithm | |
CN106527381A (en) | Fast evaluation method facing parallel batch processing machine dynamic scheduling | |
CN113204417A (en) | Multi-satellite multi-point target observation task planning method based on improved genetic and firefly combined algorithm | |
CN113627712A (en) | Method for optimizing operation sequence of shuttle vehicle of storage system | |
CN115965169A (en) | Path planning method, intelligent device and computer readable storage medium | |
CN110098964A (en) | A kind of disposition optimization method based on ant group algorithm | |
CN110147885B (en) | Shared bicycle parking point distribution method for improving genetic algorithm | |
CN110956311B (en) | Vehicle path optimization method based on super heuristic algorithm of reinforcement learning | |
CN114626632A (en) | Vehicle scheduling method, system, equipment and medium based on ant lion algorithm | |
JP2555834B2 (en) | Group management elevator control method | |
CN115271130B (en) | Dynamic scheduling method and system for maintenance order of ship main power equipment | |
CN116033016A (en) | Collaborative caching method for parking vehicles by roadside in vehicle-mounted network | |
Yi et al. | Automated design of search algorithms based on reinforcement learning | |
CN112700190B (en) | Improved method for distributing tray materials by scanning method and genetic simulation annealing method | |
CN114707707A (en) | Method and system for scheduling AGV task based on improved genetic algorithm | |
CN115907066A (en) | Cement enterprise vehicle scheduling method based on hybrid sparrow intelligent optimization algorithm | |
CN115358455A (en) | Electric vehicle path optimization method considering battery replacement and hybrid time window constraints | |
CN111639822B (en) | Express distribution method based on 0-1 knapsack problem analysis |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |