CN117891259B - Multi-agent formation control method with multi-graph configuration and related product - Google Patents

Multi-agent formation control method with multi-graph configuration and related product Download PDF

Info

Publication number
CN117891259B
CN117891259B CN202410291076.8A CN202410291076A CN117891259B CN 117891259 B CN117891259 B CN 117891259B CN 202410291076 A CN202410291076 A CN 202410291076A CN 117891259 B CN117891259 B CN 117891259B
Authority
CN
China
Prior art keywords
agent
intelligent
gray
formation
target area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202410291076.8A
Other languages
Chinese (zh)
Other versions
CN117891259A (en
Inventor
吕金虎
蔡奕辰
刘克新
孙贵宾
李容江
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beihang University
Academy of Mathematics and Systems Science of CAS
Original Assignee
Beihang University
Academy of Mathematics and Systems Science of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beihang University, Academy of Mathematics and Systems Science of CAS filed Critical Beihang University
Priority to CN202410291076.8A priority Critical patent/CN117891259B/en
Publication of CN117891259A publication Critical patent/CN117891259A/en
Application granted granted Critical
Publication of CN117891259B publication Critical patent/CN117891259B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)
  • Image Generation (AREA)

Abstract

The invention discloses a multi-agent formation control method with a multi-graph configuration and related products. The method comprises the following steps: determining a desired formation map, a plurality of gray maps corresponding to the desired formation map and tasks allocated to the intelligent agent; according to the understanding values of the intelligent agents and the neighbor intelligent agents on the positions of the expected formation diagrams in the physical space, the understanding values of the intelligent agents on the positions of the expected formation diagrams in the physical space are updated, so that the understanding values of all the intelligent agents on the positions of the expected formation diagrams in the physical space tend to be consistent; and controlling the motion state of the intelligent body according to the graph entering speed command, the graph exploring speed command and the interaction speed command. The multi-agent formation efficiently and quickly realizes the formation task, and the coverage area is a plurality of communication graphs separated from each other.

Description

Multi-agent formation control method with multi-graph configuration and related product
Technical Field
The invention relates to a multi-agent formation control method with a multi-graph configuration and a related product.
Background
Multi-agent formation control refers to control techniques that work cooperatively between multiple autonomous agents to achieve a particular task or maintain a particular formation. The field covers various applications of robots, unmanned aerial vehicles, automatic driving vehicles, internet of things equipment and the like, and aims to enable a plurality of intelligent agents to cooperate and communicate with each other and realize collective behaviors. The research background for multi-agent formation control stems from the need for modern complex systems that typically require multiple agents to perform tasks in concert, e.g., unmanned aerial vehicle formation to perform search and investigation tasks, robotic team collaboration to perform article handling, or search and rescue. Multi-agent formation control plays an important role in both civilian and military scenarios. With the development of modern technology, the unmanned cluster formed by multiple intelligent agents has a larger and larger scale, and the scale and the number of tasks which can be executed by the unmanned cluster have gradually increased. In the face of multi-agent and multi-task scenes, the control method not only needs to complete realization of formation configuration of the agents, but also needs to reasonably distribute tasks of the agents. Therefore, how to design a method for simultaneously realizing the distribution and control of the intelligent agents and the formation tasks under the large-scale multi-intelligent agents and multi-formation tasks becomes a problem to be solved.
The existing multi-agent formation control method can be divided into two modes: centralized and distributed. In the centralized method, one main controller is responsible for controlling the whole intelligent agent cluster, but the method can be difficult to implement in many practical application scenes due to the problems of high communication overhead, high single-point fault risk and the like. On the other hand, the distributed method only requires each agent to obtain local information around the agent and can only communicate with neighbor agents within a certain range. The distributed method greatly reduces the requirement of the cluster on communication capacity, and meanwhile, only the utilization of local information increases the fault tolerance of the cluster, so that the distributed method becomes a mainstream method in practical application. But in the case of large-scale clusters, the distribution may face challenges such as information transfer delays, unequal task allocation, etc. Therefore, in a distributed scenario, it is necessary to improve the performance such as the optimality and convergence rate of the control method as much as possible.
Most of the existing distributed multi-agent formation control methods are used for realizing formation configuration aiming at a single graph, tasks are realized for configuration of a plurality of graphs, and the formation control method aiming at the single graph is difficult to directly implement. The multi-graph formation control problem can be abstracted into a multi-agent multi-task problem. The existing task allocation and track planning method aiming at the multi-agent multi-task problem often considers the task as a particle, and further takes factors such as the moving distance of the agent, the energy consumption and the like as optimization indexes and gives an optimal solution. However, such methods cannot meet the demands of formation tasks, and thus cannot be applied to the problem of multi-graphic formation control. In general, the problem of multi-pattern formation control is currently studied to a small extent, and the problem of multi-pattern formation control can be effectively solved by using fresh technology.
The method proposed by the Chinese patent application CN111259327A (optimization method for multi-agent system consistency problem based on subgraph processing) is as follows: 1) Constructing a graph signal model of the multi-agent system; 2) Adding auxiliary constraint conditions to the problem to be solved; 3) Local inversion in the subgraph; 4) Fusion and averaging among subgraphs; 5) And (5) iteratively eliminating errors.
Disclosure of Invention
The invention provides a multi-agent formation control method with a multi-graph configuration and related products.
The technical scheme of the invention is as follows: a multi-agent formation control method of a multi-graphic configuration, comprising:
Determining a desired formation map, a plurality of gray maps corresponding to the desired formation map and tasks allocated to the intelligent agent;
according to the understanding values of the intelligent agents and the neighbor intelligent agents on the positions of the expected formation diagrams in the physical space, the understanding values of the intelligent agents on the positions of the expected formation diagrams in the physical space are updated, so that the understanding values of all the intelligent agents on the positions of the expected formation diagrams in the physical space tend to be consistent;
Controlling the motion state of the intelligent agent according to the graph entering speed command, the graph exploring speed command and the interaction speed command;
the method comprises the steps that a gray maximum value or a gray minimum value in an expected formation graph represents a target area, the target area is connected into a plurality of sub-images separated from each other, the sub-images are communicated with each other, the plurality of gray images are in one-to-one correspondence with the plurality of sub-images, each gray image is divided into a target area, a transition area and a non-target area, the coordinates and the gray values of grids in the target area are the same as those of the grids in the expected formation graph, the transition area is close to the target area and smoothly transits along the direction far away from the target area, the non-target area is close to the boundary of the transition area far away from the target area and the gray value is the gray minimum value or the gray maximum value different from the gray value of the target area, and the shapes and the sizes of the grids in the sub-images are equal to those of the grids in the expected formation graph;
Wherein each task corresponds to a sub-graph;
The graphic entry speed instruction is determined according to the corresponding gray level graph and is used for driving the intelligent agent to move from outside the target area to the target area;
The graphic exploration speed instruction is determined according to the corresponding gray level diagram and is used for driving the intelligent agent to enter the target area from the outer side edge of the target area, and further searching and occupying grids which are not occupied by other intelligent agents in the target area;
Wherein the interaction speed command is used for avoiding collision and driving the speed of the intelligent agent to be consistent with the speed of surrounding intelligent agents.
The technical scheme of the invention is as follows: an agent comprising a memory and a processor, the memory storing a program, the processor running the program to perform the multi-agent formation control method of the aforementioned multi-graphic configuration.
The technical scheme of the invention is as follows: a program product that, at run-time, performs the multi-agent formation control method of the aforementioned multi-graph configuration.
The beneficial effects of the invention include: in the proposed multi-agent formation control method, two parts of task allocation and agent control are innovatively combined, and the multi-graph configuration realization problem in the multi-agent formation problem is reasonably and efficiently solved; the distributed and asynchronous calculation mode is adopted, so that the requirement on the communication capacity of the intelligent agent is effectively reduced, and the engineering feasibility is increased; the convergence of formation targets can be guaranteed to be realized within a limited time, and the local calculation complexity of each intelligent agent is within an acceptable range, so that engineering realization is easy.
Drawings
FIG. 1 is a flow chart of a multi-agent formation control method of the multi-graph configuration of the present invention.
Fig. 2 is a block diagram of a controller in an agent of the present invention.
Detailed Description
The present invention will be further described with reference to specific examples, but the scope of the present invention is not limited thereto.
Referring to fig. 1, the flow of the multi-agent formation control method of the multi-graphic configuration proposed by the present invention includes the following steps.
Step 101, determining a desired formation map, a plurality of gray maps corresponding to the desired formation map and tasks allocated to the intelligent agent;
102, updating the understanding values of the intelligent agents on the positions of the expected formation diagrams in the physical space according to the understanding values of the intelligent agents and the neighbor intelligent agents on the positions of the expected formation diagrams in the physical space, so that the understanding values of all the intelligent agents on the positions of the expected formation diagrams in the physical space tend to be consistent;
Step 103, controlling the motion state of the intelligent agent according to the graph entering speed command, the graph exploring speed command and the interaction speed command;
The method comprises the steps that a gray maximum value or a gray minimum value in an expected formation graph represents a target area, the target area is connected into a plurality of sub-images separated from each other, the sub-images are communicated with each other, the plurality of gray images are in one-to-one correspondence with the plurality of sub-images, each gray image is divided into a target area, a transition area and a non-target area, the coordinates and the gray values of grids in the target area are the same as those of grids in the expected formation graph, the transition area is close to the target area and smoothly transits along the direction far away from the target area, the non-target area is close to the boundary of the transition area far away from the target area, the gray value is the gray minimum value or the gray maximum value different from the gray value of the target area, and the shapes and the sizes of the grids in the sub-images are equal after the grids in the sub-images are mapped to a physical space;
Wherein each task corresponds to a sub-graph;
The graphic entry speed instruction is determined according to the corresponding gray level graph and is used for driving the intelligent agent to move from outside the target area to the target area;
The graphic exploration speed instruction is determined according to the corresponding gray level diagram and is used for driving the intelligent agent to enter the target area from the outer side edge of the target area, and further searching and occupying grids which are not occupied by other intelligent agents in the target area;
Wherein the interaction speed command is used for avoiding collision and driving the speed of the intelligent agent to be consistent with the speed of surrounding intelligent agents.
In one example, the sum of the above three commands is equal to the expected velocity vector of the agent, which the agent is controlled by the agent's power system to move in accordance with.
The execution subject of the above method is a single agent in a multi-agent formation, which may be a controller in the agent from a hardware perspective, and may be a program running on the agent from a software perspective.
The desired formation map and the plurality of gray maps corresponding to the desired formation map are input to the agent by an external device.
In the expected formation map, the target region is represented by a region formed by the following meshes having a gradation value of 0, and the gradation of the meshes of the remaining regions in the expected formation map is 1.
Illustratively, the shape of the target area of the formation is expected to be the shape of "ABC" three letters, which are separated from each other. The target area occupied by letter a constitutes a sub-image (named sub-image a), the target area occupied by letter B constitutes a sub-image (named sub-image B), and the target area occupied by letter C constitutes a sub-image (named sub-image C).
Illustratively, the resolution of each gray map is the same as the resolution of the desired formation map. In the gray scale map corresponding to the sub-map a, the grids having the gray scale value of 0 are connected in the shape of the letter a, and the positions and the sizes are the same as those of the letter a in the desired formation map. In the gray map corresponding to sub-map B, the grid with gray value 0 is connected in the shape of letter B, and the position and size are the same as those of letter a in the desired formation map. In the gray scale map corresponding to the sub-graph C, the grids having the gray scale value of 0 are connected in the shape of the letter C, and the position and size are the same as those of the letter C in the desired formation map.
Illustratively, in each gray scale map, the gray scale value of each grid smoothly transitions from 0 to 1 in a direction away from the target region. In a region far enough from the target region, the gray value of the grid is 1.
The grid of the gray scale map is a data point on the data storage, and the shape when mapped to the physical space is, for example, square, regular hexagon, rectangle, or the like. In the following example, the grid of the gray map is being square in shape in physical space.
The present invention is not limited to how the gray scale map is generated.
Illustratively, the task of the agent coverage area forming the letter a pattern is referred to as task a, the task of the agent coverage area forming the letter B pattern is referred to as task B, and the task of the agent coverage area forming the letter C pattern is referred to as task C. Which task the agent specifically performs may be preset or specified by an external device inputting instructions to the agent.
It should be noted that, step 102 and step 103 are performed in real time.
Optionally, the intelligibility value of the position of the object shape by the agent itself is calculated according to the following formula:
Wherein, For the understanding value of agent i on the speed of the expected formation graph in physical space,/>For the understanding value of agent j on the speed of the expected formation graph in physical space,/>/>Both of which are constant gains,Is a collection of neighbor agents (agents which are located within a communication radius and can communicate bi-directionally, not limited to whether to perform the same task) of agent i,/>As a sign function,/>Calculating the sign for 2-norms,/>For the understanding value of agent i to the position of the desired formation map in physical space,/>An understanding value of the position of the expected formation map in the physical space for the agent j;
wherein c 2 is a positive constant gain, For the understanding value of the intelligent agent i to the expected formation graph in the direction of the physical space,/>The understanding value of the expected formation graph in the direction of the physical space for the agent j.
The first term in the first formula is used for enabling the understanding value of the object shape position of the object i to be consistent with the understanding value of the object shape position of the object j, and the second term is used for enabling the understanding value of the object shape speed of the object i to be consistent with the understanding value of the object shape speed of the object j.
The first term in the second formula is used for enabling the understanding value of the object shape direction of the object i to be consistent with the understanding value of the object shape direction of the object j, and the second term is used for enabling the understanding value of the object shape angular velocity of the object i to be consistent with the understanding value of the object shape angular velocity of the object j.
At the initial time, the agent understands its own position as the position of the target shape.
At the initial time, the understanding value of the object shape direction by the agent can be random or preset.
In the present invention, communication between agents is bi-directional.
A value of less than 1 allows consistency negotiations about the orientation of the target shape to be achieved in a limited time, which is of great importance for speeding up the negotiation process.
In a further embodiment of the present invention,The value is 1.
The target shape refers to the formation shape of all agents after being mapped to the physical space by the desired formation map.
Optionally, the calculation formula of the graph entry speed command is:
Wherein, For the pattern entry velocity command of agent i, k 1 is a positive control gain constant,/>For the grid nearest to the center of agent i,/>For grid/>Gray value in gray map corresponding to agent i,/>To grid/>, in gray scale map corresponding to agent iA grid having gray values closest to those of the target region in the set region (for example, in the matrix region of 3*3) for the center is located in the physical space,/>Is the location of agent i in physical space.
For example, the graphics entry speed command causes the agent to move toward the target area in the direction in which the gray value decreases.
Optionally, the calculation formula of the graph exploration speed instruction is:
Wherein, Exploring speed instructions for graphics of agent i,/>For the location of agent i in physical space,/>Is a grid/>In the position of physical space,/>For the perceived radius of agent i,/>For the collection of grids representing the target area in the gray-scale map corresponding to agent i,/>Representing the grid occupied by any agent in the gray map corresponding to agent i,/>, for exampleRepresenting a set of grids representing a target region within an agent i perception radius in a gray scale map corresponding to agent i,/>Control gain constant for positive value,/>Is a weight function,/>A non-negative function, satisfying:,/> is a piecewise continuous function and satisfies/>
For example, a weight functionThe calculation formula of (2) is
When the intelligent agent is at the edge of the corresponding target area and does not enter the target area, the graphic exploration speed instruction pulls the intelligent agent into the target area; and after the intelligent agent enters the target area, the graph exploration speed instruction can push the intelligent agent to explore the graph area which is not occupied by other intelligent agents, so that the multi-intelligent agent formation uniformly covers the target graph.
Optionally, the calculation formula of the interaction speed instruction is:
Wherein, Instructions for interaction speed for agent i,/>Control gain constant of positive value,/>For the set of neighbor agents of agent i, o i is the set of obstacle points within the perceived radius of agent i,/>For the location of agent i in physical space,/>For the position of the agent or obstacle j in physical space,/>For the speed of agent i,/>For the velocity of the agent j,Is a weight function and monotonically decreases over a range of 0 to the expected distance of the agent.
The first term of the equation is a repulsive velocity that pushes agent i away from the surrounding environment to avoid collisions. The second term aims at keeping the speed of agent i consistent with the speed of the surrounding agents. The interaction speed instruction is helpful to reduce the probability of collision between agents and helps to realize the final stabilization of multi-agent formation.
When the distance between the agents is desired to be as small as possible, the desired distance may be a collision avoidance distance between the agents. The expected distance may also be greater than the collision avoidance distance.
Optionally, a weight functionThe calculation formula of (2) is as follows:
Optionally, the method further comprises the step of task allocation:
Each agent And performing strategy updating through the following iterative steps, thereby completing task allocation. Each agentThe information stored, and also the information that needs to be updated for each iteration, is a 4-metadata set: . Wherein/> Is an agent/>Currently, the task allocation policies considered, in particular,Is one from the collection of agents/>To task set/>In particular if in the allocation policy/>Will agent/>Assigning execution tasks/>(In particular, use/>Representing an empty task, i.e., not executing a task), then use/>A representation; Is a satisfaction criterion, representing agent/> Whether to allocate policy/>Is satisfied, if satisfiedExpressed in allocation policy/>Medium intelligent agent/>Unwilling to change its selected task, otherwiseRepresenting agent/>Dissatisfied with the current allocation; /(I)Representing agent/>Is the number of iterations; /(I)Representing a timestamp, which is a uniform random variable, whenever an agent/>Update policy/>Time randomly generated (i.e., random time stamps).
Specifically, each agentPolicy updating is performed by the iterative steps of:
(1) Initially, the agent Consider that all agents perform an empty task, i.e./>,Is an agent set; each agent/>Initial iteration number/>; Each agent/>Is randomly generated, i.e./>, by the initial timestamp s i Wherein/>Representing the generation of random numbers in the [0,1] interval; each agentInitial satisfaction criterion/>; And (2) rotating.
(2) For intelligent bodyIf the criterion/>Turning to (3); otherwise, turning to (4).
(3) Intelligent bodyBased on the allocation policy currently considered/>Select and make agent/>Task/>, which has the greatest current individual benefit per seI.e. task/>Satisfy/>=/>Wherein/>Representation of policies/>Task execution considered/>Set of all agents,/>Representing collections/>The number of elements in the list; agent/>Updating the allocation policy considered by itself/>So that/>; Agent/>Iteration number/>; Agent/>Satisfaction criterion/>; And (4) rotating.
(4) Intelligent bodyWith neighbor agent/>Interactions are performed in which/>Representing agent/>Neighbor set of (i.e. at/>)And (3) other intelligent agent sets capable of two-way communication in the communication range. Specifically, agent/>To all neighbor agents/>Broadcasting its current information, M i =/>And receives from the neighbor/>Current information of (2)Thus obtaining the information of allocation strategy, iteration times, time stamp, satisfaction state and the like considered by the neighbor agent; agent/>After receiving the information of all neighbor agents, a set is formed from all the informationIn selecting the iteration number/>If a plurality of agents with the same iteration times and the maximum number of iterations exist, selecting the agent with the maximum s k as the authoritative agent by comparing the time stamps, and recording the selected authoritative agent as/>; If the agent/>Is not itself an authoritative agent, i.e./>Then the complete information of authoritative agent is used for replacing the agent/>Is the whole information of/>Then update the satisfaction statusTurning to (2); if the agent/>By itself, exactly authoritative, i.e./>Agent/>Is unchanged, and the process goes to (5).
(5) If the agentSatisfactory state/>And all neighbors received by it/>Satisfactory state information of (a)Are all 1, then agent/>Terminating the iteration and outputting the allocation strategy/>Agent/>According to the currently considered allocation strategy/>Executing; otherwise, go to (2).
The method can realize stability after finite step iteration, namely, the allocation strategy considered by each agent can be agreed, and each agent is satisfied with the current state under the allocation strategy. Thus, the task allocation policy can be obtained by the above algorithm.
The timestamp initial value is not or can be set by default, so long as all agents are ensured to be inconsistent with each other. When the timestamp initial values are randomly generated, the timestamp initial values of the two agents are the same as each other and are small probability events which can be ignored.
Wherein the benefit of an allocation policy may be defined by the sum of the individual benefits of all agents under the allocation policy.
Is provided withIndividual agent,/>And (3) task allocation is carried out by adopting a reasonable algorithm rule. Recording deviceRepresenting a population of agents, wherein/>Representing agent i,/>Representing a population of tasks, wherein/>Representing task j,/>,/>Indicating an empty task, i.e., the agent is currently not performing a task.
Definition of agentIndividual benefit/>It is the task that agent i is assigned and the total agent number/>, that performs the taskIs a function of (2). /(I)Representing agent/>Individual benefit when executing task t j with p-1 peers. Further, for a formation task of a multi-graph configuration, defining individual benefits as:
Wherein, Is intelligent body,/>For/>Selected task,/>Representation and/>The number of peers (including themselves) that commonly perform task t j; /(I)Representing agent/>Location of/(I)Representing the position of task t j;
Representing the maximum distance between the agent performing task t j and task t j; /(I) As weight parameters, generally 0.1-0.5 is preferable; n tj represents the minimum number of agents required for task t j, calculated as/>Wherein/>Representing the area covered by task t j in physical space,/>Indicating removal of the whole.
The design of this individual utility function enables the population benefit of task t j to be exactly maximized when the number of agents for task t j is chosen to be exactly n tj. Therefore, the task allocation based on the benefit function can make the number of agents allocated to the task t j approach n tj, which also meets the practical requirement. Note that the design of the benefit function is not unique and can be adjusted according to actual scene requirements. For any taskAgent/>For any two positive integersAnd meet/>It is necessary to have/>
The step of task allocation may be performed before the agent performs the formation task.
Through the distributed asynchronous algorithm, each agent can obtain the same allocation strategy, and the allocation strategy is Nash stable, namely, if any agent changes own task execution selection, the benefit of the agent cannot become larger.
Based on the same inventive concept, the present invention also provides a program product that performs the aforementioned multi-agent formation control method of the multi-graphic configuration at runtime.
Based on the same inventive concept, referring to fig. 2, the present invention also provides an agent including a memory and a processor, the memory storing a program, the processor running the program to perform the aforementioned multi-agent formation control method of the multi-graphic configuration.
The embodiments of the present invention are described in a progressive manner, and the same and similar parts of the embodiments are all referred to each other, and each embodiment is mainly described in the differences from the other embodiments.
The scope of the present invention is not limited to the above-described embodiments, and it is apparent that various modifications and variations can be made to the present invention by those skilled in the art without departing from the scope and spirit of the invention. It is intended that the present invention also include such modifications and alterations insofar as they come within the scope of the appended claims or the equivalents thereof.

Claims (10)

1. A multi-agent formation control method of a multi-graphic configuration, comprising:
Determining a desired formation map, a plurality of gray maps corresponding to the desired formation map and tasks allocated to the intelligent agent;
according to the understanding values of the intelligent agents and the neighbor intelligent agents on the positions of the expected formation diagrams in the physical space, the understanding values of the intelligent agents on the positions of the expected formation diagrams in the physical space are updated, so that the understanding values of all the intelligent agents on the positions of the expected formation diagrams in the physical space tend to be consistent;
Controlling the motion state of the intelligent agent according to the graph entering speed command, the graph exploring speed command and the interaction speed command;
the method comprises the steps that a gray maximum value or a gray minimum value in an expected formation graph represents a target area, the target area is connected into a plurality of sub-images separated from each other, the sub-images are communicated with each other, the plurality of gray images are in one-to-one correspondence with the plurality of sub-images, each gray image is divided into a target area, a transition area and a non-target area, the coordinates and the gray values of grids in the target area are the same as those of the grids in the expected formation graph, the transition area is close to the target area and smoothly transits along the direction far away from the target area, the non-target area is close to the boundary of the transition area far away from the target area and the gray value is the gray minimum value or the gray maximum value different from the gray value of the target area, and the shapes and the sizes of the grids in the sub-images are equal to those of the grids in the expected formation graph;
Wherein each task corresponds to a sub-graph;
The graphic entry speed instruction is determined according to the corresponding gray level graph and is used for driving the intelligent agent to move from outside the target area to the target area;
The graphic exploration speed instruction is determined according to the corresponding gray level diagram and is used for driving the intelligent agent to enter the target area from the outer side edge of the target area, and further searching and occupying grids which are not occupied by other intelligent agents in the target area;
Wherein the interaction speed command is used for avoiding collision and driving the speed of the intelligent agent to be consistent with the speed of surrounding intelligent agents.
2. The method of claim 1, wherein the graphic entry speed command is calculated as:
Wherein, For the pattern entry velocity command of agent i, k 1 is a positive control gain constant,/>For the grid nearest to the center of agent i,/>For grid/>Gray value in gray map corresponding to agent i,/>To grid/>, in gray scale map corresponding to agent iThe position of a grid with the gray value closest to the gray value of the target area in the set area of the center in the physical space,/>Is the location of agent i in physical space.
3. The method of claim 1, wherein the graphic exploration rate instruction is calculated by the formula:
Wherein, Exploring speed instructions for graphics of agent i,/>For the location of agent i in physical space,/>Is a gridIn the position of physical space,/>For the perceived radius of agent i,/>For the collection of grids representing the target area in the gray-scale map corresponding to agent i,/>Representing the grid occupied by any agent in the gray map corresponding to agent i,/>, for exampleRepresenting a set of grids representing a target region within an agent i perception radius in a gray scale map corresponding to agent i,/>Control gain constant for positive value,/>Is a weight function,/>A non-negative function, satisfying:,/> is a piecewise continuous function and satisfies: /(I)
4. The method of claim 1, wherein the interaction speed command is calculated as:
Wherein, Instructions for interaction speed for agent i,/>Control gain constant of positive value,/>For the set of neighbor agents of agent i, o i is the set of obstacle points within the perceived radius of agent i,/>For the location of agent i in physical space,For the position of the agent or obstacle j in physical space,/>For the speed of agent i,/>For the velocity of agent j,/>Is a weight function and monotonically decreases over a range of 0 to the expected distance of the agent.
5. The method of claim 4, wherein the weight functionThe calculation formula of (2) is as follows:
Wherein, Is the expected distance.
6. The method of claim 1, wherein the intelligibility of the orientation of the target shape by the agent itself is calculated according to the following formula:
Wherein, For the understanding value of agent i on the speed of the expected formation graph in physical space,/>For the understanding value of agent j on the speed of the expected formation graph in physical space,/>/>Both are constant gains,/>Is the collection of neighbor agents of agent i,/>As a sign function,/>Calculating the sign for 2-norms,/>For the understanding value of agent i to the position of the desired formation map in physical space,/>An understanding value of the position of the expected formation map in the physical space for the agent j;
wherein c 2 is a positive constant gain, For the understanding value of the intelligent agent i to the expected formation graph in the direction of the physical space,/>The understanding value of the expected formation graph in the direction of the physical space for the agent j.
7. The method of claim 1, further comprising the step of task allocation:
(1) The initial state judges that all the agents execute the empty task, the initial iteration times are 0, the initial time stamp is different from the initial time stamp of the other agents, and the initial satisfaction degree is judged to be unsatisfied;
(2) If the satisfaction degree is judged to be unsatisfactory, the process goes to (3), otherwise, the process goes to (4);
(3) Based on the distribution strategy generated by the intelligent body, selecting a task which maximizes the individual benefit of the intelligent body, and updating the distribution strategy and the iteration times of the intelligent body, so that the satisfaction degree is judged to be satisfactory;
(4) The method comprises the steps of (1) judging interaction allocation strategies with neighbor intelligent agents, iteration times, time stamps and self satisfaction, selecting the intelligent agent with the largest iteration times from the self and the neighbor intelligent agents of the self as an authoritative intelligent agent, selecting the intelligent agent with the largest time stamp from the authoritative intelligent agent if the number of the intelligent agent with the largest iteration times is a plurality of intelligent agents, if the intelligent agent with the largest time stamp is the authoritative intelligent agent, judging the self allocation strategies, the iteration times, the time stamps and the self satisfaction without change, and turning to (5), otherwise, assigning the self allocation strategies, the iteration times, the time stamps and the self satisfaction by using the allocation strategies, the iteration times, the time stamps and the self satisfaction of the authoritative intelligent agent;
(5) If satisfaction degree judgment of the intelligent agent and all the neighbor intelligent agents is satisfied, the intelligent agent stops iteration, otherwise, the process goes to (2).
8. The method of claim 7, wherein the individual benefit calculation formula is as follows:
Representing agent/> Individual benefit when executing task t j with p-1 peers,/>Representing agent/>Location of/(I)Representing the position of task t j, d max represents the maximum distance between the agent performing task t j and task t j,/>As a weight parameter, n tj represents the minimum number of agents required for task t j.
9. A program product, characterized in that it performs, at run-time, the multi-agent formation control method of the multi-graphic configuration according to any one of claims 1 to 8.
10. An agent comprising a memory and a processor, the memory storing a program that is executed by the processor to perform the multi-agent formation control method of the multi-graphic configuration according to any one of claims 1 to 8.
CN202410291076.8A 2024-03-14 2024-03-14 Multi-agent formation control method with multi-graph configuration and related product Active CN117891259B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202410291076.8A CN117891259B (en) 2024-03-14 2024-03-14 Multi-agent formation control method with multi-graph configuration and related product

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202410291076.8A CN117891259B (en) 2024-03-14 2024-03-14 Multi-agent formation control method with multi-graph configuration and related product

Publications (2)

Publication Number Publication Date
CN117891259A CN117891259A (en) 2024-04-16
CN117891259B true CN117891259B (en) 2024-05-14

Family

ID=90649189

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202410291076.8A Active CN117891259B (en) 2024-03-14 2024-03-14 Multi-agent formation control method with multi-graph configuration and related product

Country Status (1)

Country Link
CN (1) CN117891259B (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006029297A2 (en) * 2004-09-10 2006-03-16 Hoftberg Steven Game theoretic prioritization scheme for mobile ad hoc networks permitting hierarchal deference
CN102270305A (en) * 2011-08-11 2011-12-07 西北工业大学 Multi-agent cooperative target identification method based on MSBN (Multiple Sectioned Bayesian Network)
CN112000108A (en) * 2020-09-08 2020-11-27 北京航空航天大学 Multi-agent cluster grouping time-varying formation tracking control method and system
CN114815856A (en) * 2022-06-27 2022-07-29 中国人民解放军国防科技大学 Time-varying formation inclusion control method, device and equipment for incomplete multi-agent system
CN115616913A (en) * 2022-07-05 2023-01-17 北京理工大学 Model prediction leaderless formation control method based on distributed evolutionary game
CN117155616A (en) * 2023-08-15 2023-12-01 重庆大学 Method for predicting deception attack intention in multi-agent system based on inverse reinforcement learning

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006029297A2 (en) * 2004-09-10 2006-03-16 Hoftberg Steven Game theoretic prioritization scheme for mobile ad hoc networks permitting hierarchal deference
CN102270305A (en) * 2011-08-11 2011-12-07 西北工业大学 Multi-agent cooperative target identification method based on MSBN (Multiple Sectioned Bayesian Network)
CN112000108A (en) * 2020-09-08 2020-11-27 北京航空航天大学 Multi-agent cluster grouping time-varying formation tracking control method and system
CN114815856A (en) * 2022-06-27 2022-07-29 中国人民解放军国防科技大学 Time-varying formation inclusion control method, device and equipment for incomplete multi-agent system
CN115616913A (en) * 2022-07-05 2023-01-17 北京理工大学 Model prediction leaderless formation control method based on distributed evolutionary game
CN117155616A (en) * 2023-08-15 2023-12-01 重庆大学 Method for predicting deception attack intention in multi-agent system based on inverse reinforcement learning

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
"多视觉传感器协同弱小目标检测";王田等;《指挥与控制学报》;20240215;全文 *
采用MSBN多智能体协同推理的智能农业车辆环境识别;郭文强;高晓光;侯勇严;周强;;智能系统学报;20130515(第05期);全文 *

Also Published As

Publication number Publication date
CN117891259A (en) 2024-04-16

Similar Documents

Publication Publication Date Title
CN107491086B (en) Unmanned aerial vehicle formation obstacle avoidance method and system under time-varying network topology
CN114281104B (en) Multi-unmanned aerial vehicle cooperative regulation and control method based on improved ant colony algorithm
CN111580545B (en) Heterogeneous unmanned aerial vehicle cluster distributed space-time coordination control method
CN112068586B (en) Space-time joint optimization four-rotor unmanned aerial vehicle trajectory planning method
CN111142567B (en) Unmanned aerial vehicle target position exchange method and device in unmanned aerial vehicle system
CN113050677A (en) Control method, system and storage medium for maintaining and changing formation of multiple unmanned aerial vehicles
CN116307464A (en) AGV task allocation method based on multi-agent deep reinforcement learning
CN113342034A (en) Combined strategy algorithm for unmanned aerial vehicle channel inspection and refined inspection
CN112230677A (en) Unmanned aerial vehicle group task planning method and terminal equipment
Pivtoraiko et al. Differentially constrained motion replanning using state lattices with graduated fidelity
CN117891259B (en) Multi-agent formation control method with multi-graph configuration and related product
EP4015997A1 (en) Map representation system and computer-readable medium
CN108182243B (en) Spark-based distributed evolutionary algorithm island model parallelization method
CN117610899A (en) Multi-robot task allocation method based on priority
CN116820129A (en) Unmanned aerial vehicle cluster formation method and system based on autonomous collaboration mechanism
Su et al. Collaborative motion planning based on the improved ant colony algorithm for multiple autonomous vehicles
CN116629511A (en) Multi-star dynamic task planning method and device based on two-stage hybrid scheduling in uncertain environment
CN116203990A (en) Unmanned plane path planning method and system based on gradient descent method
CN116382331A (en) Multi-unmanned aerial vehicle rapid scheduling method, device, equipment and readable storage medium
CN116185035A (en) Unmanned cluster dynamic task allocation method and system based on improved bionic wolf clusters
CN114916013A (en) Method, system and medium for optimizing unloading time delay of edge task based on vehicle track prediction
CN115187056A (en) Multi-agent cooperative resource allocation method considering fairness principle
CN115016540A (en) Multi-unmanned aerial vehicle disaster situation detection method and system
CN114167896B (en) Unmanned aerial vehicle cluster control method for artificial potential energy field based on state prediction
Liu et al. Research on Cooperative Search Technology of Heterogeneous UAVs in Complex Environments

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant