WO2004018158A2 - Organizing groups of self-configurable mobile robotic agents - Google Patents

Organizing groups of self-configurable mobile robotic agents Download PDF

Info

Publication number
WO2004018158A2
WO2004018158A2 PCT/US2003/026764 US0326764W WO2004018158A2 WO 2004018158 A2 WO2004018158 A2 WO 2004018158A2 US 0326764 W US0326764 W US 0326764W WO 2004018158 A2 WO2004018158 A2 WO 2004018158A2
Authority
WO
WIPO (PCT)
Prior art keywords
mras
mra
ofthe
mrs
agents
Prior art date
Application number
PCT/US2003/026764
Other languages
French (fr)
Other versions
WO2004018158A3 (en
Inventor
Neal Solomon
Original Assignee
Neal Solomon
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Neal Solomon filed Critical Neal Solomon
Priority to AU2003262893A priority Critical patent/AU2003262893A1/en
Priority to JP2004531235A priority patent/JP2005539296A/en
Priority to EP03793423A priority patent/EP1563348A2/en
Publication of WO2004018158A2 publication Critical patent/WO2004018158A2/en
Publication of WO2004018158A3 publication Critical patent/WO2004018158A3/en

Links

Classifications

    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/0088Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots characterized by the autonomous decision making process, e.g. artificial intelligence, predefined behaviours
    • FMECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
    • F41WEAPONS
    • F41HARMOUR; ARMOURED TURRETS; ARMOURED OR ARMED VEHICLES; MEANS OF ATTACK OR DEFENCE, e.g. CAMOUFLAGE, IN GENERAL
    • F41H13/00Means of attack or defence not otherwise provided for

Definitions

  • Kawakami mobile robot control system
  • U.S. Patent No. 5,652,489 Asama et al.
  • Asama et al. mobile robot sensor system
  • U.S. Patent No. 5,819,008 and Wallach et al.
  • U.S. Patent No. 6,374,155 involve multiple mobile robots. These patents involve using sensors for navigation and obstacle avoidance. In addition, one mobile robot can transmit information to another mobile robot for some effect. These inventions offer only rudimentary connections between robots and lack advanced system functions.
  • SFI Sante Fe Institute
  • SFI theorists have developed the swarm intelligence model of artificial computer societies primarily for simulating economic systems.
  • the swarm intelligence model by emulating biological system operation, uses ideas of emergent behavior to describe the complex social interactions of relatively simple insects according to straightforward decentralized rules governing group activity.
  • Arkin developed a behavior-based model of robotics
  • Arkin describes behavior-based robotic architectures as well as experiments in the field with sophisticated hybrid robotic architectures.
  • An example of this hybrid approach is NASA's Atlantis system (1991) that synthesizes deliberative planning with group behavior.
  • NASA's Atlantis system (1991) that synthesizes deliberative planning with group behavior.
  • the aim of these models is to develop autonomous robots that are adaptive to their environment.
  • the development of robotic teams with social behavior is one ofthe most difficult challenges, according to Arkin' s pioneer study.
  • MRS models have been developed.
  • the Nerd Herd is an example of an MRS using rule-based social behaviors for subsumption based foraging popularized by Brooks.
  • the Alliance architecture developed a modular approach to robot team behavior that includes inter-robot communication. Such communication allows for emergent cooperation.
  • An additional version of Alliance (L- Alliance) accommodates the learning aspect of robotic agents in order to achieve a form of adaptation.
  • Arkin developed a "multiagent schema-based robotic architecture" in which team cooperation was modeled using a behavior-based approach without explicit inter-robot communication.
  • Dias and Stentz provide a market-based model for multirobotic coordination in which individual robots in a distributed environment negotiate with each other in order to agree upon a course of action.
  • Such a model applies the contract-net protocol used with software agents in a distributed network to the robotics context for operation of groups of autonomous robots in dynamic environments.
  • MRSs possess several common traits, including mobility, intelligence, communications, group behavior and specific functionality.
  • a related research stream involves A-NN, which has utilized GA in order to establish weight values of neural nodes.
  • One main aim ofthe neural networks is to develop self-configuring and self-organizing learning systems for complex problem solving. This is useful in real time collective robotics situations in which rapid adaptation to a changing environment is necessary.
  • the present inventions involve multi robotic systems, multi agent systems, collective robotics, artificial group behaviors, aggregation of robotic agents, coalition formation, dynamic coalitions, self-organization of robotic agents, emergent behavior of intelligent agents, cooperation of intelligent agents, multi agent learning, problem solving between conflicting intelligent agents, artificial intelligence, artificial neural networks and multi robotic operating systems.
  • Multi-robotic systems are complex networks that facilitate the interaction between autonomous robotic agents according to specific rules of behavior in order to perform a specific function or combination of functions.
  • the present invention describes a system for multiple mobile robotic behavior by applying the logic of advanced computer science, in particular artificial intelligence (Al), with advanced robotic electronics and mechanics.
  • Al artificial intelligence
  • the focus here is on artificial robotic collectives. So far very little research has been developed on the group behavior aspects of robotic societies as they plan, and then achieve, a coordinated goal.
  • any such collective robotic system including (1) the computation, electrical and mechanical hardware of each autonomous robot unit, (2) a hardware network layer that links the individual robots together with wireless communications, (3) a metacomputing layer (that performs complex memory, database and computation analysis functions) in a node to node distributed computing model, (4) an omni- nodal artificial neural network (A-NN) layer for distributed Al, (5) an evolutionary A-NN layer - driven by genetic algorithms and genetic programming - for adaptive group learning in order to develop real-time cellular automata (CA) based simulations to seek optimal system solutions, (6) an OS layer and (7) a layer for specific functional applications.
  • A-NN omni- nodal artificial neural network
  • the present invention describes a sophisticated MRS that is dynamic, interactive and evolving, adaptive to its environment and capable of exhibiting emergent behavior.
  • the system is designed as a hybrid of behavior-based and central planning control processes in a distributed network environment.
  • groups of autonomous robotic agents can learn together, make group decisions together (cooperatively and competitively), negotiate and solve problems together, congregate together in various sub-sets and re-configure in non-overlapping sub- groups.
  • autonomous robotic agents can form and reform into various configurations of groups in a self-organized way interacting with each other and with the environment in order to achieve pre-programmed, or evolved, goal parameters.
  • Artificial intelligence is used in a number of MRS processes, including individual robot learning and decision making using genetic algorithms (GAs), genetic programming (GP) and other evolutionary computation (EC) approaches as well as group robotic agents that uses A-NN and hybrid evolutionary A-NN approaches (including GA, GP, FL, etc.) that provide tools for adaptive collective learning and decision making.
  • GAs genetic algorithms
  • GP genetic programming
  • EC evolutionary computation
  • group robotic agents that uses A-NN and hybrid evolutionary A-NN approaches (including GA, GP, FL, etc.) that provide tools for adaptive collective learning and decision making.
  • the use of both individual agent and group learning tools are important because though the collective resources are far greater, when the system defaults to behavior-based biases, for instance, in situations with diminished computation resources, it is necessary for the individual robotic agents to have the tools to maintain autonomy.
  • a relatively simple architecture would consist of a leader robot with various followers in a hierarchy.
  • the leader possesses increased autonomy and orders the followers (super-drones).
  • pre-selected squadrons are formed, the control for which can be manually intervened by human interaction processes such as a video feed for mission objective alteration.
  • Reprogrammable orders and priorities can be uploaded at any time.
  • supplementary external computation resources can be kept outside ofthe MRS and fed in as needed by satellite.
  • computation is performed externally to the MRS, analytical results can be used to control the system.
  • reporting on agent behavior can be provided to an off-site blackboard so as to unify control at a central command center.
  • the MRS system is linked together in a distributed network of autonomous robotic agents that employ powerful computation resources and Al processes, the system can automatically "think” like a group and constantly reconfigure to the best available situation while interacting with and adapting to its environment.
  • a pure behavior-based reactive MRS architecture has advantages of local control and emergent behavior but disadvantages ofthe inability to control large groups in complex adaptive environments.
  • a central deliberative MRS control architecture has the ability to develop large self-organizing interactive systems and sub- systems but has the limits of being cumbersome and dependent on substantial computation resources.
  • What is needed in order to build and operate a complex and high performance MRS is a hybrid architecture.
  • the MRS architecture is a complex, continuously reconfiguring, operating system that links together robotic agents with computation, communications and software subsystems.
  • Such a system must be modular (so that upgrades in a subsystem can be seamlessly performed), scalable (so that nodes can be added or removed) and reconfigurable.
  • the system uses mobile software program code that provides inputs and outputs to robot machine agents.
  • the "Harness" dynamic reconfigurable metacomputing model is a pioneer for this mobile self-organizing MRS hybrid approach because it continuously seeks to re-route the system to the optimal computation and communication pathways.
  • each robotic machine unit has sensors, actuators, microprocessors, communication receivers and transmitters, power supply, a specific functionality and (system and applications) software.
  • MRS specific mobile robotic unit sensors when they are linked together, the opportunity exists for the MRS specific mobile robotic unit sensors to be organized into a network for collective data acquisition.
  • the group's collective computation resources can analyze the sensor data.
  • the group of mobile robotic agents can use complex Al induced learning processes to make group decisions, even in the face of noisy, error-prone and conflicting data streams. By maximizing the efficiency ofthe available group MRS resources, intelligent group behavior can emerge.
  • A-NN artificial neural networks
  • GA and FL soft computing
  • the A-NN system can optimize adaptation to its environment.
  • E-A-NN Evolutionary A-NN
  • neurevolution is useful for reinforcement learning.
  • A-NN's work by using genetic algorithms to adapt input features, learning rules and connection weights.
  • One ofthe most effective applications for A-NN is nonlinear statistical models such as pattern recognition.
  • Bayesian networks use hypotheses as intermediaries between data and predictions to make probability-based estimates of solutions. Hopfield networks are used to remember an earlier network configuration and to revert to an old network when noisy data limits continuing network development.
  • the present invention uses a hybrid approach to Al that combines GA and GP with A-NN and D-AI architectures.
  • the combination of evolutionary computation approaches with distributed neurocomputing models produces a system that constantly rewires itself as the system is reconfigured. This approach is necessary because finite computation resources need to be maximized even while the distributed mobile MRS changes. Not only is this scheme scalable but increased computation capacity can be provided on demand if needed by specific under used MRAs.
  • Such a hybrid Al architecture is best suited for learning by groups in a distributed network as well as for optimal adaptation to dynamic environments.
  • Hybrid Al approaches can be useful when solving complex problems.
  • Two main problem solving models involve either cooperative (altruistic) or conflict (self- interested) oriented agent behavior.
  • One main computational challenge that involves MRS is the distributed problem solving that requires negotiation among conflicting autonomous agents.
  • MRS action starts with a plan.
  • group action plans we can model the optimal configuration or allocate the most efficient resources.
  • Decision logic processes lead to identifying trade offs (parameters) between possible solutions that lead to an optimal problem solving choice.
  • MRAs use computation optimization techniques to select optimal solutions to complex problems in uncertain environments. By mapping various scenarios, using Al and decision processes in a distributed network, MRAs select the best plan to achieve objectives.
  • MRSs use advanced hybrid Al methods in order to achieve optimal grouping patterns of behavior. Unlike purely computational MASs, a MRS have physical dimension and motion in space. These physical and geometric realities about the practical operations of MRSs involve the need to organize spatial interactions and movements. It is useful to model these MRA movements before actually performing specific maneuvers primarily through the use of simulations.
  • CA Cellular automata
  • the results of combinatorial optimization approaches to seek the best solutions to solve problems can be represented by CA simulations and, thereby, tested, before actually implementing these decision choices.
  • the MRS solves problems and can seek improved solutions that can capture subtle contingencies in complex operational situations.
  • MAS swarms are tested in particle simulations using CA models, but MRSs have not applied these important CA driven simulations for real geometric behaviors. Therefore, the present invention uses simulations in a dynamic, rather than merely static, way, for real time testing. In the simulation, virtual robots are provided the valuable advantage of trial and error of potentialities of activity so as to learn from complex contingencies, in order to optimize the chances for mission success.
  • ants use pheromones (chemicals that have an odor to attract others) to develop complex foraging behaviors. By laying down pheromones, which, though temporary, can be increasingly intense if compounded, ants provide a natural reinforcement mechanism (stigmation) with positive feedback. This positive reinforcement learning mechanism suggests a self-organizing system.
  • Insects may communicate with each other indirectly.
  • the process of stigmation operates with an insect affecting, or changing, the environment, which then catalyzes other insect behavior.
  • the use of pheromones illustrates this process because the ants lay down an attracting chemical that may be acted upon by others in a limited time.
  • Flocking is a case in point. Each bird in a flock has limited information about flockmates. Instead, they have neighbors they provide local information on direction and speed. The big challenge is to avoid collision with neighbors even as they signal trajectory and velocity data through their behavior. Consequently, both attractive and repulsive forces are involving in flocking behaviors.
  • any individual can be a leader that initiates action, the recruitment of other individuals through attracting the cooperation of similarly interested neighbors is key to the process because these individuals then respond by attracting more neighbors, and so on.
  • any individual can initiate a swarm or flock; this initiation is a sort of initial request to procure resources for a specific (defensive or offensive) function or activity.
  • specific decentralized individuals can trigger group activity in a sort of local reactive chain reaction process that has the effect of overwhelming an enemy.
  • specialists alone, such as soldier ants may swarm for an attack process.
  • Insect and animal social behaviors are important to understanding complex social processes involving simple individuals. Attempts have been made to emulate biological system swarm intelligence for development of artificial systems of robots. For instance, Arkin's (1998) use of Brooks' simple modular reactive robot for group behavior shows an attempt to model complex behaviors from simple robots.
  • Simple swarm behaviors have anonymous homogeneous simple members (in uniform roles) with primitive local communication, minimal computation capacity and the limits of reactive behaviors using a narrow set of rules for learning and action. The limits of this biologically inspired system can be improved by development of an advanced MRS that exhibits social intelligence.
  • Our system has autonomous individual MRAs with highly advanced computation, Al and communications capabilities, complex learning and simulation functions, specialization features and team behaviors in a heterogeneous system.
  • the present invention emulates human social behavior by using artificially thinking mobile robotic agents for a range of functions.
  • aggregation is important as a process for organizing groups of MRAs within an MRS.
  • Intelligent aggregation of MRAs involves automatic selection, formation, combination, reformation and dissipation of groups.
  • Each new set of intelligent agents represents a new configuration.
  • Emergent behavior ofthe MRS leads to a complex self-organizing system that never settles on an equilibrium because it is constantly changing.
  • the application in an MRS involves the geometry of space and extension and the physics and mechanics of motion.
  • Aggregation is a process of grouping entities together.
  • game theory As applied to an MRS, game theoretic models have a geometric dimension. Game theoretic approaches to modeling an MRS is useful particularly because they can be multi-phasal and interactive. Not only are MRA interactions nicely modeled but complex interactions between sub-groups can be more optimally represented as well as interactions with the environment. Game theory can model cooperating agent behavior as well as conflicting or heterogeneous behaviors.
  • An example of a heuristic for MRS game theoretic modeling parallels chess playing maneuvers, with openings, gambits and traps providing MRA models for the inter-operation of artificial societies. Robotic agents work together to develop winning game strategies for achieving goals or solving problems.
  • One ofthe aims ofthe present invention is to develop methods for MRAs to constantly develop shifting groups. We are interested in discovering how intelligent autonomous robotic agents form and reform into dynamic coalitions of collectives. Understanding precisely how sub-groups of MRAs organize, self-configure, reconstitute, adapt to their environment and regroup is the key to understanding complex emergent group behavior in intelligent self-organizing systems.
  • squads of MRAs break off from larger groups in an MRS.
  • the squads can share the larger computation, communication and sensor resources and decision processes of the larger group.
  • the squads operate as teams of nodes in a neural net that constantly reconfigures on the fly. Since some ofthe sensors in some ofthe squads are exogenous to each team, the squads have access to data streams beyond any limited team. Sub-teams are synchronized into the distributed network using hybrid Al approaches. Nevertheless, each squad, and its reconfiguring teammates, can work independently with local behaviors.
  • differences between agents in a squad for example, specialists or different "personalities," can create complexity in squad behavior within the practical constraints of their programming, as they inter-relate in different configurations.
  • the present system has a number of innovations and advantages over earlier inventions. These innovations involve (1) multi-robotic system architecture, (2) computation resource structure, dynamics and allocation, (3) Al dynamics, (4) group negotiation, learning and decision structures and processes, (5) intelligent social behavior involving mobile robots and (6) dynamic coalitions of MRAs.
  • the present invention utilizes a novel hybrid MRS architecture that dynamically adjusts from manual operation of groups of MRAs to wholly automated socially intelligent MRAs in order to accommodate severe resource restrictions as well as extremely complex behaviors.
  • the system optimally adjusts to environmental conditions. For instance, very small MRAs may be resource constrained and would thereby employ simpler local reactive behavioral rules.
  • the architecture ofthe present system is also both modular and scalable so that growth or shrinkage will not affect performance.
  • the present system uses a distributed wireless grid supercomputing model.
  • This approach allows the sharing of computation resources, including memory, database storage and data analysis capacity, thereby far extending previous constraints, hi addition, this distributed model is optimal for equal node parallel processing within a collective. Computation processing speeds of dozens of teraops could be maintained in this system, thereby providing ample resources for complex group behaviors.
  • the present system also uses advanced routing procedures to maximize the most efficient geodesic heuristics.
  • the present system employs a novel use of a MAS within a MRS in order to communicate, negotiate, control and organize group behaviors.
  • Intelligent mobile software agents LMSAs
  • INAs intelligent negotiation agents
  • the present invention uses a dynamic reconfigurable evolutionary A-NN that provides optimal adaptation to the changing environments of an intelligent MRS.
  • the A-NN uses hybrid Al techniques, including combinations of GA, GP, FL and EC. As nodes are added or subtracted to the network, the A-NN is automatically rewired for maximum efficiency.
  • the system uses feedback loops to learn.
  • the A-NN is useful to train the system in group learning processes.
  • the present system In order for the present system to learn, it employs FL processes that use probabilities to make group decisions by selecting the best available option among a range of contestant options.
  • the system utilizes combinatorial optimization approaches to select the best solution to solve problems. Particularly in conflicting situations between agents, there is a need to negotiate a settlement by developing a method of winner determination.
  • the system employs novel approaches to asymmetric problem-solving by using multi-lateral negotiation methods.
  • the present invention uses game theoretic approaches and cellular automata schemas in order to simulate tactical system opportunities for an MRS in novel ways.
  • an MRS can automatically select an optimal problem-solving path and, hence, model complex interaction dynamics among MRAs and between MRAs and the environment.
  • simulation modeling for action planning and contingency scenario testing is necessary to achieve highly intelligent MRS behavior.
  • the present system is novel because it is heterogeneous.
  • the MRS employs specialty robots for diverse functions. Some MRAs may have multiple functions, alternative functions or work in teams with complementary functions. This approach increases efficiency of task execution because it promotes an automated division of labor in an MRS.
  • any agent can initiate group behaviors.
  • the attraction of MRAs to collectives can be demand-initiated in a novel implementation of group behavior in an MRS. This approach enhances system performance. In one implementation, stronger data inputs may constitute invitations to act beyond a specific threshold and thereby initiate MRA grouping behaviors.
  • the present system uses novel group attraction initiation methods. [0086] The present system synthesizes local control with deliberative planning. This hybrid architecture is novel and is possible only with the unique convergence of advanced computation technologies disclosed herein.
  • the present system uses novel approaches to dynamic coalition formation. Using these approaches, the MRS constantly reconfigures its structure and dynamics in order to adapt to environmental changes. This more effective adaptation provides increased speed, precision, efficiency and effectiveness in mission critical situations.
  • the present system develops a way for groups of robotic agents to make decisions in cooperative and in conflicting situations in real time. This is a novel and important advance over earlier systems.
  • the present system implements novel MRS approaches involving tactical cooperating teams of MRAs. This sophisticated use ofthe system transcends earlier notions of artificial group intelligence.
  • robot groups allow an increased speed to do a task. Like in nature, groups are increasingly reliable since some may fail but the group still finishes the task. In addition, using robot groups to perform tasks can be more flexible than only individual robots. The present system offers higher performance benchmarks for these traditional advantages.
  • the present system most efficiently implements complex group behavior in an artificial robotic system. For example, unlike earlier artificial systems that seek to emulate insect behaviors, the present invention seeks to emulate, and transcend, complex human group judgment to develop a true social intelligence. Consequently, the present invention goes beyond robotic systems that focus primarily on local control ofthe nearest neighbor and reactive behaviors.
  • Robots can have specific functions for specialized purposes. One robot can clean, while another can dry. But specialized robots can have particularly high utility as they function in teams. While specific purpose robots are useful, multiple function robots are increasingly productive. Multiple function robots can switch roles or change forms as needed to complete complex tasks. The more tasks a robot can do because of its multiple specialties, the more plasticity and flexibility it has.
  • Multi-functional teams of robots can perform more tasks than specific specialty robots. The more tasks that robots can do, the more plasticity of tasks a team of robots can perform because ofthe efficiency benefits ofthe maximized division of labor.
  • the present system enhances factory production, assembly and distribution processes. Methods for groups of robots to work together may greatly accelerate production techniques. For instance, by using groups of multi-functional autonomous robots, a host products can be produced faster, more efficiently and cheaper than with earlier methods.
  • Self-organizing teams of autonomous robots can build and repair roads and structures. From laying track or pipe to electrical, plumbing, framing and roofing, an MRS can be useful in performing laborious time-intensive routine structure building construction functions. Similarly, MRAs can be useful in the repair of buildings and streets. These novel MRA processes can save time and reduce costs of building construction as well as road work and repair. In one practical application, pot holes can be automatically detected and repaired by teams of MRAs.
  • medi-bots groups of medical robots
  • medi-bots can be used in critical field situations to stabilize a patient.
  • Autonomous medi-bots work together to (a) diagnose a patient's trauma, (b) resuscitate, via electronic pulse or CPR, a patient whose cardiac or pulmonary functions have ceased, (c) cauterize wounds to stop (or minimize) bleeding, (d) apply an IN for intravenous solution transmission in order to replace vital fluids and (e) call for more medical resources by providing a precise physical location position.
  • Multiple medi-bots can much more efficiently rescue and stabilize patients, thereby saving lives.
  • medi-bot can assist doctors in clinical situations by performing functions typically attributed to nurses and assistants. Such medi-bots can monitor patient functions during procedures as well as actively support the surgeon or dentist so as to save time. These medi-bots can also supply expertise in critical operating room environments. In critical emergency room situations, where time and precision can make a difference, medi-bots can save lives.
  • MRAs can transmit real-time vision and sound to off-site locations, typically via satellites or terrestrial communications systems.
  • the MRAs can be very small micro robots (more fully referenced below) that provide stealth advantages for reconnaissance and surveillance purposes.
  • MRAs can be disguised as natural phenomena, such as animals, birds, insects, etc. for evasive and stealthy advantages. By emulating natural animal behaviors, mission effectiveness can be maximized.
  • an MRA in this system can erase its programming and be rendered a useless pile of sensors, while the remaining network nodes automatically reconfigure for effective performance.
  • Groups of MRAs can be used to perform complex clean-up operations that may be hazardous to humans. These clean-up categories include: (a) toxic waste dumps, (b) nuclear reactor cleaning, (c) oil spill events and (d) sewer cleaning.
  • MRAs can use self-organizing maps of a local terrain to devise plans to most efficiently and safely provide toxic clean-up operations, thereby saving lives and protecting the environment.
  • an MRS can be used to fight fires.
  • Ground MRAs can dig trenches and plot trajectories for the expanding fire territory, while aerial MRAs can drop fire retardant at tactical locations for optimal effect.
  • MRAs use self- organizing mapping processes to assess the scope and dynamics ofthe full-motion fire situation. Fire-fighting MRAs can save lives and protect property. This application can be useful for forest fires, urban fires or industrial structure fires that require complex problem solving and decisive action. Medi-bots can be used in conjunction with these fire fighting applications for maximum benefits.
  • MRA teams can be very useful for mining minerals in remote locations. Robots can identify the most promising locations to dig and then help with laborious digging and sifting tasks. Groups of MRAs can work faster and more efficiently than current automation processes, in part because they are mobile, autonomous and self-organizing.
  • Farming has enjoyed increased automation processes for generations so as to maximize production. Groups of MRAs can continue this automation evolution, particularly in the planting and harvesting contexts in which greater care is required for specific crops such as fruit and vegetables. In general, MRAs replace the routine functions of migrant pickers.
  • MRAs can operate effectively on specific problems. Groups of MRAs can provide effective automated solutions to hazardous functions, thereby reducing risks and saving time and money.
  • MRAs One main activity for MRAs involves demining. Groups of autonomous robots can work together to either dis-assemble or explode mines that are discovered in a self-organized search process. In addition, disarming bombs can be a useful function for groups of MRAs.
  • MRS vehicle categories may include cars, trucks, trains, aircraft and ships.
  • cargo may be moved on various groups of autonomous vehicles for greater efficiency, timeliness and cost-benefit.
  • traffic coordination systems may develop complex routing algorithms that emulate, and transcend, bird flocking or ant foraging behaviors. Elevator and Dam System Regulation
  • the present system is useful to organize groups of weather balloons or aircraft to gather and disseminate data.
  • the MRS is ideally suited to complex adaptive environments such as detecting dangerous weather conditions such as tornados or hurricanes. Groups of self-organizing MRAs can more rapidly predict dramatic weather system changes.
  • MRAs can not only predict poor weather but can influence its outcome. In a drought situation, MRAs can seed clouds to increase the likelihood of inducing rain. In an extreme case, MRAs can prevent tornadoes by influencing their movement very early in their development and changing the immediate environmental conditions. Only self-organizing groups of automated mobile robotic agents with specific functions - such as warming cool air in limited areas so as to retard or minimize a turbulent cyclic force — could execute this precisely or rapidly.
  • Groups of satellites can work together to perform distinctive functions such as optimally tracking moving objects by using the present system.
  • the present system can also be used to have groups of self-organized autonomous MRAs repair or readjust a satellite remotely.
  • the present system can be used in underwater applications.
  • the underwater context can be used with other applications, including surveillance, reconnaissance, search and rescuer and demining.
  • Groups of MRAs can be used as an automated system of sentries for security protection purposes. Sentries can be used not only for surveillance but also for defensive uses in order to protect structures or personnel. Such MRA sentries detect and respond to invasive action by unauthorized personnel by tracking and evading the intruders and calling for assistance. In a more aggressive mode, automated sentries can respond to invasive behaviors by disarming and subduing unauthorized activities until the authorities can arrive.
  • the present system can be used by groups of MRAs that operate video or film cameras in order to capture dynamic movie scenes. Because the MRAs can be constantly moving and can be both self-organizing and synchronized, an MRA can facilitate a new generation of film-making techniques, particularly for the popular action sequences. While moving in synchronized or random ways, MRAs are well suited to capture moving scenes in distinctive cinemagraphic ways only possible in an MRS.
  • Routine restaurant food preparation and delivery and commercial laundry functions can be done by teams of MRAs. Working as a group of short order cooks, MRAs can produce more variety of recipes in a shorter time than professional chefs or waitresses. Similarly, a commercial laundry service can be optimized by using groups of MRAs to organize, clean and package clothes. One hour discount cleaning is now possible by using an MRS.
  • disaggregated collectives of micro MRAs can form together into a larger composite robot exhibiting unified behavior. This is important so as to allow larger robots to disassemble into constituent (specialized) parts if necessary in order to evade a predator or disguise a maneuver.
  • micro MRAs using the present system could inspect and assemble micro-electronic systems or could inspect biological entities for abnormalities.
  • nanorobots can assemble into larger composites that themselves work together as autonomous groups.
  • Nanorobots include surveillance and reconnaissance. But more fanciful uses include biological applications that include cleaning arteries by injecting a group of nano-MRAs into a patient's blood stream. The nanorobots will go to the affected area, perform the operation internally and regroup for extraction. Nanorobots could also be used to identify and repair microelectronic abnormalities.
  • Groups of anthropological MRA "androids" can work together to form complex expert systems.
  • robot expert groups can behave like specialist teammates to collect and analyze data, perform forecasting, develop alternative scenarios, make predictions and give advice in the form of reports.
  • Such groups of expert consulting opinions can involve numerous substantive industry categories and topics, including optimal telecom and energy routing algorithms and economic, business industry and scientific analyses.
  • the personalities, experience and learning processes ofthe android MRAs evolve.
  • Such expert systems constitute a think tank.
  • such a group of autonomous self-organizing robotic agents can form and reform coalitions of specialist experts similar to a sophisticated consulting firm.
  • evolutionary learning and combining various opinions such complex systems can be creative and capable of original thinking approaches that far surpass chess playing supercomputers.
  • MRAs can take numerous forms. Since there are numerous applications ofthe present system in divergent industrial and technical contexts, it is appropriate to identify the structure and function ofthe variety of MRAs that can perform various jobs.
  • MRA vehicles can include various forms of aircraft, such as airplane, glider, helicopter, balloon, blimp, satellite or spacecraft. MRAs can operate in water as ships, boats, submarines or hovercraft. On land, MRAs can be automobiles, trucks, farm equipment, mining equipment, factory equipment, etc. There may be entirely new forms of MRAs as well, such as remote exploration devices, anthropological androids, micro-robots intended to emulate insect appearances, nano-robots and so on. The range of sizes and forms of MRA are very broad.
  • INA Intelligent negotiation agent
  • D-AI Distributed artificial intelligence
  • A-NN Artificial neural network
  • E-A-NN Evolutionary artificial neural network
  • CA Cellular automata
  • Fig. 1 is a list of system layers
  • Fig. 2. is a schematic diagram of a synthetic hybrid control system for an
  • Fig. 3 is a table of a dynamic database organization
  • Fig. 4 is an illustration of three MRAs identifying MRA locations with sensors
  • Fig. 5 is a diagram of an MRA assessing its environmental situation and coordinating change in state
  • Fig. 6 illustrates a diagram of a metacomputing model for distributed MRS in which flexible mobile grid architecture is organized into dynamic clusters
  • FIG. 7 is an illustration showing the sharing of computation resources among
  • MRA nodes in a wireless mobile MRS including the efficient routing of database and analytical functions
  • Fig. 8 is a diagram showing database coordination in a distributed MRS
  • FIG. 9 is a diagram showing a dynamic distributed object relational database data flow process
  • Fig. 10 is a diagram showing temporal objects in an object relational database management system
  • Fig. 11 is a diagram showing mobile grid dynamics
  • Fig. 12 is a diagram showing autonomous blackboards for MRAs
  • Fig. 13 illustrates a diagram showing intelligent mobile software agents operations control in MRAs
  • Fig. 14 is a flow chart showing MRA juvenile and adult training levels
  • Fig. 15 is a diagram showing MRA attitude biases
  • Fig. 16 is a flow diagram showing the learning and adaptation from environmental interaction
  • Fig. 17 is a flow diagram showing the MRA training process
  • Fig. 18 is a flow diagram showing reinforcement learning
  • Fig. 19 is a flow diagram showing hybrid learning with time constraints
  • Fig. 20 is an illustration of social learning in which MRAs learn from other MRAs
  • Fig. 21 is an illustration showing MRAs that teach other MRAs
  • FIG. 22 is an illustration showing asymmetric MRA leadership and the emergence of temporary hubs
  • FIG. 23 is an illustration showing specialized learning in self-organizing teams
  • FIG. 24 is an illustration showing automated specialization in which self- organization by task division occurs for individual specialization
  • Fig. 25 is a flow diagram showing a self-organizing map
  • Fig. 26 is a flow diagram showing a genetic algorithm
  • FIG. 27 is an illustration showing a binary genetic algorithm
  • Fig. 28 is an illustration showing a genetic programming tree architecture
  • FIG. 29 is an illustration showing parallel subpopulations fitness evaluation
  • Fig. 30 is an illustration showing a two layer neural network
  • Fig. 31 illustrates an artificial neural network connection weights
  • Fig. 32 illustrates genetic programming in the calculation of initial weights
  • Fig. 33 illustrates genetic programming applied to indeterministic artificial neural networks
  • Fig. 34 is an illustration showing an evolutionary artificial network connection and node additions
  • Fig. 35 illustrates evolutionary indeterministic artificial neural network feed forward progress
  • Fig. 36 illustrates an evolutionary search for connection weights in an ANN
  • Fig. 37 is a flow diagram showing a fuzzy logic module
  • Fig. 38 is an illustration of a neuro fuzzy controller with two input variables and three rules
  • Fig. 39 illustrates a five layer evolving fuzzy neural network
  • Fig. 40 illustrates an adaptive network based fuzzy inference system
  • Fig. 41 illustrates a self-organizing neural fuzzy inference network architecture
  • Fig. 42 illustrates a dynamic evolving fuzzy neural network
  • Fig. 43 illustrates a flexible extensible distributed ANN in which ANN computation is shared between MRAs
  • Fig. 44 is an illustration showing intelligent mobile software agents (IMSA) dynamics in a multi-agent system with an emphasis on MRA interactions;
  • IMSA intelligent mobile software agents
  • Fig. 45 is an illustration showing IMSA relations between MRAs
  • Fig. 46 is a flow diagram showing the operation of analytical agents
  • Fig. 47 is a flow diagram showing the operation of search agents
  • Fig. 48 is a flow diagram showing the initial operation of intelligent negotiation agents (INAs).
  • INAs intelligent negotiation agents
  • Fig. 49 is a flow diagram showing IMSA intercommunications
  • Fig. 50 is a flow diagram showing INA architecture
  • Fig. 51 is a flow diagram showing the pre-negotiation process
  • Fig. 52 is a flow diagram showing INA logistics
  • Figs. 53 A and 53B are a flow diagram showing negotiation in a distributed system with mobility
  • Fig. 54 is an illustration showing the simultaneous multi-lateral negotiation process with multiple variables
  • Fig. 55 is an illustration showing multivariate negotiation factors
  • Fig. 56 is a flow diagram showing winner determination in a competitive INA framework
  • Fig. 57 is a table showing the argumentation process
  • Fig. 58 is a flow diagram showing anticipation of opposing LNA strategies
  • Fig. 59 is a flow diagram showing problem identification in which a group of
  • Fig. 60 is a flow diagram showing solution option development between
  • Fig. 61 is a flow diagram showing a solution option selection method
  • Fig. 62 is a flow diagram showing how the MRAs select the best available solution to a problem in the present circumstance while waiting for more recent relevant information;
  • Fig. 63 illustrates MRA group agreement
  • Fig. 64 is a table that shows the temporal aspect ofthe decision process
  • Fig. 65 is a flow diagram showing the application of multivariate analysis to problem solving
  • Fig. 66 is a flow diagram showing the application of regression analysis to problem solving of conflicting MRAs for winner determination
  • Fig. 67 is a flow diagram showing the application of pattern analysis and trend analysis to problem solving of conflicting MRAs for winner determination;
  • Fig. 68 illustrates the modeling of MRS activity with simulations in which situation assessment is performed
  • Fig. 69 is a flow diagram showing the synchronization of simulations within an MRA cluster
  • Fig. 70 illustrates the contingency cellular automata (CA) scenario option simulations
  • Fig. 71 illustrates reversible CA projecting backwards from a goal
  • Fig. 72 illustrates adaptive geometric set theory applied to an MRS
  • Fig. 73 illustrates the optimal simulation selection in which simulation scenarios are (temporarily) converged
  • Fig. 74 is a flow diagram showing the initiation ofthe aggregation process in which sets of MRAs form from the larger collective;
  • Fig. 75 illustrates the initiation of homogeneous MRA group formation
  • Fig. 76 illustrates the initiation of common heterogeneous MRA group formation
  • Fig. 77 illustrates the initiation of complementary heterogeneous (specialized) MRA group formation
  • Fig. 78 is a flow diagram illustrating the initial phase of demand-initiated environmental adaptation
  • Fig. 79 illustrates continuous MRA group composition reconfiguration
  • Fig. 80 illustrates the continuous reconfiguration of sub-networks
  • Fig. 81 illustrates dynamic group behavior adaptation to environmental interaction
  • Fig. 82 is a flow diagram illustrating the parallel dynamic traveling salesman problem (TSP) with cooperating autonomous agents;
  • Fig. 83 illustrates the altruistic sacrifice of MRAs (gambit tactic) in order to acquire sensor information to increase chances of overall mission success;
  • Fig. 84 is a flow diagram illustrating the general dynamic coalition process
  • Fig. 85 illustrates group MRA coordination and obstacle avoidance
  • Fig. 86 illustrates specific MRA functionality via specialization
  • Fig. 87 illustrates specialized MRAs working as a team
  • Fig. 88 illustrates multi-functional self-organizing MRAs
  • Fig. 89 illustrates surveillance and reconnaissance of a mobile object sensed and tracked by multiple micro-MRAs
  • Fig. 90 illustrates remote exploration with initial tracking of multiple objects with multiple micro-MRAs
  • Fig. 91 illustrates sentry behavior within limited perimeters
  • Fig. 92 illustrates cinematography applications with MRAs in which objects are sensed and tracked
  • Fig. 93 illustrates land based toxic site clean up with multiple MRAs
  • Fig. 94 illustrates dynamic cleanup of an oil spill within limited hydro perimeters by multiple MRAs
  • Fig. 95 illustrates fire fighting with multiple MRAs as a dynamic interaction between the MRS and a complex environment
  • Fig. 96 illustrates manufacturing production in which an object is created by using multiple MRAs
  • Fig. 97 illustrates the assembly of objects in which parts are combined to create a whole object using multiple MRAs
  • Fig. 98 illustrates road generation using MRAs, and
  • Fig. 99 illustrates surgical micro MRAs used for trauma intervention and stabilization.
  • the system and methods incorporated in the present invention are implemented by using software program code applied to networks of computers.
  • the present invention represents a multirobotic system (MRS) that includes at least two mobile robotic agents (MRAs).
  • MRAs mobile robotic agents
  • MRAs have various useful purposes in the context of industrial and practical applications.
  • the MRAs use complex software program code, including mobile software agents, to execute specific instructions involving robotic and computation operations.
  • the software capabilities activate specific robotic functions within MRAs involving movement and decision-making.
  • the present invention focuses on how groups of autonomous MRAs operate in a distributed MRS.
  • the invention or cluster of methods, solves problems in the area of computation for groups of mobile robots in a distributed network.
  • the system shows novel ways for groups of MRAs to work together to achieve specific goals such as mapping the environment, coordinating missions, aggregating into dynamic coalitions and engaging in complex self-organizing activities.
  • the system employs hybrid models for collective robotic control that combines not only synthetic control methods that combine central and behavior- based approaches but also hybrid artificial intelligence methods.
  • Distributed artificial intelligence approaches are used in several contexts ofthe present system, including learning, negotiation, simulation and decision-making of MRAs and intelligent mobile software agents (LMSAs).
  • the main approach for decision making of MRA collectives is decentralized.
  • the MRS engages in learning and decision processes that employ extensive use of IMS As.
  • IMS As interact with each other to handle routine matters between MRAs, including communication, analysis and negotiation.
  • Intelligent negotiation agents (LNAs) provide a medium for multilateral interaction of MRAs for group decisions. Simulations are used extensively to model and select optimal pathways for MRA group action and for the evaluation of scenarios for action.
  • Fig. 1 illustrates the layers ofthe multi-robotic system architecture.
  • the first level shows a synthetic hybrid control system for MRAs including central planning control and behavior-based control aspects, which are further described in Fig. 2.
  • MRAs are independent autonomous agents that use Al to interact with their environment using the hybrid control model.
  • the second layer is the level ofthe mobile robotic system in a distributed network which connects together individual MRAs using communications.
  • the Grid computing architecture is used to link the MRAs together at layer three in order to share computation and database resources between the individual MRAs for maximum network efficiency.
  • the MRA network develops dynamic clusters for optimal computation and storage capability.
  • the mobile Grid network model is critical in order to accomplish complex tasks.
  • the dynamic distributed database system is used.
  • This extension ofthe Grid computing hardware architecture uses object relational databases and temporal data objects to organize data between databases in the MRAs.
  • LMSAs Intelligent mobile software agents
  • MAS multi-agent system
  • IMSAs are complex agents that perform a number of important functions within each MRA, such as analysis and decision-making, and between MRAs, such as data search, negotiation and collaboration.
  • the MRAs produce complex simulations to represent their relative positions and movements as well as to map out the possible scenarios for future action. These simulations are represented as mobile cellular automata in level seven.
  • the specific functional application of each implementation ofthe system comprises level eight.
  • the main application categories of remote sensing, hazard management and manufacturing processes each use specific functional representations that are closest to the environment with specific hardware types.
  • Fig. 2 shows a multi-layer architecture of an MRA synthetic hybrid control system.
  • the first level shows specific central (0270) and behavior-based (0280) control processes, in which the former uses abstract logic and the latter is reactive to the environment.
  • layer two the two main processes are intermediated (0260) in synthetic control approaches.
  • Layer three illustrates several main hybrid control systems that combine both central planning and behavior-based control models: (1) planning driven (0220), (2) advice mediation (0230), (3) adaptation (0240) and (4) postponement (0250).
  • the planning-driven approach to combining the main control methods determines the behavioral component; it is primarily a top-down model.
  • the advice mediation approach models the central planning function as advice giving, but allows the reactive model to decide; it is primarily a down-up model.
  • the adaptation model uses the central planning control module to continuously alter reaction in changing conditions.
  • the postponement model uses a least commitment approach to wait to the last moment to collect information from the reactive control module until it decides to act.
  • the suite of synthetic control systems (0210) is constructed of various combinations of these main hybrid control models.
  • a robotic unit may use a suite of hybrid control systems in order to optimize specific situations.
  • a single MRA unit includes a hardware component with an object-relational database. Within this MRA, software agents perform tasks such as analysis, negotiation and decision- making. On a more advanced level, a single MRA has complex computation resources to manage, including Al and ANN.
  • Fig. 3 shows that while a single autonomous unit is important, when combined with other similar units in a network and provided with mobility, and when also combined with both software agent system integration and Al and ANN capabilities, the system produces a complex adaptive collective capable of autonomous mobile interaction.
  • Figs. 4 and 5 show simple MRA operations such as using sensors to locate other MRAs or changing position by avoiding obstacles.
  • Fig. 4 shows a simple communication between three MRAs using sensors.
  • Each MRA uses its sensors to detect the positions ofthe other MRAs. In this way, each MRA can identify each others' positions.
  • the position of each MRA may be transmitted to other MRAs in the network by way of wireless communications.
  • positions of MRAs can be transmitted to other MRAs by satellite, radar or other external GPS tracking system. In these ways, the positions of MRAs can be tracked by other MRAs in the network.
  • MRA position tracking of other MRAs is important is that in a noisy environment, there are multiple methods for MRAs to track other MRAs. In the total absence of communication, an individual MRA may default to a behavior-based reactive mode of interacting with other MRAs and with the environment.
  • An individual MRA can detect an object (0520) in the environment with its sensors and change position from 0520 to 0530 as illustrated in Fig. 5.
  • Fig. 6 shows a metacomputing model for a distributed mobile robotic system (MRS).
  • MRA 1 requests (at (l)(a), (l)(b) and (l)(c)) computation resources and data storage capacity from other MRAs.
  • MRAs 2, 3 and 4 (at 0620, 0630 and 0640, respectively) then respond to the request (at (2)(a), (2)(b) and (2)(c), respectively) of MRA 1 (at 0650).
  • Fig. 7 illustrates the sharing of computation resources among MRA nodes in a wireless mobile MRS, with an emphasis on the routing of.database and analytical functions.
  • the distributed network of MRAs can work together as one dynamic unit. Messages are input to the report status distributor (0720) and the request coordinator (0730) The report status distributor feeds messages to the MRS (0740) which interacts with the cache (0750) and the data stream (0760). The cache also interacts with the analytical (0770) functions of the system. Messages are output from the data stream and from the request coordinator.
  • the mobile wireless grid computing architecture uses the most recent version ofthe message passing interface (MPI) for distributed computer networks. The use of grid architecture in a mobile wireless distributed network allows for a maximum of flexibility and scalability in providing massive resources in adaptive environments.
  • MPI message passing interface
  • Fig. 8 shows database coordination in a distributed MRS.
  • the front end (0810) inputs queries at the query initiator (0820) which inputs to the query executor (0830), which has buffers (0870) with other MRAs.
  • the multiple data sources (0850 and 0860) supply information to the query executor.
  • the query executor outputs its queries to output queues (0840) at various other MRAs (0880). This process is further illustrated in Fig. 9.
  • Fig. 9 the dynamic distributed object relational database data flow process is described.
  • the query origination (0910) moves to the various databases (0920), DB1 through DB5, internal to MRA 1 through MRA 5.
  • the query executor (0950) which is buffered (at 0970), searches the same databases (0980), which have sensor data stream inputs (0930) as data sources (0940). Once accessed, the databases output their data at the output queues (0990).
  • This distributed model shows a parallel network approach to database organization.
  • the system uses active storage databases in which the computer processing capacity is internal to the database, which is itself continuously mining objects for analytical functionality.
  • Temporal objects reveal their temporal priority in order to be listed in a higher or lower relative priority in the database for storage retrieval purposes.
  • Objects are "tagged" with temporal priorities such as "now”, “imminent”, “very soon”, “in the future”, “possibly useful in the future”, “past”, “near past”, “immediate past”, “urgent priority”, etc.
  • the query generator (1010) requests the query executor (1020) to access databases at DB1 (1030) and DB2 (1050) in sequential order. These databases access the data object (1060), which is tagged as it undergoes temporal change and is given temporal priority (1040) and is then provided back to the query executor (1020). Once again the databases are accessed with temporal information about the data object. The data object is then directed to the query manager (1070) for feedback to the system. By prioritizing data according to temporal priority, the system can route data efficiently and effectively anticipate functions. Temporal data is useful in the present system in the context of evolving learning, evolving ANN, evolving game theoretic negotiation applications, evolving environmental conditions and general systemic adaptation processes.
  • Fig. 11 shows the mobile grid dynamics. Data sets at a specific location inform the system analysis at 1120. The data sets are analyzed and interpreted at 1110 in order to determine where the system should move. The system moves to the new position at 1130. Yet this change of position provides new data sets, which are, in turn, provided to the system for analysis in order to determine where the system should move. This dynamic process optimizes the functionality ofthe system.
  • MRAs It is necessary for MRAs to obtain and transmit information from other MRAs about specific data such as physical position, analysis, negotiation and decision-making. Concise data sets are transmitted between MRAs in real time about the location and analytical state ofthe MRAs. These abbreviated data sets are consolidated in each MRA by autonomous blackboards, which act as "radar readouts" informing MRAs about the state of the network.
  • Fig. 12 autonomous MRA blackboards are described.
  • limited information is referenced involving spatial position, vector and speed so that each MRA can get a snapshot ofthe present situation of every other MRA in the system.
  • MRAs 1 through 4 readout specific data sets in a spreadsheet format at 1210 during phase one. New data sets are presented to the same MRAs in phase two to signify a change in state ofthe network.
  • an external blackboard keeps track ofthe data as a form of back up.
  • the leader would maintain the consolidated information function. If such a consolidated approach were used in a further embodiment ofthe system, the leader may shift, thereby providing fluidity for centralized leadership ofthe system.
  • Fig. 13 describes the operation of intelligent mobile software agents (LMSAs) among MRAs.
  • LMSAs intelligent mobile software agents
  • MRA 1 receives a collaboration agent sent by MRA 2 (1320), as it launches a search agent to both MRA 2 and 3 (1330).
  • An interaction process is engaged between MRA 1 and MRA 2.
  • an analytical agent is launched by MRA 3 to MRA 1, while a messenger sub-agent is launched from MRA 3 to MRA 2.
  • negotiation agents LNAs
  • Figs. 14 through 19 deal with MRA training and learning, while figures 20 through 25 deal with social learning.
  • Fig. 14 is a flow chart depicting the evolution of training level states.
  • an MRA initiates a training exercise (1410)
  • it increases levels of training (1420).
  • It may employ a learning module with specific learning tasks (1460) and refinement of learning tasks (1470) or it may interact with various environmental inputs (1430) in order to leam.
  • a juvenile training level is achieved (1480).
  • it improves learning with positive reinforcement (1440) and an adult training level is reached (1450), which is constantly reinforced with a feedback loop.
  • MRA attitude biases are shown in Fig. 15.
  • passive behavior the MRA acts with slower judgment but generally with more information
  • aggressive behavior the MRA acts with faster judgment but within information constraints because ofthe time limits of quicker action.
  • Environmental interaction is critical for learning and adaptation.
  • Fig. 16 shows a flow chart in which MRAs interact with both other MRAs and with the environment. After an MRA initiates a training exercise (1610), it either interacts with other MRAs (1620) or with its environment (1630).
  • an MRA queries other MRAs about a specific question (1640), while the MRAs then access databases and respond to the data query (1660). Inter-MRA feedback is then shared between MRAs for efficient learning (1680), akin to a tutorial.
  • the MRA feedback changes (1650). In this case, negative feedback is avoided (1665) while positive feedback is attractive behavior (1670) which leads to reinforcement learning (1675) and a feedback loop with the environment. As the environment changes, new data about these changes is supplied to MRA databases in order for them to access these environmental changes. When provided with positive feedback, the MRA constantly updates its beliefs about the environment (1690).
  • the MRA training process includes a combination of environmental interaction with group sensor data as illustrated in Fig. 17.
  • the MRA initiates learning (1710) and accesses either the sensor data from other MRA team members (1720) or the environment (1730). By accessing the environment direction, the MRA collects raw sensor data (1740). Whether obtained from other MRAs or directly from the environment, the MRA analyzes and interprets the sensor data (1750) and initiates a decision to act based on the data (1760). In this way, training processes may be implemented based on the data obtained, contingent on the method of originating the data (whether from the environment directly or from other MRAs). Whereas figure 17 shows the two main ways of obtaining data, Fig. 18 shows the two main qualities of information, viz., intensity and quantity of data, which provide MRA learning reinforcement.
  • sensor data is input into an MRA (1810) while the intensity of inputs is measured (1820) or the quantity of inputs is measured from different sources (1830). In either event, the inputs are compared to databases (1840) while each is provided a weighted value, with high intensity input weighting (1850) and quantity input weighting (1860), respectively.
  • the MRA evaluates the weighted value from different sources (1865) and interacts with the environment based on input evaluation (1870).
  • MRA 2 (2015) and MRA 3 interact with objects (2030) in the environment in an initial phase, hi the second phase, the MRAs interact with each other by sharing information about the object-interaction.
  • This descriptive phenomenology about the objects is used in the third phase by further interactions between the MRAs and the objects.
  • Fig. 21 illustrates an MRA that teaches another MRA.
  • MRA 2 (2120)
  • Fig. 22 illustrates this process.
  • MRAs (2210) interacts with a moving object (2220). But the leader is knocked out of action (2240) in the second phase, while a new leader emerges for the group (2230) as the new leader seeks the moving object (2250) and it is also removed from action. Finally, in phase three, yet another new leader emerges for the group (2260) while the mobile object (2470) continues to elude the group. At each new phase, a new hub is created with a new leader of the MRA cluster. In each case, the goal is to seek out the elusive mobile object.
  • Fig. 23 A division of labor can occur in specialized teams for increasingly efficient performance as shown in Fig. 23.
  • Each MRA is designated with a letter to signify its role as a specialist, while the whole group interacts with a mobile object (2320).
  • the MRAs reorganize into new positions in order to optimize the sharing of data and resources and to organize an interaction between the various specialists and the object (2340).
  • Fig. 24 further illustrates the self-organization process by task division for automatic individual specialization.
  • the group of MRAs (2410) interact with the object (2420).
  • the MRAs automatically activate a specific specialization mode (2430) to attack the increasingly elusive object (2440) as shown in the second phase.
  • the MRAs automatically reorganize to a new specialization mode (2450) to catch the object (2460).
  • Fig. 25 is a flow chart that shows the process ofa self-organizing map for a group of MRAs.
  • MRAs move to new locations to fulfill a mission (2520), where they receive sensor feedback from the environment (2530).
  • the MRAs create an initial map based on initial sensor data organization (2540) and obtain more sensor data (2550) as they cover more terrain to include in more refined mapping phases. In this way, the MRAs fill out the initial map to create a fuller picture of tenain to include formerly missing parts (2560).
  • the MRAs can perform this filling in procedure by using caching techniques that add the most recent information to a map outline. More complete data from sensors continue to refine the map (2570) as the MRAs continue to generate more and better data from continued mobility and data gathering. As objects in the environment change position, the MRA sensor data inputs that represent these changes continue to update the maps (2580).
  • FIGS. 26 through 29 show the main Al procedures of genetic algorithms and genetic programming. These techniques are then applied, in figures 30 through 43, to artificial neural networks. These discussions are important because Al and ANN are also applied to IMS As, to the negotiation process and to simulations, which will be addressed later in the figures.
  • Fig. 26 describes a flow chart of a genetic algorithm. After a population is created (2610) (and mutations added to the population (2620)), each member ofthe population is evaluated for fitness (2630). The weak members are pruned out (2640) and removed (2650) while the strongest members are selected for crossover (2660), such as breeding, which is then performed (2670). A feedback loop is generated in order to generate multiple generations of a population or a range of sub- populations.
  • Fig. 27 shows an example of a binary genetic algorithm crossover in which 2710 is bred with 2720 to achieve 2730.
  • a combination of "zero" and “one” yields a one, while two zeros or two ones combined in a specific position produces a zero.
  • FIG. 28 shows a genetic programming model with a crossover from the first phase of a tree on the left with a tree on the right.
  • the triangular grouping on the upper left (in the box) (2830) is combined (2880) with the tree ofthe upper right (including the triangular grouping in box (2970)), though the two groupings are "switched" right to left in the examples.
  • This tree structure modeling approach more closely resembles the actual genetic representation of evolutionary processes.
  • Fig. 29 uses the tree structure model to illustrate parallel subpopulation fitness evaluation in which two main triangular structures (2910 and 2920) break into a large number of smaller sub-populations (2930 and 2940) in order to assess the fitness ofthe best set of pairings. A final pairing is then selected (2950). Rather than mnning through a single sequence ofthe fitness assessment procedure, the parallel approach is much more time sensitive. This time sensitivity is more conducive to adaptive systems in which real-time interaction is critical.
  • Genetic algorithms, genetic programming and evolutionary computation techniques are applied to artificial neural networks in order to (1) calculate the initial weight and the connection weights ofthe signal between neurons, (2) train and optimize the connection weights, (3) generate the architecture and topology of a NN and (4) analyze the pattern, structure and phase state of a NN.
  • GA, GP and EC are also applicable to a range of complex computation problems, including (1) distributed problem solving, (2) group learning, (3) group cellular automata simulations, (4) routing of computation resources in the distributed system, (5) scheduling in a dynamic distributed system, (6) creating a self- organizing map, (7) solving optimization problems, (8) performing game theoretic simulations, (9) performing parallel data mining and (10) selecting a winner from among complex aggregation choices.
  • Figs. 30 through 43 deal with artificial neural networks.
  • ANNs and evolutionary ANNs have numerous applications to the present system, particularly (1) organizing and optimizing distributed networks, (2) performing dynamic data mining, (3) organizing indeterministic learning, (4) ordering and adapting simulations, (5) modeling and optimizing dynamic game theoretic interactions (6) structuring adaptive self-organization and (7) general problem solving.
  • the field of neural networks has evolved in the last generation from a purely theoretical endeavor of logicians, mathematicians and neuro-biologists to include applications that are useful for practical systems.
  • the present system is an example of an application of complex neural networks to learning, simulation and adaptive processes.
  • the neural networks are computational representations within the program code of MRA hardware that provide useful tools for calculations of specific solutions to problems.
  • ANNs are parallel computational systems including interconnected nodes.
  • ANNs have inputs and outputs in the connection weights between nodes.
  • An ANN node represents an artificial neuron that is modeled after biological neurons in a brain.
  • a perceptron is the structure that represents the sum of a neuron's inputs and outputs.
  • Fig. 30 shows a two layer neural network in which inputs are entered on the left side and outputs are registered on the right side ofthe figure. A feedback connection can be added that directs the connections back to the left side ofthe nodes.
  • a multilayer ANN is represented, with 3120 and 3140 representing the first layer, 3110, 3125 and 3150 representing the second, hidden, layer, and 3130 representing the output layer.
  • the ANN structure is a multilayer perceptron (MLP).
  • MLP multilayer perceptron
  • the connection weights are illustrated in numerical terms in this figure, with the bottom part having higher numbers than the upper part.
  • MLP multilayer perceptron
  • These types of ANNs can be classified as having feed-forward recall or feedback recall, being deterministic or indeterministic and, finally, possessing supervised learning or unsupervised learning.
  • Fig. 32 genetic programming is used to calculate initial connection weights.
  • the GP randomly generates a population, computes the fitness of its members, generates a new population by performing a crossover ofthe first generation and adding random mutations and, finally, seeks to identify the fitness of specific members of this most recent population by comparing the best fit members with the criteria to satisfy the problem of identifying the initial weight ofthe connection.
  • Fig. 33 shows how genetic programming is applied to an indeterministic ANN.
  • the multi-layer ANN has inputs that register higher relative numbers at the top (connections between 3315 and 3310, between 3310 and 3320 and between 3325 and 3320) ofthe network than at the bottom (connections between 3315 and 3325, between 3325 and 3330 and between 3325 and 3320) ofthe network.
  • the network grows, shown in phase two, it emphasizes growth at the top, where there is significantly higher activity, and adds nodes at 3340 and 3350, while lower positioned nodes at 3355, 3360 and 3365 become inactive.
  • Fig. 34 shows the automatic generation of a new node (3450) and a new connection (between 3420 and 3450) through a mutating process.
  • Fig. 35 illustrates an evolutionary ANN indeterministic feed forward progression from the first phase to the second phase.
  • new nodes 3560 and 3580
  • connections are added, while less active nodes (3550 and 3570) and their connections are made inactive. In this way, ANNs constantly "rewire" the network towards more productive nodes.
  • Fig. 36 shows a 3-2-1 multilayer network in which connection weights are calculated by genetic algorithms.
  • the algorithms are represented as binary units in order to calculate the connection weights.
  • the network is trained by fine-tuning the connection weights through a process of optimization that the successive generations of genetic algorithms perform.
  • Fuzzy logic is a method to provide new approaches to computing that includes terms like "maybe,” “possibly” and other partial and soft descriptions. Also called soft computing, FL represents a departure from traditional hard computing with mutually exclusive logic. FL uses statistical methods to compute solutions to complex real world problems. FL is applied to ANN to produce complex adaptive networks.
  • Fig. 37 describes a FL module. A sensor provides crisp data input (3710) to a fuzzifier module (3720), which is fed random mutations (3770) and proceeds with the fuzzification process. At this point, a fuzzy analysis proceeds in a fuzzy inference engine (3730) that operates according to fuzzy rules (3780) which are themselves adapted (3790).
  • the defuzzification of data occurs in the defuzzifier module (3740) where crisp data is output (3750) and presented to actuators (3760) for functional performance.
  • This process is similar to a the process a signal undergoes in conversion from an analogue waveform to a digital mode by way of an analogue-digital (A to D) converter, or, contrarily, from a digital to an analogue signal by way of a digital-to-analogue converter (DAC).
  • a to D analogue-digital
  • DAC digital-to-analogue converter
  • Fig. 38 shows a neuro fuzzy controller with two input variables and three rules.
  • the input variables Al (3810) and A2 (3820) provide connections to the rule base RI (3830), R2 (3840) and R3 (3850), which then provide an output at X (3860).
  • Fig. 39 shows a five layer evolving fuzzy neural network, with the input layer (3910), the fuzzification layer (3920), the rule node layer (3930), the decision layer (3940) and the output layer (3950).
  • a more complex ANN architecture is described in Fig. 40.
  • an adaptive network based fuzzy inference system is shown in which inputs are presented to the initial presentation layer, which is shown here in a parallel configuration, with R and S nodes.
  • a training process occurs in the multilayer network (4040) that contains hidden layers. The outputs of this training process are fed to the consequent parameters (4050) that then lead to outputs.
  • a multilayer neural fuzzy inference network is illustrated in Fig. 41.
  • the first layer (4110) generates the offspring (4115), which produce neural nodes at level three (4120) that are evaluated for fitness at level four (4125).
  • the nodes breed a new generation with inactive nodes (4130) at level five.
  • the surviving nodes (4135) again breed a new population of nodes that result in two active members (4140) in layer seven.
  • the successful mating of these nodes yields an output node (4145) at level eight.
  • Fig. 42 shows a dynamic evolving fuzzy neural network. With five layers, including an input layer (4250), a fuzzy quantification layer (4255), an evolving rule nodes layer (4260), a weighted least square estimator layer (4265) and an output layer (4270). This model shows a complex synthesis of simpler ANN representations.
  • One ofthe advantages of applying evolutionary computation to ANN is that such advanced computing can be performed more efficiently by using parallel approaches to break down a problem into smaller parts so that a larger number of computer processors may solve the problem simultaneously. In this way, multiple MRAs may work on a problem together in order to accomplish the task in real time.
  • One application of this approach is in the fitness evaluation part ofthe genetic algorithm population production process. The problem of identifying the successful candidates in a population can be performed, and expedited, by using parallel processing.
  • An example ofthe application of EC and ANN to an MRS is the modeling of game theoretic interactions.
  • a particular strategy may be evolved for a particular player based on a basic rule pattern selection organized by a multilayered feed forward perceptron. Each layer performs a calculation ofthe weights of inputs, connections and biases.
  • a random number of nodes is selected in the multilayer network, with a random number of offspring replicated from each parent and randomly mutated.
  • a number of rules of game moves are identified and consistently applied.
  • Each network generation is evaluated for accurate effectiveness of achieving a successful game move. The network is trained and retrained with full information, i this way, the learning process is refined so that each player is able to optimally move according to the rules.
  • MRA 1 (4410) launches a collaboration agent that is received by MRA 2 (4420) and collaboration between the two MRAs is initiated.
  • a search agent is launched from MRA 1 to search databases in MRA 3 (4430) and MRA 4 (4440).
  • a negotiation between MRA 3 and MRA 4 occurs by using intelligent negotiation agents (LNAs). LNAs are further discussed in figures 48 and 50 through 58 below.
  • analytical agents are launched by MRA 4 to MRA 2 and MRA 3 in order to analyze a specific problem.
  • Fig. 45 shows IMSA relations between MRAs. MRAs are able to communicate with each other about complex tasks simultaneously by using IMSA specialist agent roles.
  • IMSA specialist agent roles There are number of specific types of IMSAs, including analytical agents, search agents, collaboration agents and negotiation agents.
  • Figs. 46 to 48 briefly describe analytical, search and negotiation agents.
  • the process of initiating the AA begins with the generation of a search agent, which is sent to multiple MRAs' databases with an initial query (4630).
  • the search agent reports back to the initiating MRA with the priorities of data in MRA databases (4640).
  • the AA is then sent to the MRA in the order of priority sequence (4650) revealed by the search.
  • the AA analyzes the problem using specific methods (4660) detailed at 4670 including MVA, regression analysis, pattern analysis, trend analysis and hybrid analyses.
  • the AA develops solution options to the problem (4680) and shares the results with relevant MRAs (4690).
  • Search agents are described in Fig. 47.
  • An MRA generates a search agent (4710) to query distributed MRA databases (4720).
  • the search agent receives initial feedback from databases regarding initial query (4730), refines the query with specific databases (4740), evolves search parameters (4750) and seeks specific data sets among databases (4760).
  • the search agent finds data sets as a result ofthe refined search (4770) and retrieves them for the MRA.
  • the general negotiation process is described with reference to intelligent negotiation agents (LNAs) in a distributed network in Fig. 48.
  • the initiator INA meta- agent (4810) begins the process by launching initiator INA micro-agents to several other MRAs.
  • LNA micro-agent 1 is launched to a negotiation session at LNA 2's location (4820)
  • LNA micro-agent 2 is launched to a negotiation session at LNA 3's location (4825)
  • LNA micro- agent 3 is launched to a negotiation session at INA 4's location (4830).
  • Each respective negotiation session occurs at each INA's location within its MRA (2, 3 and 4, respectively).
  • the initiator INA interacts with LNAs at the various remote MRA locations (or at its home location) (4850), while a winner is determined at its home location (4855).
  • Mutual agreement is reached, in this case between INA 3 and the initiator LNA (4860), while sessions are closed between the INA 2 and INA 4 negotiations (4865) and the overall negotiation process is closed (4870).
  • Fig. 49 describes an LMSA intercommunication with messenger sub-agents.
  • the content ofthe decision is translated into specific instructions of action (4920) and the MRA creates messenger sub-agents (4930).
  • the MRA launches the messenger sub-agents to other MRAs (4940), which then deliver the message with the instructions to the MRAs (4950).
  • LNAs are used in a critical way in a distributed mobile multi-robotic system, they are further developed in figures 50 through 58, including a description ofthe INA architecture, pre-negotiation process, INA logistics, negotiation process in a distributed network, multi-lateral negotiation process, multivariate negotiation factors, winner detennination process, argumentation process and opposing LNA strategies.
  • LNAs work by negotiating between at least two MRAs. LNAs use argumentation methods to negotiate by presenting arguments with variable weights. INAs also negotiate about the best simulation to use in a specific situation. In general, LNAs use multi-lateral and multivariate negotiation in order to come to agreement between noncooperating MRAs. In the case of competitive MRAs that negotiate for a compromise, problems are solved using group problem-solving and analytical techniques. Solutions to complex MRA group problems include the optimal or a temporary choice between solution options. Group problem solving is discussed in figures 59 to 67. In all cases, Al is used in order to facilitate the negotiation and problem solving processes.
  • Fig. 50 the main INA architecture is described.
  • LNAs including an initiator LNA (5020) and INA 2 (5010), INA 3 (5015) and LNA 4 (5025) enter into a pre- negotiation session (5030), which is discussed more fully in figure 51 below.
  • all INAs negotiate in a first session between the initiator INA and the several INAs (5040), but stops negotiating with LNA 4 (5045).
  • the initiator LNA continues to negotiate with LNA 2 and LNA 3 in session two (5050), it eventually stops the negotiation process with LNA 2 (5070).
  • the initiator LNA continues to negotiate with LNA 3 in session 3 (5060) where it reaches agreement (5080) and closes the session (5090).
  • INA INA micro-agent
  • SI SI
  • S2 S2
  • Sn 5150
  • the LNAs agree on the rules of negotiations, the number of negotiation sessions and so on, based on the constraints (5170) and the initiator INA proceeds to the negotiation sessions with the other LNAs based on these pre-negotiation protocols, rules and methods (5180).
  • INA logistics are described. After initiating the session (5210), agents are generated and identified by codes (5215). The initial agent interaction protocols are generated (5220) in order for the agents to establish a common communication methodology. Such communication processes involves translation (5225) and synchronization (5230). Failure to synchronize communication leads to a termination at 5245. Once fully synchronized, LNAs may construct unique negotiation strategies using Al (5240) utilizing analytical agents (5235). At this point, agents signal the intention (5250) to negotiate with other agents. After signaling to other agents, LNAs send out communication streams (5255) to their home base, thereby constantly revealing to the home base their locations, status and plans.
  • the initiator LNA enters into a pre-negotiation session with the selected LNAs (5260) and launches micro-agents to negotiate with INAs at different locations (5265).
  • the LNAs then enter into the negotiation process (5270) and either cease negotiation (5275) or come to an agreement (5280). If they cease negotiation, the LNA settings are saved for later (5285) and the session closed. On the other hand, if there is agreement, the MRA functions are activated consistent with the agreement reached (5290).
  • Figs. 53A and 53B illustrate the negotiation process in a distributed system with mobility between INAs.
  • the present example focuses on a one-to-one negotiation between an initiator LNA and INA 2.
  • the INAs identify possible locations (5315) and specify agreed locations (5320) at which to negotiate.
  • the initiator LNA moves to LNA 2's location (5323) with program code.
  • LNA 2 identifies incoming initiator-LNA entry after activation and security protocol approval (5326) at INA 2's location.
  • the agents engage in (5330) and complete (5333) negotiation tasks, after which the initiator LNA notifies its home MRA of its remote location activities by sending a message (5336). After reviewing more tasks at the remote INA 2 location (5340), the initiator LNA either terminates (or returns home) (5343) or assesses additional tasks using internal database and analysis (5347), assessment (5350) and identification (5353) ofthe next location for task execution and moves to another location (5356).
  • the initiator LNA After moving its program code (5360), the initiator LNA identifies a need for
  • Al computation requests Al computation resources at a specified location (5367), identifies available Al computation resources (5370) and messages a request for Al computation resources to be sent to a specific location (5373).
  • the initiator INA receives (5377) and tests the Al computation resources at a specific negotiation site (5380). The negotiations are completed at the remote location (5385) and the initiator LNA returns home (5390).
  • an initiator LNA may negotiate simultaneously with at least two INAs at two or more INA locations in another embodiment.
  • Fig. 54 shows a simultaneous multi-lateral negotiation process with multiple variables.
  • INA 1 is in the position of negotiating with six LNAs, listed here as 2 through 7.
  • LNA 1 negotiates with LNA 3 in the second session.
  • LNA 1 negotiates simultaneously with LNA 2 and INA 3 on the second phase of negotiation with each.
  • LNA 1 begins to negotiate with LNA 4, while it continues to negotiate with LNA 2 in a third phase.
  • LNA 1 continues to negotiate with LNA 4 in a second phase, while it begins to negotiate with LNA 5 in a first phase.
  • the sixth session continues this approach of continuing with LNA 5 in a second phase while it initiates a negotiation with INA 6, and so on in session seven.
  • Fig. 55 shows multivariate negotiation factors in which, in the first phase,
  • MRA 1 negotiates over specific variables with MRA 2, rejecting successive possible variables until finally agreeing on, and thus selecting, "Z".
  • MRA 2 negotiates over specific variables with MRA 3 in a similar way, also resulting in the agreement over, and selection of, "Z".
  • This process of negotiating over a number of factors shows the key element of "convergence" to negotiation. By repeating this process a number of times, many INAs may agree with each other about numerous factors in a complex dynamic system.
  • Fig. 56 shows the tournament style winner determination process in a competitive LNA framework.
  • LNAs (2 through 5) enter into a negotiation with an initiator LNA (5650) in phase one.
  • the initiator INA agrees to narrow down the field to LNA 2 (5660) and INA 4 (5670) in phase two. Between these finalists, the initiator INA then selects the winner, LNA 4 (5690) in the third phase.
  • Figs. 59 through 67 describe group problem solving.
  • problems are identified by MRAs and the collective agrees to narrow the focus ofthe problem.
  • Any MRA in the group can identify a problem (5910), in sequence, such as "How to carry out a mission with other MRAs?" (5920), “How to combine with other MRAs for a common mission?" (5930), "How to target an object with a group of MRAs?” (5940) or other mission or goal based problems (5950).
  • the group of MRAs prioritize problems by assigning values to each problem and ordering them by rank in real time (5960) so that potential solutions can be made in the ranking order (5970).
  • Fig. 60 Solution options between MRAs are described in Fig. 60.
  • a shared four- dimensional grid is created by MRAs in order to represent the framework of a potential field (6010).
  • Simulation scenarios from the MRA group are tested in order to detect the best fitting solution for a specific option (6040) after analyses are performed on specific solution options by MRAs (6030).
  • a competition is then established between various potential solutions for the best solution available (6050) and weights are attached to each solution option (6060) which allows the simulation scenario solution options to be ranked (6070).
  • Fig. 61 describes the solution option selection method developed and applied by MRAs.
  • An MRA develops a benchmark of methods in order to select a simulation scenario (6110) and then applies an experimentation process to test possible solutions (6120).
  • the shortest path option is selected as a default without environmental interaction (6125).
  • the MRAs interact with the environment (6160), a process that is informed by actual environmental change (6150).
  • the MRAs receive the results ofthe environmental interaction (6170) and evaluate the results (6180).
  • Each MRA has a distinct vantage and thus applies a unique analysis (6190).
  • the MRAs prioritize the results by weighting them for probability of success and by ranking them in the order of highest probability (6130).
  • the methods of solution selection are refined (6135) and a feedback loop is structured to apply continued experimentation, when combined with continued environmental interaction, in order to continue to refine the methods of solution selection.
  • a winner is selected from the possible solution options (6140) and the optimal solution is selected for a possible scenario (6145).
  • Fig. 62 describes this process of selecting the best available, not the optimum, solution, to a problem, while waiting for the most recent relevant information.
  • the MRAs work together to establish a list of solution options (6210), which are filtered according to constraints (6220) by time, optimization, combinatorial optimization, accuracy, quality of information and by pruning out what is not probable (6230).
  • the MRAs then apply solution option methods (6240) which are refined by interaction with the environment (6250).
  • the MRAs either (1) undergo a convergence of agreement (6250), in which case they select a specific simulation scenario solution option (6275) and carry out a mission (6285), (2) partially agree with an overlap of interests within constraints (6260) and (3) temporarily agree (within constraints) (6265), in which cases they select merely the best available simulation scenario solution option (6280) and carry out a mission within these constraints (6290).
  • the MRAs may not agree at all (6270), in which event they must return back to the earlier phases ofthe process of filtering the solution options (at 6220).
  • Fig. 63 shows an illustration of MRA group agreement.
  • Fig. 64 shows the temporal aspect ofthe decision process, with the left column representing the temporal component, the second column representing the physical state ofthe multi-robotic system and the right column representing the analytical state ofthe multi-agent system.
  • past physical experiences influence past data flows, while past data flows affect future scenarios.
  • Future scenarios affect present analysis and decision-making, which influence the selection of a preferred scenario of action.
  • This section ofthe preferred scenario influences the present course of action.
  • the analytical and physical states ofthe system have causal connections over time.
  • the group problem solving process requires specific analytical methods, including multivariate analysis, regression analysis, trend analysis and pattern analysis, in order to select a successful candidate.
  • Figs. 65, 66 and 67 describe these analytical tools.
  • Fig. 65 multivariate analysis is applied to problem solving.
  • a problem is forwarded to an MNA filter (6510), which strips the variables from the problem and analyzes each variable in isolation (6520).
  • the MNA filtering process forwards the variable analysis procedure to multiple MRAs (6530) using parallel processing, where each MRA analyzes variables and compares this analysis with other MRA analyses (6540).
  • the MRAs rank the multiple variables and share with the results between the MRAs (6550).
  • the variables are evaluated in each solution option (6560) and the best available solution is selected from solution options (6570).
  • the MRAs for a winner determination.
  • the MRAs analyze a problem with a regression analysis filter (6610), sort through various variables (6620) and share the data between them (6630). Again, the MRAs divide the analysis between them in order to benefit from the advantages of parallel computation.
  • the MRAs weight the variables by establishing priorities and comparing each variable with program parameters (6645).
  • the MRAs evaluate the importance ofthe variables by comparing them with data sets in the distributed database (6650) and then rank the priorities of variables (6660) and apply the ranking ofthe problem variables to solution options (6670). MRAs select the best solution option by applying the program parameters (6680).
  • Fig. 67 pattern analysis and trend analysis are applied to problem solving of conflicting MRAs for winner determination.
  • a problem is formulated (6710) and either pattern analysis (6720 and 6730)) or trend analysis (6725 and 6735) is applied.
  • the pattern analysis approach analyses regularities in spatial coordinates using statistical methods (6740), while trend analysis analyses regularities in temporal coordinates using statistical methods (6745).
  • each analysis is evaluated (6750), the results ranked (6760) and the analyses are applied to MRA decision logic (6770).
  • the MRA group then makes a decision based on these analyses and formulates a plan (6780) that the group is able to activate (6790).
  • MRAs are mechanical entities that assume physical shape and mobility in space and time, it is possible to model them by using simulations.
  • the MRS may use a number of types of simulations, including cellular automata simulations, particle simulations and game theoretic simulations. All three main types of simulation add valuable qualities to the representation of complex activities in a mobile distributed multi-robotic system, including structuring the dynamics of aggregation processes.
  • Figures 68 through 73 describe the cellular automata simulation of MRA group activities.
  • CA Cellular automata
  • Fig. 68 shows the modeling of MRS activity with simulations in a situation assessment.
  • a cubic space is occupied by mobile agents, represented here as A, B and C.
  • the map describes the change in spatial position ofthe agents from Al to A2 to A3 (6840), from Bl to B2 to B3 (6850) and from Cl to C2 to C3 (6860).
  • Fig. 69 describes synchronizing simulations within an MRA cluster.
  • An MRA sensor detects other MRA locations (6910) and converts the analogue sensor data to a digital form (6920).
  • the MRA data about other MRA positions is analyzed in real time to show phase state changes (6930) and a simulation is constructed to represent data about MRA position changes (6940).
  • Each MRA continuously tracks all MRAs in the system in real time (6950) by using this approach and each MRA constructs a simulation to represent MRAs in the system (6960).
  • Fig. 70 describes a CA scenario option simulation. Two scenario options are presented for A and B. For scenario option A, MRA 1 (7010) and MRA 2 (7020) move across four phases to objects X and Y. For scenario option B, MRA 1 (7030) and MRA 2 (7040) move across the four phases towards objects X and Y but in a different path.
  • Fig. 71 describes a reversible, or deterministic, CA in which a simulation is constructed by projecting backwards from a goal.
  • scenario option representations look very similar to figure 70, the phasal process that is used is exactly opposite the causal approach. Rather, in this simulation model, the MRAs begin with the goal and project backwards. By using this reversible approach, the CA simulation is presented with a more goal-oriented solution.
  • Fig. 72 shows how adaptive geometric set theory is applied to an MRS.
  • the three CA models of A (7210), B (7220) and C (7230) show three different sequences from one to three reflecting different positions.
  • the converged model (7240) a combination of the three models is reached which synthesizes the three by compromising the outcomes of B and C.
  • Geometric set theory is useful to represent the overlap of aggregated sets.
  • Fig. 73 shows the selection of an optimal simulation as a (temporary) convergence of simulation scenarios.
  • MRA 1 is represented by actual positions at 1 ' , 1 " , 1 ' “ and 1"" (7320) while a possible scenario is represented by 1R", 1R'" and 1R'" (7310).
  • MRA 2 (7330) and the possible simulation scenario (7340).
  • the outcome for these sequences is a convergence of MRA 1 at 7325 and of MRA 2 at 7335.
  • Figs. 74 through 78 describe the aggregation process in a multi robotic system.
  • Figures 75 through 84 describe the dynamic coalition (or reaggregation) process in a MRA and figures 85 through 88 show autonomous MRS self-organizing processes.
  • Fig. 74 describes the aggregation initiation process in which sets of MRAs form from a larger collective.
  • the MRAs develop and present simulations (7410), test the simulations (7415), prune out the least useful simulations (7420), and compare the best simulations with the environment (7425) and with (updated) program parameters (7430).
  • the best simulations (within constraints) are selected (7435) and converged (7450) in order to create overlap. From the converged simulations, a map is created (7455) and individual MRA locations are identified relative to their positions on the map (7460).
  • the MRAs then move their physical locations in an efficient way according to the geometric location ofthe converged simulation map (7465).
  • an object X (7510) is confronted with seven similar MRAs (7520). After undergoing an aggregation initiation phase, the MRAs (7540) are shown in the second section as changing their position with regard to object X (7530) by moving towards the object. [0346] hi Fig. 76, the initiating process is shown involving common heterogeneous
  • an MRA with type "S” (7610) initiates a group of specialized MRAs (7620).
  • the "S" MRAs (7630) concentrate in order to perform a specific task, while the other types of MRAs (7640) retain their positions.
  • a particular type of specialized MRA is "picked out” in order to perform a specific function as a specialized unit.
  • Fig. 77 a complementary heterogeneous MRS group formation initiation is described.
  • the IVIRA with type "S” (7710) initiates a group of specialized MRAs in a similar was as with common heterogeneous MRAs.
  • the "Y" and "T” types (7730) from the second column which leaves the other MRAs in their stable positions (7740).
  • complementary specialists may work together as a team to perform complex functions in tandem.
  • Fig. 78 The first phase of a demand-initiated environmental adaptation is described in Fig. 78. From the combination of static environmental data maps (7830) and actual environmental changes (7820), dynamic environment data maps (7825) are created. These maps inform past and present simulations (7850), which are analyzed (7840). The analysis is itself informed by learning methods (7810). Given the simulations and their analysis, negotiations occur between the MRAs (7855), which reach a decision, within limits (7865). This decision is also informed by limited (converged) scenario simulations (7870). Once a decision is made by MRAs, the selection is made about the specific form of aggregation to use (7875) and the actual special positions ofthe MRAs are changed in accordance with this new decision (7880).
  • Figs. 79 to 84 dynamic coalitions, or re-aggregation processes, are discussed.
  • Fig. 79 the continuous MRA group composition reconfiguration process is described.
  • a group is concentrated (7910) that includes MRAs 1, 2, 3 and 4.
  • a new grouping is organized (7930) that includes MRAs 3, 4, 5 and 6.
  • yet another grouping is organized to include MRAs 4, 5, 6, 7 and 8.
  • the movement through the system from the left part ofthe group to the right part ofthe group illustrates the changing interaction response to the environment that requires the grouping to adapt to different sub-sets ofthe larger collective.
  • Fig. 80 the continuous reconfiguration of sub-networks is described.
  • the right column shows an object that the MRA group(s) on the left move towards.
  • the first phase ofthe process is identified in the right column. In the first part of the process, at 8005, the first sub-set ofthe collective moves towards the object. In a second phase, the MRAs reconstitute the configuration ofthe MRA grouping (8010) and move toward the object. In a later phase, in the middle map, a larger initial grouping, including six MRAs (8015) move toward the object, while a second grouping (8020) moves to the object later. This second group includes the overlapping two members of both groups. However, in the third part ofthe process, the demand for MRAs changes again from the second part ofthe process.
  • Fig. 81 illustrates dynamic group behavior adaptation to environmental interaction.
  • the first MRA grouping (8120) moves towards a group of objects (8130).
  • one object is knocked out, represented by an X, but two MRAs are also removed.
  • the reconstituted group of MRAs (8150) which includes the combination of 8120 and 8110, move towards two more objects in the group of objects (8160) and three MRAs are removed from action, as represented by an X's.
  • the newly reconstituted MRA group (8170) that includes a combination of 8150 and 8140, move towards the three remaining objects (8180).
  • the parallel dynamic traveling salesman problem is described with cooperating autonomous agents in Fig. 82.
  • a group of MRAs collect environmental data by sharing sensor data (8240) and use the initial prioritization of environmental data consistent with program parameters (8250).
  • program parameters 8250
  • an interaction between MRAs and the environment occurs (8820) which informs the MRA sensor data stream (8210).
  • the environmental data changes (8270) also reprioritize the order of priorities with the latest information ofa changing environment (8280); this reprioritization ofthe order of priorities are largely based on the MRAs' prioritization of a physical sequence (8260) based on a reprioritization of MRA program parameters (8230).
  • FIG. 64 illustrates a data flow process that accommodates both physical state and analytical state dynamics across time.
  • Fig. 83 shows altruistic MRAs sacrificing themselves in order to acquire sensor information to increase the chances of overall mission success.
  • the MRAs shown with an X move toward the object (8320) and are knocked out. However, the information that is obtained in this gambit mission is then sent back to the collective so that they are better able to defeat the object.
  • Fig. 84 The general dynamic coalition process is described in Fig. 84.
  • sensor data and various sources examine the terrain (8460).
  • the simultaneous parallel computation by numerous agents is performed by sharing data and by dividing computation resources (8465).
  • the sensor data is then evaluated by various MRAs (8470). Groups of MRAs begin to emerge by agreeing to aggregate (8475). Decisions are made to form smaller groups in order to meet evolving mission parameters and priorities.
  • Specified MRAs update the navigation plans and activate the mission (8485). As the mission evolves, groups of MRAs are added or removed as needed, for instance if the opposition is particularly hostile (8490).
  • Fig. 85 describes the group coordination and obstacle avoidance process that is involved in autonomous MRS self-organizing processes. Obstacles X, Y and Z (8510) move towards MRAs A, B and C (8550) from their initial positions. As the objects get closer, at 8520, the MRAs detect the objects as obstacles, at 8540, and begin to avoid them by moving out ofthe trajectory ofthe moving objects (8530).
  • Fig. 86 specific MRAs A (8610), B (8630) and C (8650) move towards specific objects X (862), Y (8640) and Z (8660), with A attacking Z, B attacking X and C attacking Y.
  • This specialization of a self-organizing process is further developed in Fig. 87 as a specialized group of MRAs work together as a team.
  • MRAs A (8710), B (8720), C (8780) and D (8790) move into positions 8730, 8740, 8760 and 8770, respectively, in a phase in the process towards assembling together at 8750.
  • the specialized MRAs work together sharing specific functions for greater usefulness on a mission, hi Fig.
  • multi-functional MRAs are described in a self-organizing process. Whereas in figure 87, the MRAs are specialized, in figure 88, the MRAs have multiple functions that may switch in specific changing circumstances. As the figure shows, MRA A in position Al (8810) and MRA B in position Bl (8850) move towards object X (8830). As they move towards the object, the MRAs detect the need to change from one specialized function to another. At positions A2 (8820) and B2 (8840), the MRAs change their functional mode to a different specialty in order to be more effective in their mission against the object.
  • Figs. 89 through 99 describe specific applications ofthe present system.
  • Remote sensing (described in Figs. 89 to 92), (2) hazard management (described in figures 93 to 95) and (3) building processes (described in Figs. 96 to 99).
  • Remote sensing activities that use an MRS include surveillance, reconnaissance, remote exploration, sentry activities and cinematography.
  • Hazard management activities include toxic site clean-up, oil spill and fire fighting activities.
  • Building processes include manufacturing production and assembly, road building and surgical activities.
  • Fig. 89 surveillance and reconnaissance is described using multiple micro objects for sensing and tracking of a mobile object.
  • X (8910) and Y (8960) move in parallel tracks to positions X2 (8920) and Y2 (8970), respectively, they track object A (8940).
  • MRA X moves to position X3 (8930) and then to position X4 (8935)
  • MRA Y moves to position Y3 (8980) and then to position Y4 (8990) by using sensors and by tracking the object closely.
  • Fig. 90 a remote exploration process is described in which the initial tracking of multiple objects is performed by multiple micro-MRAs.
  • MRAl (9010) moves towards object RI to position X'.
  • the object itself moves, from position R' to position R" and is followed by the MRA, which moves to position X".
  • MRA 2 (9020) tracking object R2 (9060)
  • MRA 3 (9030) tracking object R3 (9070).
  • Fig. 91 describes sentry activity within limited perimeters defending multiple objects with a multiple number of MRAs.
  • the MRAs are spaced evenly apart in order to occupy a constrained field around the perimeters of a field.
  • the current system is also applicable to cinematography, wherein one mobile object (or cluster of mobile objects) are sensed and tracked with MRAs. This process is described in Fig. 92.
  • MRA 1 (9210) and MRA 2 (9270) track object X (9240) as it moves to positions 9250 and 9260.
  • MRA 1 tracks the object along a path to position 9220 and 9230, while MRA 2 tracks the object along a path to position 9280 and 9290.
  • This process may be variable so that as the object stops to pause, the MRAs stop as well.
  • the MRAs have automated digital photographic capabilities with on-board auto-focus zoom lenses and data storage.
  • the MRAs can be used to track multiple objects as well. One MRA may track the object(s) in a close in view while the other MRA(s) may track the object(s) from a distance in order to obtain a different view ofthe same scene.
  • Fig. 93 describes a toxic site cleanup.
  • a static cleanup occurs within land perimeters by multiple MRAs.
  • A-type MRAs (9310) are used to confine a limited amount of toxic contamination (9320) in a specific physical space.
  • the MRAs move by using a side-to-side sweeping approach.
  • the spill (9340) has been reduced and the MRS calls in the B-type MRAs (9330) in order to continue to eliminate the contamination by using a similar sweeping technique.
  • the MRS calls in the C-type MRAs (9350) to complete the mop up operation.
  • Fig. 94 describes a dynamic clean up of an oil spill within limited hydro perimeters by multiple MRAs.
  • the oil spill (9420) is surrounded by MRAs (9410), which operate to limit the damage and remove the oil.
  • MRAs (9410) operate to limit the damage and remove the oil.
  • the oil spill is rendered smaller (9440) and MRAs (9430) continue to operate to remove the oil by operating in specific "cells” that act to sweep up the spill.
  • This process continues in the final phase in which the oil spill (9470) is confined and the final drops of oil are mopped up by the MRAs (9460).
  • Fig. 95 describes the automated fire fighting process in which dynamic interaction occurs with a complex environment by multiple MRAs.
  • MRAs (9510) are dropped to the fire (9520) on one facade only (because the fire is initially inaccessible on the other side).
  • the MRAs (9530) are able to surround the fire (9540)
  • the MRAs may be air launched or ground launched and retrieved.
  • the fire is reduced (9560) and the MRAs (9550) complete the task of extinguishing the fire.
  • Fig. 96 describes the manufacturing production process in which an object is created by using multiple MRAs.
  • MRAs A (9610), B (9640), C (9650) and D (9630) work together to create the object (9620).
  • One way to do this is for each MRA to attach parts of the object together from different spatial positions.
  • Fig. 97 shows the assembly of an object by using MRAs to combine the parts.
  • MRA A at position Al(9720) and MRA B at position Bl (9730) act to assemble objects. Rather than having a movable assembly line, in this case, the MRAs themselves move. MRA A moves to position A2 (9740) and MRA B moves to position B2 (9750) in order to complete the assembly task.
  • This process of organization of assembly tasks provides the opportunity for specialized functional MRAs to work together as a team in order to assemble objects by combining parts more quickly.
  • Roads can be built by using multiple MRAs as illustrated in Fig. 98.
  • MRA A (9810) and MRA B (9820) proceed to create a road by laying down asphalt along adjacent tracks.
  • Fig. 99 describes micro surgery using MRAs for trauma intervention and stabilization, ha this case, MRA A and MRA B guide themselves to the patient.
  • the MRAs ascertain, by using sensors, the symptoms of trauma in order to identify problems.
  • the MRAs then move to various positions on the patient in order to solve the problems.
  • the MRA will seek to stop the bleeding by cauterizing the wound with a laser or by applying pressure.
  • the MRA will administer an electric shock.

Landscapes

  • Engineering & Computer Science (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • General Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Game Theory and Decision Science (AREA)
  • Medical Informatics (AREA)
  • Aviation & Aerospace Engineering (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Automation & Control Theory (AREA)
  • Feedback Control In General (AREA)
  • Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)

Abstract

A system of self-organizing mobile robotic agents (MRAs) (6) in a multi robotic system (MRS) (2). MRAs cooperate, learn and interact with the environment. The system uses various AI technologies including genetic algorithms, genetic programming and evolving artificial neural networks to develop emergent dynamic behaviors. The collective behaviors of autonomous intelligent robotic agents are applied to numerous applications. The system uses hybrid control architectures. The system also develops dynamic coalitions of groups of autonomous MRAs (2) for formation and reformation in order to perform complex tasks.

Description

SYSTEMS, METHODS AND APPARATUS FOR ORGANIZING GROUPS OF SELF-CONFIGURABLE MOBILE ROBOTIC AGENTS IN
A MULTI-ROBOTIC SYSTEM
CROSS-REFERENCES TO RELATED APPLICATIONS [0001] The present application claims the benefit of priority under 35 U.S.C. §119 from U.S. Provisional Patent Application Serial Nos. 60/404,945 and 60/404,946, filed on August 21, 2002, the disclosures of which are hereby incorporated by reference in their entirety for all purposes.
BACKGROUND OF THE INVENTION [0002] There are several categories of prior art patents that apply to the present invention. These patents involve mainly mobile robots and groups of mobile robots. [0003] Matsuda (robot system and control device), U.S. Patent No. 5,825,981; Peless et al. (method for operating a robot), U.S. patent application publication number # 20010047231; and Nourbakhsh et al. (socially interactive autonomous robot), U.S. patent application publication number # 20020013641, mobile robots are used automatically, or with manual intervention to perform tasks such as multifunctional manufacturing, cleaning, mowing, snow blowing or interacting with humans. These pedestrian approaches to robotic control fit into the main paradigm of robotic applications.
[0004] Kawakami (mobile robot control system), U.S. Patent No. 5,652,489; Asama et al., (mobile robot sensor system), U.S. Patent No. 5,819,008; and Wallach et al.
(autonomous multi-platform robot system), U.S. Patent No. 6,374,155 involve multiple mobile robots. These patents involve using sensors for navigation and obstacle avoidance. In addition, one mobile robot can transmit information to another mobile robot for some effect. These inventions offer only rudimentary connections between robots and lack advanced system functions.
[0005] Most of the research history involving the technologies of the present system - including (1) intelligent agents and self-organizing systems, (2) Al and D-AI in coordinated systems, (3) negotiation and problem solving and (4) cooperating agents and aggregation - are represented in the academic literature, described below. [0006] The development of complexity theory is fairly recent. Theorists from economics and biology advanced the view in the 1980s that systems are self-organizing and adaptive of their environments, hi particular, biologists have studied ant and insect social organization and have observed the complex adaptive behaviors of these societies.
[0007] Researchers at the Sante Fe Institute (SFI) have developed complexity theory by looking at the fields of biology, economics, mathematics, epistemology and computer science. One ofthe aims ofthe SFI is to develop a complex self-organizing computer model representing artificial autonomous agents that emulate the biological functions of complex insect social behavior.
[0008] SFI theorists have developed the swarm intelligence model of artificial computer societies primarily for simulating economic systems. The swarm intelligence model, by emulating biological system operation, uses ideas of emergent behavior to describe the complex social interactions of relatively simple insects according to straightforward decentralized rules governing group activity.
[0009] The challenge for computer scientists lies in how to develop a system of self- organized autonomous robotic agents. The development of societies of behavior-based robotics that fuse elements of system control with elements of decentralized local control is one ofthe most difficult challenges in computer science and robotics. A key part of this problem lies in how to configure Al systems for problem solving in a MRS for collective behavior, hi short, how can we design an intelligent MRS for optimal adaptation to dynamic environments?
[0010] The computer science fields of robotics and Al have evolved in the past decade in such a way that a convergence of technologies allows an explosion in research in collective robotics and in intelligent systems in order to achieve the goals of developing an intelligent MRS for group behavior. For example, rapid advances in computation resources, communications and networking allow the combination of integrated technologies necessary for a development of a sophisticated MRS. In addition, in the area of Al research, several trends have emerged, including GA, GP, A-NN and distributed Al, that allow computer systems to not only learn but achieve some degree of autonomy.
[0011] In the early 90s, Brooks developed a decentralized modular approach to robotics at MIT's Media Lab. Revolutionary at the time because it spurned conventional wisdom of highly computation-intensive deliberative robotic control approaches, his modular approach used less than three percent of traditional computer approaches. This leap in efficiency was achieved by separating the subsystems for automatic reactive control (he called it subsumption) rather than deliberative top-down robot system control. The mobility, navigation and pick-up functions ofthe robot could be separated for increased efficiency.
[0012] By exploiting this research stream, Arkin (1998) developed a behavior-based model of robotics, hi this model, Arkin describes behavior-based robotic architectures as well as experiments in the field with sophisticated hybrid robotic architectures. An example of this hybrid approach is NASA's Atlantis system (1991) that synthesizes deliberative planning with group behavior. The aim of these models is to develop autonomous robots that are adaptive to their environment. The development of robotic teams with social behavior is one ofthe most difficult challenges, according to Arkin' s pioneer study.
[0013] Bonabeau et al. (1999), an SFI fellow, develops a research stream that connects the study of ant and insect behavior in complex biological social systems with the development of complex artificial robotic societies. In their vision of swarm intelligence, they use key notions of system self-organization, reactive behavior and environmental adaptation to point to a model for artificial robotic systems that might emulate biological systems.
[0014] In 2001, Kennedy and Eberhart focused on the social and theoretical aspects of swarm intelligence. Their examination of group behavior develops a computer model of adaptive self-organized systems, similar to economic "particle" simulations by the SFI, by emulating the social behavior of biological systems. In order to develop an artificial swarm system, the authors look to complex pattern emergence, which has a lineage from Von Neumann to Burks to Wolfram. In this research stream, cellular automata are used to simulate a complex but stable self-organizing system. Though the authors refer to research experiments with robot societies, their focus remains on computer and theoretical models of complex social behavior involving autonomous entities.
[0015] Another important research stream involves the application of Al to networks.
The emergence ofthe Internet has presented novel ways to conduct commerce automatically with autonomous software agents in a MAS. Originally developed by Smith, the contract-net protocol established an early model for distributed problem solving. As the Internet evolved rapidly, new computational systems emerged to emulate commercial systems. Solomon has developed demand-initiated self-organizing commercial systems for both intermediated and dis-intermediated transactions that employ novel multivariate and multilateral negotiation models.
[0016] One niche of the automated commerce system lies in the aggregation of autonomous agents. Precisely how to combine pools of autonomous agents for wholesale discounts presents an opportunity to remove a layer of distribution from commercial systems. This research stream is important because it provides clues as to how to develop coalitions of robotic agents for common purposes.
[0017] MRS models have been developed. The Nerd Herd is an example of an MRS using rule-based social behaviors for subsumption based foraging popularized by Brooks. Second, the Alliance architecture developed a modular approach to robot team behavior that includes inter-robot communication. Such communication allows for emergent cooperation. An additional version of Alliance (L- Alliance) accommodates the learning aspect of robotic agents in order to achieve a form of adaptation.
[0018] Arkin developed a "multiagent schema-based robotic architecture" in which team cooperation was modeled using a behavior-based approach without explicit inter-robot communication.
[0019] Dias and Stentz provide a market-based model for multirobotic coordination in which individual robots in a distributed environment negotiate with each other in order to agree upon a course of action. Such a model applies the contract-net protocol used with software agents in a distributed network to the robotics context for operation of groups of autonomous robots in dynamic environments.
[0020] Finally, Solomon developed a hybrid MRS model with military and industrial applications in which a hierarchical leader-follower approach is implemented in a hybrid central-control and behavior-based control architecture.
[0021] Most MRSs possess several common traits, including mobility, intelligence, communications, group behavior and specific functionality.
[0022] One critical aspect of robotic group behavior lies less in the value of intelligence that in the importance of methods of aggregation. It is a key challenge of robotic systems of determine ways for robotic agents to synchronize, cooperate and collaborate and, in sum, to work together as a team. The emergence of dynamic coalitions of robotic groups is one ofthe most interesting and important areas of robotic research. [0023] The effort to achieve the development of complex MRSs that may emulate, and even transcend, emergent natural self-organizing processes, has become primarily a computation challenge that involves the need to create sophisticated Al architectures. Al systems have themselves emulated biological systems, with the advent, from Holland and Koza to the present, of genetic algorithms, genetic programming and evolutionary computation methods in order to solve complex problems. A related research stream involves A-NN, which has utilized GA in order to establish weight values of neural nodes. One main aim ofthe neural networks is to develop self-configuring and self-organizing learning systems for complex problem solving. This is useful in real time collective robotics situations in which rapid adaptation to a changing environment is necessary.
[0024] The development of hybrid Al technologies that synthesize various methods for specified problem solving would provide a robust and successful option in the computer scientist's arsenal of weapons that may be useful for the development of sophisticated MRS architectures.
BRIEF SUMMARY OF THE INVENTION
[0025] The present inventions involve multi robotic systems, multi agent systems, collective robotics, artificial group behaviors, aggregation of robotic agents, coalition formation, dynamic coalitions, self-organization of robotic agents, emergent behavior of intelligent agents, cooperation of intelligent agents, multi agent learning, problem solving between conflicting intelligent agents, artificial intelligence, artificial neural networks and multi robotic operating systems.
[0026] Multi-robotic systems are complex networks that facilitate the interaction between autonomous robotic agents according to specific rules of behavior in order to perform a specific function or combination of functions. The present invention describes a system for multiple mobile robotic behavior by applying the logic of advanced computer science, in particular artificial intelligence (Al), with advanced robotic electronics and mechanics. The focus here is on artificial robotic collectives. So far very little research has been developed on the group behavior aspects of robotic societies as they plan, and then achieve, a coordinated goal.
[0027] There are several layers of any such collective robotic system, including (1) the computation, electrical and mechanical hardware of each autonomous robot unit, (2) a hardware network layer that links the individual robots together with wireless communications, (3) a metacomputing layer (that performs complex memory, database and computation analysis functions) in a node to node distributed computing model, (4) an omni- nodal artificial neural network (A-NN) layer for distributed Al, (5) an evolutionary A-NN layer - driven by genetic algorithms and genetic programming - for adaptive group learning in order to develop real-time cellular automata (CA) based simulations to seek optimal system solutions, (6) an OS layer and (7) a layer for specific functional applications.
[0028] The present invention describes a sophisticated MRS that is dynamic, interactive and evolving, adaptive to its environment and capable of exhibiting emergent behavior. The system is designed as a hybrid of behavior-based and central planning control processes in a distributed network environment. By decentralizing numerous functions in a distributed architecture model, groups of autonomous robotic agents can learn together, make group decisions together (cooperatively and competitively), negotiate and solve problems together, congregate together in various sub-sets and re-configure in non-overlapping sub- groups. Using these unique approaches, autonomous robotic agents can form and reform into various configurations of groups in a self-organized way interacting with each other and with the environment in order to achieve pre-programmed, or evolved, goal parameters.
[0029] Artificial intelligence (Al) is used in a number of MRS processes, including individual robot learning and decision making using genetic algorithms (GAs), genetic programming (GP) and other evolutionary computation (EC) approaches as well as group robotic agents that uses A-NN and hybrid evolutionary A-NN approaches (including GA, GP, FL, etc.) that provide tools for adaptive collective learning and decision making. The use of both individual agent and group learning tools are important because though the collective resources are far greater, when the system defaults to behavior-based biases, for instance, in situations with diminished computation resources, it is necessary for the individual robotic agents to have the tools to maintain autonomy. By building on the lower layers of behaviors of reactive approaches, a more complex MRS can evolve beyond ant society emulation.
[0030] In practical terms, MRS operation in unknown environments presents numerous challenges and problems to solve. In the absence of a centralized "mission control" station to solve all the problems a robotic system may encounter, there must be a number of fall back system modes in order for the mission to be successful, which leads to a hierarchy of system structures. These system modes are dependent on computation resources, communications resources, levels of robotic agent autonomy, levels of learning and levels of group behavior.
[0031] In earlier multirobotic systems, a relatively simple architecture would consist of a leader robot with various followers in a hierarchy. The leader possesses increased autonomy and orders the followers (super-drones). In this model, pre-selected squadrons are formed, the control for which can be manually intervened by human interaction processes such as a video feed for mission objective alteration. Reprogrammable orders and priorities can be uploaded at any time.
[0032] In one embodiment, supplementary external computation resources can be kept outside ofthe MRS and fed in as needed by satellite. Alternatively, though computation is performed externally to the MRS, analytical results can be used to control the system. In addition, reporting on agent behavior can be provided to an off-site blackboard so as to unify control at a central command center.
[0033] As the system and its agents gain autonomy, increased capacities are brought in, such as computation power, communication bandwidth and Al capabilities. Still, only reactive behavior-based autonomous robotic agent interaction approaches would yield a relatively simple system that appears to generate group behavior but merely mimics collective actions because ofthe outcome of interactions between autonomous agents. The system in this mode is merely semi-autonomous, which reflects its resource limits.
[0034] As the MRS system is linked together in a distributed network of autonomous robotic agents that employ powerful computation resources and Al processes, the system can automatically "think" like a group and constantly reconfigure to the best available situation while interacting with and adapting to its environment.
[0035] It is therefore valuable that the system, though using a hybrid architecture, employ a number of distinct embodiments that accommodate changes and that automatically default to the most complex task achievable.
Hybrid MRS Architecture with Distributed Resource Management and Command Structure
[0036] A pure behavior-based reactive MRS architecture has advantages of local control and emergent behavior but disadvantages ofthe inability to control large groups in complex adaptive environments. On the other hand, a central deliberative MRS control architecture has the ability to develop large self-organizing interactive systems and sub- systems but has the limits of being cumbersome and dependent on substantial computation resources. What is needed in order to build and operate a complex and high performance MRS is a hybrid architecture. In effect, the MRS architecture is a complex, continuously reconfiguring, operating system that links together robotic agents with computation, communications and software subsystems. Such a system must be modular (so that upgrades in a subsystem can be seamlessly performed), scalable (so that nodes can be added or removed) and reconfigurable. The system uses mobile software program code that provides inputs and outputs to robot machine agents. The "Harness" dynamic reconfigurable metacomputing model is a pioneer for this mobile self-organizing MRS hybrid approach because it continuously seeks to re-route the system to the optimal computation and communication pathways.
[0037] On a lower level, each robotic machine unit has sensors, actuators, microprocessors, communication receivers and transmitters, power supply, a specific functionality and (system and applications) software. However, when they are linked together, the opportunity exists for the MRS specific mobile robotic unit sensors to be organized into a network for collective data acquisition. The group's collective computation resources can analyze the sensor data. In addition, the group of mobile robotic agents can use complex Al induced learning processes to make group decisions, even in the face of noisy, error-prone and conflicting data streams. By maximizing the efficiency ofthe available group MRS resources, intelligent group behavior can emerge.
[0038] The aggregation of IVIRAs into subgroups can occur, further reconfiguring in complex ways in dynamic and changing environments. By learning and working as a group, specific autonomous agents are altruistic and may be sacrificed for the greater whole if it is necessary in a specific critical operation. Further, specific sub-groups may conflict and split the herd in order to achieve different objectives. The convergence of technologies that allows teams of autonomous MRAs to work together - computation resources and reconfiguration, communications bandwidth capacity and complex system software - make possible a revolution that emulates how groups of humans think and behave.
[0039] In order to make this technology convergence operability possible, it is necessary to develop a distinctive hybrid MRS architecture for a distributed self-organizing system. Such a hybrid system accommodates lower-level bottom-up reactive modular behavior-based approaches as well as the use of sophisticated hybrid Al resources (D-AI, A- NN, GA, GP, etc.) that work in a distributed system for group learning processes applied to complex decision processes, optimal simulation and collective robotics actions in dynamic environments. Such a hybrid model allows for adaptation in uncertain environments while also being able to carry out initial, and evolving, program objectives.
[0040] If one compares how animals work in groups we see a resemblance to our system. Though specific animals have sensory data, memory, navigation, data analysis, decision-making and action sub-system abilities, as a group collectives of animals can achieve marked performance improvements because they have more data and analytical capacities and the integration of successful actions that increase the probabilities of winning at foraging for food or defending against attacks. Why, then, cannot an MRS be developed that emulates, and even transcends, the performance ofthe animal (and insect) group model?
[0041] Historically, one ofthe main problems in building such an intelligent MRS of autonomous self-organizing MRAs has been computer resource constraints. There is the limit, not only of computer capacity, but also time, constraints. A huge amount of data must be processed in a short time while the MRS is operational; in essence, the system must compute on the fly as it gathers and understands data and decides what to do and then how to act as a group. There are practical solutions to these resource constraint problems. First, the application of Grid computing models provides an appropriate distributed model for maximizing computation capacity by sharing resources among MRAs in real-time. This model can be scalable so that new MRAs can be added as needed even if others are subtracted as the mission requires. In fact, each agent can be re-tooled and upgraded in each reuse ofthe modular system.
[0042] Second, multiple communications topologies can be used to re-route data streams to the most efficient use within the distributed system, including using advanced caching techniques for optimal collective effect. Finally, Al software can be employed for learning, negotiation, decision and simulation of complex collective behaviors. The system then determines, while it is mobile, what to do and then acts as a team to cooperatively achieve the objective. By overcoming the resource constraints with collective action, an intelligent MRS emerges. [0043] The present system is therefore far more advanced than previous MAS approaches that seek to emulate the behavior of groups of simple insects because our system is endowed, not only with autonomous agent intelligence, but with collective group intelligence that transcends simple group behaviors. It is clear, then, that in order to develop such an advanced MRS, hybrid or meta-architectures must be employed that combine both local and global aspects.
Towards a Hybrid MRS Al Model: Distributed Problem Solving, Integrated Group Learning, Decision Processes and Dynamic Optimization Simulations with Cellular Automata
[0044] Al has emerged in the past generation as a valuable tool for solving complex problems. Genetic algorithms, developed by Holland and others, are a problem solving method to evolve, through reproduction, crossover and mutation techniques, algorithms. Genetic programming and other evolutionary computation approaches seek to solve different domains of problems. These complex strategies seek to emulate natural evolution processes so as to find the fittest, most efficient or optimal solutions.
[0045] The development of artificial neural networks (A-NN) was initially intended to emulate brain function. Referred to as connectionism, A-NN uses GA and FL (soft computing) techniques to map out, train and reconfigure a network of nodes for solving problems. By using an adaptive network architecture topology, the A-NN system can optimize adaptation to its environment. By training the network over distributed groups of agent nodes, the A-NN can learn. Evolutionary A-NN (E-A-NN), or neurevolution, is useful for reinforcement learning. A-NN's work by using genetic algorithms to adapt input features, learning rules and connection weights. One ofthe most effective applications for A-NN is nonlinear statistical models such as pattern recognition. A-NN's leam by altering synaptic weights; synaptic weight variables change by using fuzzy logic techniques to assess probabilities and thresholds. Bayesian networks use hypotheses as intermediaries between data and predictions to make probability-based estimates of solutions. Hopfield networks are used to remember an earlier network configuration and to revert to an old network when noisy data limits continuing network development.
[0046] The present invention uses a hybrid approach to Al that combines GA and GP with A-NN and D-AI architectures. The combination of evolutionary computation approaches with distributed neurocomputing models produces a system that constantly rewires itself as the system is reconfigured. This approach is necessary because finite computation resources need to be maximized even while the distributed mobile MRS changes. Not only is this scheme scalable but increased computation capacity can be provided on demand if needed by specific under used MRAs. Such a hybrid Al architecture is best suited for learning by groups in a distributed network as well as for optimal adaptation to dynamic environments.
[0047] Hybrid Al approaches can be useful when solving complex problems. Two main problem solving models involve either cooperative (altruistic) or conflict (self- interested) oriented agent behavior. One main computational challenge that involves MRS is the distributed problem solving that requires negotiation among conflicting autonomous agents.
[0048] Conflicting MRAs use Al approaches to negotiate a settlement so as to solve complex multilateral disagreements. One way for groups to solve problems in a conflicted MRS is by finding proper matches for shared common interests, thereby focusing on the limited remaining variables and disagreements. This pruning process can settle an issue either by pre-determined (or changing) rules or by a vote between involved agents. In this way teams of MRAs can compete for effective solutions. Another method to find solutions in conflicted MAS situations is to set up a competition for the strongest strategies according to agreed upon rules. Finally, an agent can persuade other agents to its position.
[0049] All of these models involve inter-agent collaboration for complex group problem solving. The resolution of competing rival MRAs conflicts result in agreement about an optimal solution. Through conflict and competition, not only is common ground sought, but a winning algorithm solution is determined for complex problems. This problem solving negotiation approach is useful for organizing heterogeneous MRAs for common objectives.
[0050] How are negotiations between MRAs in a MRS processed? Autonomous robotic agents use complex decision processes that ultimately affect group behavior. Decisions can be made by either individual agents or by groups of agents. Rules are used to prioritize specific possible choices over others. Upon achievement of a specific threshold, a decision process yields a resulting choice of possible options. Once a threshold is achieved, a plan of action can be implemented.
[0051] Since it is important to configure group decision processes for MRS problem- solving, a range of decision choice constraints present the lower and upper bounds of potential optimal solutions. Further, these parameters are constantly shifting in dynamic environments. Hence, methods need to be devised to find the shortest path to perform specific tasks. One way to do this is to perform specific tasks. One way to do this is to use statistical weighting to prioritize problems and solution possibilities. In the context of complex changing environments, an MRS must simultaneously work on solving numerous constantly changing problems. The Markov decision process makes decisions by prioritizing possible choice as measured by evolving values criteria.
[0052] MRS action starts with a plan. By mapping the parameters of group action plans, we can model the optimal configuration or allocate the most efficient resources. Decision logic processes lead to identifying trade offs (parameters) between possible solutions that lead to an optimal problem solving choice. MRAs use computation optimization techniques to select optimal solutions to complex problems in uncertain environments. By mapping various scenarios, using Al and decision processes in a distributed network, MRAs select the best plan to achieve objectives.
[0053] MRSs use advanced hybrid Al methods in order to achieve optimal grouping patterns of behavior. Unlike purely computational MASs, a MRS have physical dimension and motion in space. These physical and geometric realities about the practical operations of MRSs involve the need to organize spatial interactions and movements. It is useful to model these MRA movements before actually performing specific maneuvers primarily through the use of simulations.
[0054] Cellular automata (CA) models provide an important tool to simulate the changing movements of MRAs in an MRS. By using Al approaches, each robot is represented as a cell in a larger system. Cells can interact with neighbor cells in the neighborhood ofa CA system, with two dimensional, three dimensional or four dimensional models representing the change in cellular states.
[0055] The results of combinatorial optimization approaches to seek the best solutions to solve problems can be represented by CA simulations and, thereby, tested, before actually implementing these decision choices. By modeling group behavior in real time, the MRS solves problems and can seek improved solutions that can capture subtle contingencies in complex operational situations. MAS swarms are tested in particle simulations using CA models, but MRSs have not applied these important CA driven simulations for real geometric behaviors. Therefore, the present invention uses simulations in a dynamic, rather than merely static, way, for real time testing. In the simulation, virtual robots are provided the valuable advantage of trial and error of potentialities of activity so as to learn from complex contingencies, in order to optimize the chances for mission success. Swarms, Flocks, MRS Aggregation and the Formation and Reconfiguration of Dynamic Coalitions
[0056] Nature provides analogies for computer scientists in the contexts of Al and group robotics. In the case of Al, GAs and GPs seek to emulate natural selection by breeding the best fit problem-solving programs using principles of sexual reproduction, pruning and random mutation. In the case of robotic group behavior, scientists have sought to emulate insect (ant and bee) social behaviors in order to understand how compolex patterns emerge from simple individuals. How can MRSs be developed that have the self-organizing properties of insects? The two main behaviors that have intrigued observers are foraging (food location search and procurement) and swarming behaviors.
[0057] Scientists have discovered that ants use pheromones (chemicals that have an odor to attract others) to develop complex foraging behaviors. By laying down pheromones, which, though temporary, can be increasingly intense if compounded, ants provide a natural reinforcement mechanism (stigmation) with positive feedback. This positive reinforcement learning mechanism suggests a self-organizing system.
[0058] There are other ways for insects to communicate with each other. Some ants and bees have developed ways of communicating with their nearest neighbors about food sources, for instance, to get help with or altruistically share information with the group. This nearest neighbor communication approach, which is primarily sense based, is key to the formation of flocking, herding and schooling behaviors in animals and fish.
[0059] In the case of bees or ants, there may be specialists that perform specific functions in the hive or nest in order for the whole organization to function more smoothly. This division of labor has evolved for millions of years as an efficient social system.
[0060] Insects may communicate with each other indirectly. The process of stigmation operates with an insect affecting, or changing, the environment, which then catalyzes other insect behavior. The use of pheromones illustrates this process because the ants lay down an attracting chemical that may be acted upon by others in a limited time.
[0061] Animal and insect group behaviors emerge at the local level. Though insects are not intelligent in some ways, their complex group behaviors suggest that they have evolved social intelligence. By working in groups, they have defended against predators and survived in hostile environments. But here are limits to this kind of swarm intelligence. [0062] Though they have an initiator, most swarm or flocking behaviors do not have a single persistent leader. Instead, such social behaviors focus on local and reactive interactions.
[0063] Flocking is a case in point. Each bird in a flock has limited information about flockmates. Instead, they have neighbors they provide local information on direction and speed. The big challenge is to avoid collision with neighbors even as they signal trajectory and velocity data through their behavior. Consequently, both attractive and repulsive forces are involving in flocking behaviors.
[0064] There are, then, simple flocking rules that are useful to MRS designers because they illustrate local reactive behaviors: (A) Fly at a steady state speed of neighbors, (B) Tend to the center ofthe flock and (C) Avoid collisions with neighbors. This is similar to driving on highways because we have limited information (visibility) restricted primarily to our nearest neighbors, with which we seek to avoid collision but also maintain a consistent pace. Flocking, like herding, school and swarm behaviors, have evolved to allow groups of insects, birds, fish or animals to move in a hostile environment while avoiding peripheral members from being picked off by predators. In nation, then, avoidance of obstacles, neighbors and predators has become integrated into the rules of survival that social group behaviors maintain.
[0065] How does a swarm form? An event will stimulate an individual insect to attract neighbors to the swarm activity. Though any individual can be a leader that initiates action, the recruitment of other individuals through attracting the cooperation of similarly interested neighbors is key to the process because these individuals then respond by attracting more neighbors, and so on. Thus, any individual can initiate a swarm or flock; this initiation is a sort of initial request to procure resources for a specific (defensive or offensive) function or activity. Rather than a centralized mission control issuing orders to the troops, specific decentralized individuals can trigger group activity in a sort of local reactive chain reaction process that has the effect of overwhelming an enemy. In some cases, specialists alone, such as soldier ants, may swarm for an attack process.
[0066] Insect and animal social behaviors are important to understanding complex social processes involving simple individuals. Attempts have been made to emulate biological system swarm intelligence for development of artificial systems of robots. For instance, Arkin's (1998) use of Brooks' simple modular reactive robot for group behavior shows an attempt to model complex behaviors from simple robots.
Beyond Flocking: MRS Aggregation and the Formation and Reconfiguration of Dynamic Coalitions
[0067] The present invention goes beyond these interesting biological emulations.
Because our system is layered, with simpler default modes of operation, we will use simple swarm behaviors in an MRS that employ reactive local interactions. But the main objective is development of complex aggregated MRS systems that are capable of intelligent social behavior as well as the operation of dynamic coalitions. Whereas the simpler group behaviors have severely limited computation and conimumcation resources in a homogeneous system, the present invention does not. Simple swarm behaviors have anonymous homogeneous simple members (in uniform roles) with primitive local communication, minimal computation capacity and the limits of reactive behaviors using a narrow set of rules for learning and action. The limits of this biologically inspired system can be improved by development of an advanced MRS that exhibits social intelligence. Our system has autonomous individual MRAs with highly advanced computation, Al and communications capabilities, complex learning and simulation functions, specialization features and team behaviors in a heterogeneous system. In short, the present invention emulates human social behavior by using artificially thinking mobile robotic agents for a range of functions.
[0068] The problem of how to aggregate objects is an important one in computer science. Methods of aggregation involve collecting together disjoint sets for an organized assembly. Combinatorial optimization is the mathematical field concerned with seeking solutions to aggregation problems. Aggregation is useful for mass pooling of customers with common interests for wholesale discounts. Similarly, combinatorial auctions are a useful commercial structure to enable parties to acquire bundles of items for optimum benefit.
[0069] For the purposes ofthe present invention, aggregation is important as a process for organizing groups of MRAs within an MRS. We are not only interested in how groups of intelligent robotic agents form, but also the process by which groups break into subgroups and reform. Intelligent aggregation of MRAs involves automatic selection, formation, combination, reformation and dissipation of groups. Each new set of intelligent agents represents a new configuration. Emergent behavior ofthe MRS leads to a complex self-organizing system that never settles on an equilibrium because it is constantly changing. Finally, unlike other pure computational contexts for aggregation, the application in an MRS involves the geometry of space and extension and the physics and mechanics of motion.
[0070] The autonomy of intelligent agents leads to the opportunity for individual specialization. Whether in biological or economic systems, specialization affords the optimization of teams because it establishes an efficiency enhancing division of labor. Groups of MRA specialists can work together in an artificial system for increased benefit to the objectives ofthe whole group. The existence of specialization also makes possible the interactions of sets of agents.
[0071] Aggregation is a process of grouping entities together. One useful way to model groups is with game theory. As applied to an MRS, game theoretic models have a geometric dimension. Game theoretic approaches to modeling an MRS is useful particularly because they can be multi-phasal and interactive. Not only are MRA interactions nicely modeled but complex interactions between sub-groups can be more optimally represented as well as interactions with the environment. Game theory can model cooperating agent behavior as well as conflicting or heterogeneous behaviors. An example of a heuristic for MRS game theoretic modeling parallels chess playing maneuvers, with openings, gambits and traps providing MRA models for the inter-operation of artificial societies. Robotic agents work together to develop winning game strategies for achieving goals or solving problems.
[0072] One ofthe aims ofthe present invention is to develop methods for MRAs to constantly develop shifting groups. We are interested in discovering how intelligent autonomous robotic agents form and reform into dynamic coalitions of collectives. Understanding precisely how sub-groups of MRAs organize, self-configure, reconstitute, adapt to their environment and regroup is the key to understanding complex emergent group behavior in intelligent self-organizing systems.
[0073] With severe resource constraints, mobile agents will tend to behave according to simple rules inspired from biological systems, with local and reactive control. But endowed with sufficient computation and communications resources, an intelligent MRS will be able to perform more effectively. One ofthe areas of improvement in the operation of groups of MRSs lies in establishing methods and processes for dynamic coalition behavior. [0074] Multiple squads containing specialized MRAs can work together by sharing sensor data, data analysis, computation, communications and decision processes. Such multiple squads can form alliances and temporary coalitions for specific missions with numerous applications to industry. When group resources are restricted, specific squads can operate autonomously with limited information and still perform its objectives. Further, higher priority squads can get more resources at crucial times. Squads can reconstitute by taking resources from the larger group for continuous dynamic coalition reformation so as to more optimally adapt to changing environments. The existence of multiple micro-coalitions can be better suited to satisfying multiple goals simultaneously and thus increase the chances of a mission success.
[0075] Squads of MRAs break off from larger groups in an MRS. The squads can share the larger computation, communication and sensor resources and decision processes of the larger group. In effect, the squads operate as teams of nodes in a neural net that constantly reconfigures on the fly. Since some ofthe sensors in some ofthe squads are exogenous to each team, the squads have access to data streams beyond any limited team. Sub-teams are synchronized into the distributed network using hybrid Al approaches. Nevertheless, each squad, and its reconfiguring teammates, can work independently with local behaviors. In addition, differences between agents in a squad, for example, specialists or different "personalities," can create complexity in squad behavior within the practical constraints of their programming, as they inter-relate in different configurations. Squads self- select into various coalition configurations, but during complex missions, new squads can pick up stragglers from previous damaged squads. Similarly, squads can merge in instances where combined strength is needed to solve a problem. Robotic agent nodes can be added or subtracted as the system continuously reconfigures to achieve optimum success.
[0076] Different methods are employed in order to realize group MRA self- organization processes. In one important sense, game-theoretic and cellular automata simulations are useful in order for collectives in an MRS to map out and achieve complex plans for problem solving. By employing these processes within Al driven computation, intelligent MRAs work together to optimize complex processes in order to achieve mission success. The opportunity to simulate these processes of constant re-grouping for dynamic coalitions of MRAs allows a new generation of applications of MRS social behaviors to be possible. In this way, among others, the present system far surpasses prior approaches to emulating biological social behaviors. Our system allows intelligent MRAs to constantly shift in dynamic coalitions that are best suited for environmental interaction. It is precisely the continuously changing environment that requires development of a complex system that makes possible continuous reorganization.
Innovations and Advantages ofthe Present System
[0077] The present system has a number of innovations and advantages over earlier inventions. These innovations involve (1) multi-robotic system architecture, (2) computation resource structure, dynamics and allocation, (3) Al dynamics, (4) group negotiation, learning and decision structures and processes, (5) intelligent social behavior involving mobile robots and (6) dynamic coalitions of MRAs.
[0078] The present invention utilizes a novel hybrid MRS architecture that dynamically adjusts from manual operation of groups of MRAs to wholly automated socially intelligent MRAs in order to accommodate severe resource restrictions as well as extremely complex behaviors. By defaulting to the most complex appropriate resource level, the system optimally adjusts to environmental conditions. For instance, very small MRAs may be resource constrained and would thereby employ simpler local reactive behavioral rules. The architecture ofthe present system is also both modular and scalable so that growth or shrinkage will not affect performance.
[0079] The present system uses a distributed wireless grid supercomputing model.
This approach allows the sharing of computation resources, including memory, database storage and data analysis capacity, thereby far extending previous constraints, hi addition, this distributed model is optimal for equal node parallel processing within a collective. Computation processing speeds of dozens of teraops could be maintained in this system, thereby providing ample resources for complex group behaviors. The present system also uses advanced routing procedures to maximize the most efficient geodesic heuristics.
[0080] The present system employs a novel use of a MAS within a MRS in order to communicate, negotiate, control and organize group behaviors. Intelligent mobile software agents (LMSAs) are the analytical representatives that perform critical internal functions in the robotic system. In addition, intelligent negotiation agents (INAs) represent a core and innovative aspect ofthe present system as a vehicle for MRAs to interact and solve problems.
[0081] The present invention uses a dynamic reconfigurable evolutionary A-NN that provides optimal adaptation to the changing environments of an intelligent MRS. The A-NN uses hybrid Al techniques, including combinations of GA, GP, FL and EC. As nodes are added or subtracted to the network, the A-NN is automatically rewired for maximum efficiency. The system uses feedback loops to learn. The A-NN is useful to train the system in group learning processes. These applications to a mobile and dynamic MRS are novel. The use of comiectionism (neural nets) in a MAS and a MRS is a huge leap from earlier systems.
[0082] In order for the present system to learn, it employs FL processes that use probabilities to make group decisions by selecting the best available option among a range of contestant options. The system utilizes combinatorial optimization approaches to select the best solution to solve problems. Particularly in conflicting situations between agents, there is a need to negotiate a settlement by developing a method of winner determination. The system employs novel approaches to asymmetric problem-solving by using multi-lateral negotiation methods.
[0083] The present invention uses game theoretic approaches and cellular automata schemas in order to simulate tactical system opportunities for an MRS in novel ways. By using real-time CA and GT simulations, an MRS can automatically select an optimal problem-solving path and, hence, model complex interaction dynamics among MRAs and between MRAs and the environment. Given limited information in challenging environments with resource constraints, the use of simulation modeling for action planning and contingency scenario testing is necessary to achieve highly intelligent MRS behavior.
[0084] The present system is novel because it is heterogeneous. The MRS employs specialty robots for diverse functions. Some MRAs may have multiple functions, alternative functions or work in teams with complementary functions. This approach increases efficiency of task execution because it promotes an automated division of labor in an MRS.
[0085] Despite their specialty functions, any agent can initiate group behaviors. The attraction of MRAs to collectives can be demand-initiated in a novel implementation of group behavior in an MRS. This approach enhances system performance. In one implementation, stronger data inputs may constitute invitations to act beyond a specific threshold and thereby initiate MRA grouping behaviors. The present system uses novel group attraction initiation methods. [0086] The present system synthesizes local control with deliberative planning. This hybrid architecture is novel and is possible only with the unique convergence of advanced computation technologies disclosed herein.
[0087] The present system uses novel approaches to dynamic coalition formation. Using these approaches, the MRS constantly reconfigures its structure and dynamics in order to adapt to environmental changes. This more effective adaptation provides increased speed, precision, efficiency and effectiveness in mission critical situations.
[0088] By applying distributed artificial intelligence approaches, the present system develops a way for groups of robotic agents to make decisions in cooperative and in conflicting situations in real time. This is a novel and important advance over earlier systems.
[0089] The present system implements novel MRS approaches involving tactical cooperating teams of MRAs. This sophisticated use ofthe system transcends earlier notions of artificial group intelligence.
[0090] Why are groups of robots important? Traditionally, robot groups allow an increased speed to do a task. Like in nature, groups are increasingly reliable since some may fail but the group still finishes the task. In addition, using robot groups to perform tasks can be more flexible than only individual robots. The present system offers higher performance benchmarks for these traditional advantages.
[0091] Since the present system uses multiple hybrid architectures, at the system and
Al levels, there are nontrivial advantages over earlier systems.
[0092] The present system most efficiently implements complex group behavior in an artificial robotic system. For example, unlike earlier artificial systems that seek to emulate insect behaviors, the present invention seeks to emulate, and transcend, complex human group judgment to develop a true social intelligence. Consequently, the present invention goes beyond robotic systems that focus primarily on local control ofthe nearest neighbor and reactive behaviors.
[0093] The use of coordinated, cooperative and reconfiguring squads in dynamic coalitions in the present system provides numerous novel and useful advantages. [0094] Finally, the present system is useful for a broad range of important applications, from manufacturing to toxic clean-up and from remote exploration to traffic coordination. The sheer breadth of collective robotic applications, to industry and beyond, using the present system suggests a range of uses that could provide revolutionary implications.
Applications ofthe Present Invention
[0095] There are numerous applications ofthe present system. Robots can have specific functions for specialized purposes. One robot can clean, while another can dry. But specialized robots can have particularly high utility as they function in teams. While specific purpose robots are useful, multiple function robots are increasingly productive. Multiple function robots can switch roles or change forms as needed to complete complex tasks. The more tasks a robot can do because of its multiple specialties, the more plasticity and flexibility it has.
[0096] Multi-functional teams of robots can perform more tasks than specific specialty robots. The more tasks that robots can do, the more plasticity of tasks a team of robots can perform because ofthe efficiency benefits ofthe maximized division of labor.
[0097] The following is an extensive (but not exhaustive) list of applications of groups of robots that the present invention advances.
Manufacturing
[0098] The present system enhances factory production, assembly and distribution processes. Methods for groups of robots to work together may greatly accelerate production techniques. For instance, by using groups of multi-functional autonomous robots, a host products can be produced faster, more efficiently and cheaper than with earlier methods.
[0099] Regarding the factory assembly process, the novel use of groups of autonomous mobile robots may reshape the very idea of an assembly line because new interactive processes, reflecting an efficient modular workspace, will reconfigure approaches to activities in which parts are combined to a whole. The application of self-organized groups of multifunctional robotic systems to manufacturing assembly can promote just-in- time production processes and lean inventory to save time and increase efficiency. [0100] The distribution function of factories, such as loading and unloading, can be improved with teams of autonomous robots working together. Such a system can replace routine labor practices.
Construction and Repair of Structures and Roads
[0101] Self-organizing teams of autonomous robots can build and repair roads and structures. From laying track or pipe to electrical, plumbing, framing and roofing, an MRS can be useful in performing laborious time-intensive routine structure building construction functions. Similarly, MRAs can be useful in the repair of buildings and streets. These novel MRA processes can save time and reduce costs of building construction as well as road work and repair. In one practical application, pot holes can be automatically detected and repaired by teams of MRAs.
Medical Applications: Medi-bots
[0102] There are two categories of application ofthe present invention to the medical field. First, groups of medical robots (medi-bots) can be used in critical field situations to stabilize a patient. Autonomous medi-bots work together to (a) diagnose a patient's trauma, (b) resuscitate, via electronic pulse or CPR, a patient whose cardiac or pulmonary functions have ceased, (c) cauterize wounds to stop (or minimize) bleeding, (d) apply an IN for intravenous solution transmission in order to replace vital fluids and (e) call for more medical resources by providing a precise physical location position. Multiple medi-bots can much more efficiently rescue and stabilize patients, thereby saving lives.
[0103] Second, groups of medi-bot can assist doctors in clinical situations by performing functions typically attributed to nurses and assistants. Such medi-bots can monitor patient functions during procedures as well as actively support the surgeon or dentist so as to save time. These medi-bots can also supply expertise in critical operating room environments. In critical emergency room situations, where time and precision can make a difference, medi-bots can save lives.
Reconnaissance and Surveillance
[0104] A big category of use ofthe present system lies in reconnaissance and surveillance. Multiple autonomous robots working as a team are optimal for reconnaissance and surveillance activities. These MRAs can transmit real-time vision and sound to off-site locations, typically via satellites or terrestrial communications systems. [0105] In one mode, the MRAs can be very small micro robots (more fully referenced below) that provide stealth advantages for reconnaissance and surveillance purposes.
[0106] In other embodiments, MRAs can be disguised as natural phenomena, such as animals, birds, insects, etc. for evasive and stealthy advantages. By emulating natural animal behaviors, mission effectiveness can be maximized.
[0107] Finally, by using groups of MRAs, a more complete, more dynamic and more accurate view ofthe terrain being viewed can be maintained than with any other existing technology.
[0108] If captured, an MRA in this system can erase its programming and be rendered a useless pile of sensors, while the remaining network nodes automatically reconfigure for effective performance.
Search and Rescue
[0109] Including reconnaissance MRAs and medi-bots, teams of robotic agents can conduct search and rescue operations in difficult terrain that may be inhospitable to humans, such as in extreme weather.
Toxic Clean-ups
[0110] Groups of MRAs can be used to perform complex clean-up operations that may be hazardous to humans. These clean-up categories include: (a) toxic waste dumps, (b) nuclear reactor cleaning, (c) oil spill events and (d) sewer cleaning.
[0111] MRAs can use self-organizing maps of a local terrain to devise plans to most efficiently and safely provide toxic clean-up operations, thereby saving lives and protecting the environment.
Fire-Fighting
[0112] Using similar configurations and methods as used in toxic clean-up applications of MRAs, an MRS can be used to fight fires. Ground MRAs can dig trenches and plot trajectories for the expanding fire territory, while aerial MRAs can drop fire retardant at tactical locations for optimal effect. As with toxic clean-ups, MRAs use self- organizing mapping processes to assess the scope and dynamics ofthe full-motion fire situation. Fire-fighting MRAs can save lives and protect property. This application can be useful for forest fires, urban fires or industrial structure fires that require complex problem solving and decisive action. Medi-bots can be used in conjunction with these fire fighting applications for maximum benefits.
Mining
[0113] MRA teams can be very useful for mining minerals in remote locations. Robots can identify the most promising locations to dig and then help with laborious digging and sifting tasks. Groups of MRAs can work faster and more efficiently than current automation processes, in part because they are mobile, autonomous and self-organizing.
Agriculture
[0114] Farming has enjoyed increased automation processes for generations so as to maximize production. Groups of MRAs can continue this automation evolution, particularly in the planting and harvesting contexts in which greater care is required for specific crops such as fruit and vegetables. In general, MRAs replace the routine functions of migrant pickers.
Ship Hazards
[0115] Like toxic clean-ups, ships have a number of complex and dangerous problems to solve. Because ships function as self-enclosed physical domains, MRAs can operate effectively on specific problems. Groups of MRAs can provide effective automated solutions to hazardous functions, thereby reducing risks and saving time and money.
Clearing Minefields
[0116] One main activity for MRAs involves demining. Groups of autonomous robots can work together to either dis-assemble or explode mines that are discovered in a self-organized search process. In addition, disarming bombs can be a useful function for groups of MRAs.
Traffic Coordination
[0117] Groups of automated vehicles can use the present system for effective operation. MRS vehicle categories may include cars, trucks, trains, aircraft and ships. In particular, cargo may be moved on various groups of autonomous vehicles for greater efficiency, timeliness and cost-benefit. Such traffic coordination systems may develop complex routing algorithms that emulate, and transcend, bird flocking or ant foraging behaviors. Elevator and Dam System Regulation
[0118] Systems of elevators can be better guided and coordinated by using autonomous group logic. Similarly, dams can be regulated more efficiently by using group logic processes of an MRS.
Weather Prediction
[0119] The present system is useful to organize groups of weather balloons or aircraft to gather and disseminate data. The MRS is ideally suited to complex adaptive environments such as detecting dangerous weather conditions such as tornados or hurricanes. Groups of self-organizing MRAs can more rapidly predict dramatic weather system changes.
[0120] In an active mode, MRAs can not only predict poor weather but can influence its outcome. In a drought situation, MRAs can seed clouds to increase the likelihood of inducing rain. In an extreme case, MRAs can prevent tornadoes by influencing their movement very early in their development and changing the immediate environmental conditions. Only self-organizing groups of automated mobile robotic agents with specific functions - such as warming cool air in limited areas so as to retard or minimize a turbulent cyclic force — could execute this precisely or rapidly.
Satellites
[0121] Groups of satellites can work together to perform distinctive functions such as optimally tracking moving objects by using the present system.
[0122] The present system can also be used to have groups of self-organized autonomous MRAs repair or readjust a satellite remotely.
Underwater Applications
[0123] As with other remote domains, the present system can be used in underwater applications. Specifically, the underwater context can be used with other applications, including surveillance, reconnaissance, search and rescuer and demining.
Remote and Space Exploration
[0124] The use ofthe present system for space and remote exploration is logical. By using teams of self-organizing MRAs, complex exploration activities can be routinely performed. This technology can be applied to underwater, extreme cold or deep space missions which are optimized for the flexibility and efficiency ofthe group behavior of mobile robotic vehicles. These vehicles can have multiple functions for the collection and analysis of local environmental data. In some situations, these MRAs can conduct covert operations during which they may need evasive programming capabilities.
Sentry Protective Services
[0125] Groups of MRAs can be used as an automated system of sentries for security protection purposes. Sentries can be used not only for surveillance but also for defensive uses in order to protect structures or personnel. Such MRA sentries detect and respond to invasive action by unauthorized personnel by tracking and evading the intruders and calling for assistance. In a more aggressive mode, automated sentries can respond to invasive behaviors by disarming and subduing unauthorized activities until the authorities can arrive.
Cinematography
[0126] The present system can be used by groups of MRAs that operate video or film cameras in order to capture dynamic movie scenes. Because the MRAs can be constantly moving and can be both self-organizing and synchronized, an MRA can facilitate a new generation of film-making techniques, particularly for the popular action sequences. While moving in synchronized or random ways, MRAs are well suited to capture moving scenes in distinctive cinemagraphic ways only possible in an MRS.
Commercial Laundry or Restaurant
[0127] Routine restaurant food preparation and delivery and commercial laundry functions can be done by teams of MRAs. Working as a group of short order cooks, MRAs can produce more variety of recipes in a shorter time than professional chefs or waitresses. Similarly, a commercial laundry service can be optimized by using groups of MRAs to organize, clean and package clothes. One hour discount cleaning is now possible by using an MRS.
Micro-Robotics
[0128] One ofthe most exciting developments in robotics is the advent of small (fly- sized) robots. But the smaller the robot the greater utility is derived from working in groups. Once in groups, micro-robots can become very useful much as ants or bees are successful in groups. A number of group robotic applications involve the use of micro-robots. Given resource constraints of micro-robots, the group gains massive resource benefits while operating in a network using the present invention that make possible dramatic performance gains over merely a collection of unlinked autonomous robotic agents. These MRS micro- robotic networks could also be construed as very small scale integrated systems (VSSIS).
[0129] Generally, the smaller the micro-robotic agent, the simpler the system. Hence, some straightforward applications include surveillance and reconnaissance in which sensor data is transmitted for central use while the system is camouflaged as a natural phenomena (such as a real fly or spider).
[0130] Teams of self-organizing micro-robots utilizing local reactive operational behaviors can use traditional computer based group behavior that emulates biological system behaviors such as foraging or flocking. But the present system strives to go beyond these restrictive behaviors.
[0131] In one embodiment, disaggregated collectives of micro MRAs can form together into a larger composite robot exhibiting unified behavior. This is important so as to allow larger robots to disassemble into constituent (specialized) parts if necessary in order to evade a predator or disguise a maneuver.
[0132] hi another embodiment, micro MRAs using the present system could inspect and assemble micro-electronic systems or could inspect biological entities for abnormalities.
[0133] As microprocessors progress to ever smaller sizes and greater capabilities, the practical uses and possibilities for micro-robotics, particularly in self-organizing groups, increases dramatically.
Nanotechnolo gy
[0134] A close relative of micro-robotics is nanotechnology. The use by nanorobots and nanoprobes ofthe present system is logical. The same argument and restrictions of microrobotics apply to molecular sized nanorobotics. Like microrobots, nanorobots can assemble into larger composites that themselves work together as autonomous groups.
[0135] Uses of nanorobots include surveillance and reconnaissance. But more fanciful uses include biological applications that include cleaning arteries by injecting a group of nano-MRAs into a patient's blood stream. The nanorobots will go to the affected area, perform the operation internally and regroup for extraction. Nanorobots could also be used to identify and repair microelectronic abnormalities.
Expert Systems
[0136] Groups of anthropological MRA "androids" can work together to form complex expert systems. Operating as consultants with autonomous opinions, robot expert groups can behave like specialist teammates to collect and analyze data, perform forecasting, develop alternative scenarios, make predictions and give advice in the form of reports. Such groups of expert consulting opinions can involve numerous substantive industry categories and topics, including optimal telecom and energy routing algorithms and economic, business industry and scientific analyses. The personalities, experience and learning processes ofthe android MRAs evolve. Taken together, such expert systems constitute a think tank. Ultimately, such a group of autonomous self-organizing robotic agents can form and reform coalitions of specialist experts similar to a sophisticated consulting firm. By applying evolutionary learning and combining various opinions, such complex systems can be creative and capable of original thinking approaches that far surpass chess playing supercomputers.
MRAs
[0137] MRAs can take numerous forms. Since there are numerous applications ofthe present system in divergent industrial and technical contexts, it is appropriate to identify the structure and function ofthe variety of MRAs that can perform various jobs.
[0138] MRA vehicles can include various forms of aircraft, such as airplane, glider, helicopter, balloon, blimp, satellite or spacecraft. MRAs can operate in water as ships, boats, submarines or hovercraft. On land, MRAs can be automobiles, trucks, farm equipment, mining equipment, factory equipment, etc. There may be entirely new forms of MRAs as well, such as remote exploration devices, anthropological androids, micro-robots intended to emulate insect appearances, nano-robots and so on. The range of sizes and forms of MRA are very broad.
[0139] What unites the MRAs in the present system are common processes that make possible self-organizing group behavior of autonomous intelligent mobile robots. Nevertheless, the various specialized applications that are made possible by using the present system allow a broad range of important uses that endeavor to enhance the human condition by performing the riskiest, most remote, most complex, most routine and most important tasks imaginable.
[0140] Reference to the remaining portions of the specification, including the drawings and claims, will realize other features and advantages ofthe present invention. Further features and advantages ofthe present invention, as well as the structure and operation of various embodiments ofthe present invention, are described in detail below with respect to accompanying drawings, like reference numbers indicate identical or functionally similar elements.
List of Acronyms:
MAS: Multi-agent system
MRS: Multi-robotic system
MRA: Mobile robotic agent
INA: Intelligent negotiation agent
IMSA: Intelligent mobile software agent
Al: Artificial intelligence
D-AI: Distributed artificial intelligence
A-NN: Artificial neural network
E-A-NN: Evolutionary artificial neural network
FL: Fuzzy logic
GA: Genetic algorithm
GP: Genetic programming
EC: Evolutionary computation
OS: Operating system
CA: Cellular automata
GT: Game theory BRIEF DESCRIPTION OF THE DRAWINGS [0141] Fig. 1 is a list of system layers;
[0142] Fig. 2. is a schematic diagram of a synthetic hybrid control system for an
MRA;
[0143] Fig. 3 is a table of a dynamic database organization;
[0144] Fig. 4 is an illustration of three MRAs identifying MRA locations with sensors;
[0145] Fig. 5 is a diagram of an MRA assessing its environmental situation and coordinating change in state;
[0146] Fig. 6 illustrates a diagram of a metacomputing model for distributed MRS in which flexible mobile grid architecture is organized into dynamic clusters;
[0147] Fig. 7 is an illustration showing the sharing of computation resources among
MRA nodes in a wireless mobile MRS, including the efficient routing of database and analytical functions;
[0148] Fig. 8 is a diagram showing database coordination in a distributed MRS;
[0149] Fig. 9 is a diagram showing a dynamic distributed object relational database data flow process;
[0150] Fig. 10 is a diagram showing temporal objects in an object relational database management system;
[0151] Fig. 11 is a diagram showing mobile grid dynamics;
[0152] Fig. 12 is a diagram showing autonomous blackboards for MRAs;
[0153] Fig. 13 illustrates a diagram showing intelligent mobile software agents operations control in MRAs;
[0154] Fig. 14 is a flow chart showing MRA juvenile and adult training levels;
[0155] Fig. 15 is a diagram showing MRA attitude biases;
[0156] Fig. 16 is a flow diagram showing the learning and adaptation from environmental interaction; [0157] Fig. 17 is a flow diagram showing the MRA training process;
[0158] Fig. 18 is a flow diagram showing reinforcement learning;
[0159] Fig. 19 is a flow diagram showing hybrid learning with time constraints;
[0160] Fig. 20 is an illustration of social learning in which MRAs learn from other MRAs;
[0161] Fig. 21 is an illustration showing MRAs that teach other MRAs;
[0162] Fig. 22 is an illustration showing asymmetric MRA leadership and the emergence of temporary hubs;
[0163] Fig. 23 is an illustration showing specialized learning in self-organizing teams;
[0164] Fig. 24 is an illustration showing automated specialization in which self- organization by task division occurs for individual specialization;
[0165] Fig. 25 is a flow diagram showing a self-organizing map;
[0166] Fig. 26 is a flow diagram showing a genetic algorithm;
[0167] Fig. 27 is an illustration showing a binary genetic algorithm;
[0168] Fig. 28 is an illustration showing a genetic programming tree architecture;
[0169] Fig. 29 is an illustration showing parallel subpopulations fitness evaluation;
[0170] Fig. 30 is an illustration showing a two layer neural network;
[0171] Fig. 31 illustrates an artificial neural network connection weights;
[0172] Fig. 32 illustrates genetic programming in the calculation of initial weights;
[0173] Fig. 33 illustrates genetic programming applied to indeterministic artificial neural networks;
[0174] Fig. 34 is an illustration showing an evolutionary artificial network connection and node additions;
[0175] Fig. 35 illustrates evolutionary indeterministic artificial neural network feed forward progress;
[0176] Fig. 36 illustrates an evolutionary search for connection weights in an ANN; [0177] Fig. 37 is a flow diagram showing a fuzzy logic module;
[0178] Fig. 38 is an illustration of a neuro fuzzy controller with two input variables and three rules;
[0179] Fig. 39 illustrates a five layer evolving fuzzy neural network;
[0180] Fig. 40 illustrates an adaptive network based fuzzy inference system;
[0181] Fig. 41 illustrates a self-organizing neural fuzzy inference network architecture;
[0182] Fig. 42 illustrates a dynamic evolving fuzzy neural network;
[0183] Fig. 43 illustrates a flexible extensible distributed ANN in which ANN computation is shared between MRAs;
[0184] Fig. 44 is an illustration showing intelligent mobile software agents (IMSA) dynamics in a multi-agent system with an emphasis on MRA interactions;
[0185] Fig. 45 is an illustration showing IMSA relations between MRAs;
[0186] Fig. 46 is a flow diagram showing the operation of analytical agents;
[0187] Fig. 47 is a flow diagram showing the operation of search agents;
[0188] Fig. 48 is a flow diagram showing the initial operation of intelligent negotiation agents (INAs);
[0189] Fig. 49 is a flow diagram showing IMSA intercommunications;
[0190] Fig. 50 is a flow diagram showing INA architecture;
[0191] Fig. 51 is a flow diagram showing the pre-negotiation process;
[0192] Fig. 52 is a flow diagram showing INA logistics;
[0193] Figs. 53 A and 53B are a flow diagram showing negotiation in a distributed system with mobility;
[0194] Fig. 54 is an illustration showing the simultaneous multi-lateral negotiation process with multiple variables;
[0195] Fig. 55 is an illustration showing multivariate negotiation factors; [0196] Fig. 56 is a flow diagram showing winner determination in a competitive INA framework;
[0197] Fig. 57 is a table showing the argumentation process;
[0198] Fig. 58 is a flow diagram showing anticipation of opposing LNA strategies;
[0199] Fig. 59 is a flow diagram showing problem identification in which a group of
MRAs agree to narrow focus;
[0200] Fig. 60 is a flow diagram showing solution option development between
MRAs;
[0201] Fig. 61 is a flow diagram showing a solution option selection method;
[0202] Fig. 62 is a flow diagram showing how the MRAs select the best available solution to a problem in the present circumstance while waiting for more recent relevant information;
[0203] Fig. 63 illustrates MRA group agreement;
[0204] Fig. 64 is a table that shows the temporal aspect ofthe decision process;
[0205] Fig. 65 is a flow diagram showing the application of multivariate analysis to problem solving;
[0206] Fig. 66 is a flow diagram showing the application of regression analysis to problem solving of conflicting MRAs for winner determination;
[0207] Fig. 67 is a flow diagram showing the application of pattern analysis and trend analysis to problem solving of conflicting MRAs for winner determination;
[0208] Fig. 68 illustrates the modeling of MRS activity with simulations in which situation assessment is performed;
[0209] Fig. 69 is a flow diagram showing the synchronization of simulations within an MRA cluster;
[0210] Fig. 70 illustrates the contingency cellular automata (CA) scenario option simulations;
[0211] Fig. 71 illustrates reversible CA projecting backwards from a goal; [0212] Fig. 72 illustrates adaptive geometric set theory applied to an MRS;
[0213] Fig. 73 illustrates the optimal simulation selection in which simulation scenarios are (temporarily) converged;
[0214] Fig. 74 is a flow diagram showing the initiation ofthe aggregation process in which sets of MRAs form from the larger collective;
[0215] Fig. 75 illustrates the initiation of homogeneous MRA group formation;
[0216] Fig. 76 illustrates the initiation of common heterogeneous MRA group formation;
[0217] Fig. 77 illustrates the initiation of complementary heterogeneous (specialized) MRA group formation;
[0218] Fig. 78 is a flow diagram illustrating the initial phase of demand-initiated environmental adaptation;
[0219] Fig. 79 illustrates continuous MRA group composition reconfiguration;
[0220] Fig. 80 illustrates the continuous reconfiguration of sub-networks;
[0221] Fig. 81 illustrates dynamic group behavior adaptation to environmental interaction;
[0222] Fig. 82 is a flow diagram illustrating the parallel dynamic traveling salesman problem (TSP) with cooperating autonomous agents;
[0223] Fig. 83 illustrates the altruistic sacrifice of MRAs (gambit tactic) in order to acquire sensor information to increase chances of overall mission success;
[0224] Fig. 84 is a flow diagram illustrating the general dynamic coalition process;
[0225] Fig. 85 illustrates group MRA coordination and obstacle avoidance;
[0226] Fig. 86 illustrates specific MRA functionality via specialization;
[0227] Fig. 87 illustrates specialized MRAs working as a team;
[0228] Fig. 88 illustrates multi-functional self-organizing MRAs;
[0229] Fig. 89 illustrates surveillance and reconnaissance of a mobile object sensed and tracked by multiple micro-MRAs; [0230] Fig. 90 illustrates remote exploration with initial tracking of multiple objects with multiple micro-MRAs;
[0231] Fig. 91 illustrates sentry behavior within limited perimeters;
[0232] Fig. 92 illustrates cinematography applications with MRAs in which objects are sensed and tracked;
[0233] Fig. 93 illustrates land based toxic site clean up with multiple MRAs;
[0234] Fig. 94 illustrates dynamic cleanup of an oil spill within limited hydro perimeters by multiple MRAs;
[0235] Fig. 95 illustrates fire fighting with multiple MRAs as a dynamic interaction between the MRS and a complex environment;
[0236] Fig. 96 illustrates manufacturing production in which an object is created by using multiple MRAs;
[0237] Fig. 97 illustrates the assembly of objects in which parts are combined to create a whole object using multiple MRAs;
[0238] Fig. 98 illustrates road generation using MRAs, and;
[0239] Fig. 99 illustrates surgical micro MRAs used for trauma intervention and stabilization.
DETAILED DESCRIPTION OF THE INVENTION [0240] The present disclosures illustrate in detail the main ideas ofthe present system. Since the present invention has numerous embodiments, it is not intended to restrict the invention to a single embodiment.
[0241] The system and methods incorporated in the present invention are implemented by using software program code applied to networks of computers. Specifically, the present invention represents a multirobotic system (MRS) that includes at least two mobile robotic agents (MRAs). These MRAs have various useful purposes in the context of industrial and practical applications. The MRAs use complex software program code, including mobile software agents, to execute specific instructions involving robotic and computation operations. The software capabilities activate specific robotic functions within MRAs involving movement and decision-making.
[0242] The present invention focuses on how groups of autonomous MRAs operate in a distributed MRS. As such, the invention, or cluster of methods, solves problems in the area of computation for groups of mobile robots in a distributed network. The system shows novel ways for groups of MRAs to work together to achieve specific goals such as mapping the environment, coordinating missions, aggregating into dynamic coalitions and engaging in complex self-organizing activities. The system employs hybrid models for collective robotic control that combines not only synthetic control methods that combine central and behavior- based approaches but also hybrid artificial intelligence methods. Distributed artificial intelligence approaches are used in several contexts ofthe present system, including learning, negotiation, simulation and decision-making of MRAs and intelligent mobile software agents (LMSAs).
[0243] The main approach for decision making of MRA collectives is decentralized. In order to achieve self-organizing aggregation for specific missions in changing environments, the MRS engages in learning and decision processes that employ extensive use of IMS As. IMS As interact with each other to handle routine matters between MRAs, including communication, analysis and negotiation. Intelligent negotiation agents (LNAs) provide a medium for multilateral interaction of MRAs for group decisions. Simulations are used extensively to model and select optimal pathways for MRA group action and for the evaluation of scenarios for action.
[0244] This detailed description ofthe figures is divided into several parts that explain: (1) the main structure and operation ofthe MRS, (2) resource management ofa distributed MRS, (3) MRA learning, (4) Al and ANN, (5) EMSAs, (6) LNAs, (7) problem solving, (8) cellular automata (CA) simulations, (9) aggregation and self-organizing dynamic coalitions and (10) specific applications including (a) remote sensing, (b) hazard management and (c) building processes.
General System Architecture and Dynamics
[0245] Fig. 1 illustrates the layers ofthe multi-robotic system architecture. The first level shows a synthetic hybrid control system for MRAs including central planning control and behavior-based control aspects, which are further described in Fig. 2. MRAs are independent autonomous agents that use Al to interact with their environment using the hybrid control model.
[0246] The second layer is the level ofthe mobile robotic system in a distributed network which connects together individual MRAs using communications. The Grid computing architecture is used to link the MRAs together at layer three in order to share computation and database resources between the individual MRAs for maximum network efficiency. In this way, the MRA network develops dynamic clusters for optimal computation and storage capability. Particularly in time constrained dynamic environments, the mobile Grid network model is critical in order to accomplish complex tasks.
[0247] At level four, the dynamic distributed database system is used. This extension ofthe Grid computing hardware architecture uses object relational databases and temporal data objects to organize data between databases in the MRAs.
[0248] Artificial intelligence is used in layer five as a dynamic interactive artificial neural network that evolves. By applying Al to evolving networks of MRAs as they interact in a dynamic environment, complex learning and adaptation processes develop.
[0249] Intelligent mobile software agents (LMSAs) operate within the multi-agent system (MAS), comprising the sixth layer. The IMSAs are complex agents that perform a number of important functions within each MRA, such as analysis and decision-making, and between MRAs, such as data search, negotiation and collaboration.
[0250] The MRAs produce complex simulations to represent their relative positions and movements as well as to map out the possible scenarios for future action. These simulations are represented as mobile cellular automata in level seven.
[0251] Finally, the specific functional application of each implementation ofthe system comprises level eight. The main application categories of remote sensing, hazard management and manufacturing processes each use specific functional representations that are closest to the environment with specific hardware types.
[0252] Fig. 2 shows a multi-layer architecture of an MRA synthetic hybrid control system. The first level shows specific central (0270) and behavior-based (0280) control processes, in which the former uses abstract logic and the latter is reactive to the environment. In layer two, the two main processes are intermediated (0260) in synthetic control approaches. [0253] Layer three illustrates several main hybrid control systems that combine both central planning and behavior-based control models: (1) planning driven (0220), (2) advice mediation (0230), (3) adaptation (0240) and (4) postponement (0250). The planning-driven approach to combining the main control methods determines the behavioral component; it is primarily a top-down model. The advice mediation approach models the central planning function as advice giving, but allows the reactive model to decide; it is primarily a down-up model. The adaptation model uses the central planning control module to continuously alter reaction in changing conditions. Finally, the postponement model uses a least commitment approach to wait to the last moment to collect information from the reactive control module until it decides to act.
[0254] Finally, at layer four, the suite of synthetic control systems (0210) is constructed of various combinations of these main hybrid control models. For instance, a robotic unit may use a suite of hybrid control systems in order to optimize specific situations.
[0255] The evolution of these hybrid control models, as represented in the layered structure of figure two, is suited to complex social behaviors ofa distributed MRS used in dynamic environments.
[0256] The structure ofthe dynamic database organization is referenced in Fig. 3 as a table. A single MRA unit includes a hardware component with an object-relational database. Within this MRA, software agents perform tasks such as analysis, negotiation and decision- making. On a more advanced level, a single MRA has complex computation resources to manage, including Al and ANN.
[0257] Taken as a whole system of MRAs in a distributed network, the MRAs manage data within a network and share database organizational functions. Similarly, in the distributed network, the software agents become mobile and interact with other software agents at various MRA locations. Finally, on this network level linking MRAs, computation resources are constantly restructured so as to maximize computer power for complex time constrained applications.
[0258] On the level of mobility, MRAs change spatial positions in variable time sequences in order to perform specific tasks. The software agents are also mobile within a limited wireless range between mobile MRAs. The network of MRAs constantly rewires its computation resources by using Al and ANN in order to adapt to its environment and to optimally perform the collective mission. [0259] Thus Fig. 3 shows that while a single autonomous unit is important, when combined with other similar units in a network and provided with mobility, and when also combined with both software agent system integration and Al and ANN capabilities, the system produces a complex adaptive collective capable of autonomous mobile interaction.
[0260] Figs. 4 and 5 show simple MRA operations such as using sensors to locate other MRAs or changing position by avoiding obstacles. Fig. 4 shows a simple communication between three MRAs using sensors. Each MRA uses its sensors to detect the positions ofthe other MRAs. In this way, each MRA can identify each others' positions. In another embodiment, the position of each MRA may be transmitted to other MRAs in the network by way of wireless communications. In still another embodiment, positions of MRAs can be transmitted to other MRAs by satellite, radar or other external GPS tracking system. In these ways, the positions of MRAs can be tracked by other MRAs in the network. The reason that individual MRA position tracking of other MRAs is important is that in a noisy environment, there are multiple methods for MRAs to track other MRAs. In the total absence of communication, an individual MRA may default to a behavior-based reactive mode of interacting with other MRAs and with the environment.
[0261] An individual MRA can detect an object (0520) in the environment with its sensors and change position from 0520 to 0530 as illustrated in Fig. 5.
[0262] The individual autonomous MRAs are part of a distributed network in much the same way that inert computers are linked together into grid computing networks for supercomputing. This mobile grid computer network comprised of individual MRAs uses wireless communications in order to share computation resources. Fig. 6 shows a metacomputing model for a distributed mobile robotic system (MRS). The figure describes a flexible mobile grid architecture of dynamic clusters of mobile MRAs. At 0610, MRA 1 requests (at (l)(a), (l)(b) and (l)(c)) computation resources and data storage capacity from other MRAs. MRAs 2, 3 and 4 (at 0620, 0630 and 0640, respectively) then respond to the request (at (2)(a), (2)(b) and (2)(c), respectively) of MRA 1 (at 0650).
[0263] Fig. 7 illustrates the sharing of computation resources among MRA nodes in a wireless mobile MRS, with an emphasis on the routing of.database and analytical functions. The distributed network of MRAs can work together as one dynamic unit. Messages are input to the report status distributor (0720) and the request coordinator (0730) The report status distributor feeds messages to the MRS (0740) which interacts with the cache (0750) and the data stream (0760). The cache also interacts with the analytical (0770) functions of the system. Messages are output from the data stream and from the request coordinator. The mobile wireless grid computing architecture uses the most recent version ofthe message passing interface (MPI) for distributed computer networks. The use of grid architecture in a mobile wireless distributed network allows for a maximum of flexibility and scalability in providing massive resources in adaptive environments.
[0264] MRAs possess not only computation capability, which allow up to teraops
(one trillion operations per second) or yodaops of system processing power, but also database storage capacity as well. Each MRA possesses a database. However, taken as a whole, the MRS network comprises a distributed database system with complex coordination capabilities. The databases work together to store data objects such as a table, a calculation, a multimedia segment or other complex combinations of coherent mobile code. Such working together involves sharing database storage among a number of machines in order to ensure maximum efficiency under severe time constraints. Fig. 8 shows database coordination in a distributed MRS. The front end (0810) inputs queries at the query initiator (0820) which inputs to the query executor (0830), which has buffers (0870) with other MRAs. The multiple data sources (0850 and 0860) supply information to the query executor. The query executor outputs its queries to output queues (0840) at various other MRAs (0880). This process is further illustrated in Fig. 9.
[0265] In Fig. 9, the dynamic distributed object relational database data flow process is described. The query origination (0910) moves to the various databases (0920), DB1 through DB5, internal to MRA 1 through MRA 5. The query executor (0950) which is buffered (at 0970), searches the same databases (0980), which have sensor data stream inputs (0930) as data sources (0940). Once accessed, the databases output their data at the output queues (0990). This distributed model shows a parallel network approach to database organization. In one embodiment, the system uses active storage databases in which the computer processing capacity is internal to the database, which is itself continuously mining objects for analytical functionality.
[0266] One ofthe particular types of objects that the object-relational database management system organizes involves temporal objects. Because the MRS is typically time constrained in order to perform its primary missions, temporal objects become a prominent part ofthe distributed database system. Temporal objects reveal their temporal priority in order to be listed in a higher or lower relative priority in the database for storage retrieval purposes. Objects are "tagged" with temporal priorities such as "now", "imminent", "very soon", "in the future", "possibly useful in the future", "past", "near past", "immediate past", "urgent priority", etc. By storing, and reprioritizing, objects according to temporal priority, the system can operate much more efficiently. Fig. 10 shows how temporal objects operate in a ORDbMS.
[0267] The query generator (1010) requests the query executor (1020) to access databases at DB1 (1030) and DB2 (1050) in sequential order. These databases access the data object (1060), which is tagged as it undergoes temporal change and is given temporal priority (1040) and is then provided back to the query executor (1020). Once again the databases are accessed with temporal information about the data object. The data object is then directed to the query manager (1070) for feedback to the system. By prioritizing data according to temporal priority, the system can route data efficiently and effectively anticipate functions. Temporal data is useful in the present system in the context of evolving learning, evolving ANN, evolving game theoretic negotiation applications, evolving environmental conditions and general systemic adaptation processes.
[0268] Fig. 11 shows the mobile grid dynamics. Data sets at a specific location inform the system analysis at 1120. The data sets are analyzed and interpreted at 1110 in order to determine where the system should move. The system moves to the new position at 1130. Yet this change of position provides new data sets, which are, in turn, provided to the system for analysis in order to determine where the system should move. This dynamic process optimizes the functionality ofthe system.
[0269] It is necessary for MRAs to obtain and transmit information from other MRAs about specific data such as physical position, analysis, negotiation and decision-making. Concise data sets are transmitted between MRAs in real time about the location and analytical state ofthe MRAs. These abbreviated data sets are consolidated in each MRA by autonomous blackboards, which act as "radar readouts" informing MRAs about the state of the network.
[0270] In Fig. 12 autonomous MRA blackboards are described. In this example, limited information is referenced involving spatial position, vector and speed so that each MRA can get a snapshot ofthe present situation of every other MRA in the system. In the figure, MRAs 1 through 4 readout specific data sets in a spreadsheet format at 1210 during phase one. New data sets are presented to the same MRAs in phase two to signify a change in state ofthe network. In one embodiment, an external blackboard keeps track ofthe data as a form of back up. In the event of a centralized blackboard on board a specific MRA, such as a satellite, the leader would maintain the consolidated information function. If such a consolidated approach were used in a further embodiment ofthe system, the leader may shift, thereby providing fluidity for centralized leadership ofthe system.
[0271] Fig. 13 describes the operation of intelligent mobile software agents (LMSAs) among MRAs. Though discussed below at figures 44 to 58, LMSAs (and LNAs) are the main software based methods for MRAs to communicate, interact and collaborate with each other. MRA 1 (1310) receives a collaboration agent sent by MRA 2 (1320), as it launches a search agent to both MRA 2 and 3 (1330). An interaction process is engaged between MRA 1 and MRA 2. Meanwhile, an analytical agent is launched by MRA 3 to MRA 1, while a messenger sub-agent is launched from MRA 3 to MRA 2. Finally, the figure shows negotiation agents (LNAs) interacting between MRA 2 and MRA 3. These software based interactions represent a key method for MRAs to communicate and work with each other as a network.
[0272] Figs. 14 through 19 deal with MRA training and learning, while figures 20 through 25 deal with social learning.
[0273] Fig. 14 is a flow chart depicting the evolution of training level states. After an MRA initiates a training exercise (1410), it increases levels of training (1420). It may employ a learning module with specific learning tasks (1460) and refinement of learning tasks (1470) or it may interact with various environmental inputs (1430) in order to leam. At a specific point, a juvenile training level is achieved (1480). However, with continued experiments with the environment, it improves learning with positive reinforcement (1440) and an adult training level is reached (1450), which is constantly reinforced with a feedback loop.
[0274] MRA attitude biases are shown in Fig. 15. On a behavior spectrum between passive (1510) and aggressive (1530) behaviors lies a moderate "normal" behavior (1520). With passive behavior, the MRA acts with slower judgment but generally with more information, while with aggressive behavior, the MRA acts with faster judgment but within information constraints because ofthe time limits of quicker action. [0275] Environmental interaction is critical for learning and adaptation. Fig. 16 shows a flow chart in which MRAs interact with both other MRAs and with the environment. After an MRA initiates a training exercise (1610), it either interacts with other MRAs (1620) or with its environment (1630). When it interacts with other MRAs, an MRA queries other MRAs about a specific question (1640), while the MRAs then access databases and respond to the data query (1660). Inter-MRA feedback is then shared between MRAs for efficient learning (1680), akin to a tutorial. On the other hand, when an IVIRA interacts with the environment, as the environment changes, the MRA feedback changes (1650). In this case, negative feedback is avoided (1665) while positive feedback is attractive behavior (1670) which leads to reinforcement learning (1675) and a feedback loop with the environment. As the environment changes, new data about these changes is supplied to MRA databases in order for them to access these environmental changes. When provided with positive feedback, the MRA constantly updates its beliefs about the environment (1690).
[0276] The MRA training process includes a combination of environmental interaction with group sensor data as illustrated in Fig. 17. The MRA initiates learning (1710) and accesses either the sensor data from other MRA team members (1720) or the environment (1730). By accessing the environment direction, the MRA collects raw sensor data (1740). Whether obtained from other MRAs or directly from the environment, the MRA analyzes and interprets the sensor data (1750) and initiates a decision to act based on the data (1760). In this way, training processes may be implemented based on the data obtained, contingent on the method of originating the data (whether from the environment directly or from other MRAs). Whereas figure 17 shows the two main ways of obtaining data, Fig. 18 shows the two main qualities of information, viz., intensity and quantity of data, which provide MRA learning reinforcement.
[0277] In Fig. 18, sensor data is input into an MRA (1810) while the intensity of inputs is measured (1820) or the quantity of inputs is measured from different sources (1830). In either event, the inputs are compared to databases (1840) while each is provided a weighted value, with high intensity input weighting (1850) and quantity input weighting (1860), respectively. The MRA evaluates the weighted value from different sources (1865) and interacts with the environment based on input evaluation (1870). For instance, if a number of MRAs provide a large quantity of inputs that a mission objective is achieved, then these inputs are weighted highly in order to provide reinforcement ofa specific behavior (1880); contrarily, if a very high weighting is assigned to an individual MRA data set because ofthe intensity ofthe data, then this behavior is reinforced.
[0278] The combination ofthe aforementioned learning approaches present a hybrid learning model with time constraints illustrated in the flow chart of Fig. 19. Data from other MRAs (1910) and direct environmental inputs (1920) are analyzed (1930) before the MRA acts (1940). The MRA then proceeds to interact with the environment (1960) and receive positive feedback (1950). This environmental feedback presents behavior reinforcement in a minimal time (1980) and the MRA establishes a plan of action (1985), which is implemented by activating specific behavior (1990). Meanwhile, the MRA updates other MRAs (1970) which provides a partial feedback loop for MRAs to supply information for future sensor data.
[0279] In Fig. 20 social learning is described as MRA interaction. MRA 1 (2010),
MRA 2 (2015) and MRA 3 (2020) interact with objects (2030) in the environment in an initial phase, hi the second phase, the MRAs interact with each other by sharing information about the object-interaction. This descriptive phenomenology about the objects is used in the third phase by further interactions between the MRAs and the objects.
[0280] Fig. 21 illustrates an MRA that teaches another MRA. MRA 2 (2120), a
"student" with limited training, requests assistance from an experienced MRA 1 (2110). While MRA 1 is in motion, and thus moves to a new position (2130), the "adult" MRA 1 provides the student with a learning module via a software agent.
[0281] Given the distributed environment of the present MRS network and the learning schema presented, it is possible to have asymmetric MRA leaders. That is, if this is not a centralized system, it is still possible to have mission leaders, but they are not necessarily centralized or even consistent. Like in a flock of geese, any member ofthe flock may be a leader, though temporarily. Consequently, asymmetric MRA leadership provides the emergence of temporary hubs of MRAs that cluster together to interact with the environment.
[0282] Fig. 22 illustrates this process. In the first phase, a leader of a cluster of
MRAs (2210) interacts with a moving object (2220). But the leader is knocked out of action (2240) in the second phase, while a new leader emerges for the group (2230) as the new leader seeks the moving object (2250) and it is also removed from action. Finally, in phase three, yet another new leader emerges for the group (2260) while the mobile object (2470) continues to elude the group. At each new phase, a new hub is created with a new leader of the MRA cluster. In each case, the goal is to seek out the elusive mobile object.
[0283] A division of labor can occur in specialized teams for increasingly efficient performance as shown in Fig. 23. Each MRA is designated with a letter to signify its role as a specialist, while the whole group interacts with a mobile object (2320). In phase two, the MRAs reorganize into new positions in order to optimize the sharing of data and resources and to organize an interaction between the various specialists and the object (2340). Fig. 24 further illustrates the self-organization process by task division for automatic individual specialization. In the first phase, the group of MRAs (2410) interact with the object (2420). The MRAs automatically activate a specific specialization mode (2430) to attack the increasingly elusive object (2440) as shown in the second phase. However, at phase three, the MRAs automatically reorganize to a new specialization mode (2450) to catch the object (2460).
[0284] One key application ofthe (social) learning and environmental interaction processes is to construct a self-organizing map of an uncertain environment. This map can then be used as a benchmark for further collective action. Fig. 25 is a flow chart that shows the process ofa self-organizing map for a group of MRAs. After initial parameters are developed (2510), MRAs move to new locations to fulfill a mission (2520), where they receive sensor feedback from the environment (2530). The MRAs create an initial map based on initial sensor data organization (2540) and obtain more sensor data (2550) as they cover more terrain to include in more refined mapping phases. In this way, the MRAs fill out the initial map to create a fuller picture of tenain to include formerly missing parts (2560). The MRAs can perform this filling in procedure by using caching techniques that add the most recent information to a map outline. More complete data from sensors continue to refine the map (2570) as the MRAs continue to generate more and better data from continued mobility and data gathering. As objects in the environment change position, the MRA sensor data inputs that represent these changes continue to update the maps (2580).
[0285] Using artificial intelligence and artificial neural networks optimizes the learning process. Figures 26 through 29 show the main Al procedures of genetic algorithms and genetic programming. These techniques are then applied, in figures 30 through 43, to artificial neural networks. These discussions are important because Al and ANN are also applied to IMS As, to the negotiation process and to simulations, which will be addressed later in the figures.
[0286] In his quest for software that would solve complex optimization problems,
Holland sought a solution from nature. By emulating biological processes of breeding, mutation and survival ofthe fittest, he sought to develop a new kind of software logic that would automatically improve in order to solve problems. His revolution in software design emerged as genetic algorithms that are binary representations of genes that undergo evolutionary processes similar to biological entities. Fig. 26 describes a flow chart ofa genetic algorithm. After a population is created (2610) (and mutations added to the population (2620)), each member ofthe population is evaluated for fitness (2630). The weak members are pruned out (2640) and removed (2650) while the strongest members are selected for crossover (2660), such as breeding, which is then performed (2670). A feedback loop is generated in order to generate multiple generations of a population or a range of sub- populations. The successful candidates are put into the population to replace the weak members (2680). In this way, the population of possible algorithms evolves to an optimal solution. Fig. 27 shows an example of a binary genetic algorithm crossover in which 2710 is bred with 2720 to achieve 2730. In this example, a combination of "zero" and "one" yields a one, while two zeros or two ones combined in a specific position produces a zero.
[0287] Holland's student, Koza, developed genetic programming based not on binary algorithms but on the evolution of trees diagrams. Fig. 28 shows a genetic programming model with a crossover from the first phase of a tree on the left with a tree on the right. In this example, the triangular grouping on the upper left (in the box) (2830) is combined (2880) with the tree ofthe upper right (including the triangular grouping in box (2970)), though the two groupings are "switched" right to left in the examples. This tree structure modeling approach more closely resembles the actual genetic representation of evolutionary processes.
[0288] The process of producing multiple generations of algorithms may take an enormous amount of time because there may be many thousands of generations before a solution to a problem emerges. In order to abbreviate this, process, the evaluation part ofthe process may be performed in a parallel way. By breaking down the fitness evaluation function, the process is expedited. Fig. 29 uses the tree structure model to illustrate parallel subpopulation fitness evaluation in which two main triangular structures (2910 and 2920) break into a large number of smaller sub-populations (2930 and 2940) in order to assess the fitness ofthe best set of pairings. A final pairing is then selected (2950). Rather than mnning through a single sequence ofthe fitness assessment procedure, the parallel approach is much more time sensitive. This time sensitivity is more conducive to adaptive systems in which real-time interaction is critical.
[0289] Genetic algorithms, genetic programming and evolutionary computation techniques are applied to artificial neural networks in order to (1) calculate the initial weight and the connection weights ofthe signal between neurons, (2) train and optimize the connection weights, (3) generate the architecture and topology of a NN and (4) analyze the pattern, structure and phase state of a NN. GA, GP and EC are also applicable to a range of complex computation problems, including (1) distributed problem solving, (2) group learning, (3) group cellular automata simulations, (4) routing of computation resources in the distributed system, (5) scheduling in a dynamic distributed system, (6) creating a self- organizing map, (7) solving optimization problems, (8) performing game theoretic simulations, (9) performing parallel data mining and (10) selecting a winner from among complex aggregation choices.
[0290] Figs. 30 through 43 deal with artificial neural networks. ANNs and evolutionary ANNs have numerous applications to the present system, particularly (1) organizing and optimizing distributed networks, (2) performing dynamic data mining, (3) organizing indeterministic learning, (4) ordering and adapting simulations, (5) modeling and optimizing dynamic game theoretic interactions (6) structuring adaptive self-organization and (7) general problem solving. The field of neural networks has evolved in the last generation from a purely theoretical endeavor of logicians, mathematicians and neuro-biologists to include applications that are useful for practical systems. The present system is an example of an application of complex neural networks to learning, simulation and adaptive processes. The neural networks are computational representations within the program code of MRA hardware that provide useful tools for calculations of specific solutions to problems.
[0291] ANNs are parallel computational systems including interconnected nodes.
Sometimes called connectionism because of an emphasis on the connections between the nodes, ANNs have inputs and outputs in the connection weights between nodes. An ANN node represents an artificial neuron that is modeled after biological neurons in a brain. A perceptron is the structure that represents the sum of a neuron's inputs and outputs. [0292] Fig. 30 shows a two layer neural network in which inputs are entered on the left side and outputs are registered on the right side ofthe figure. A feedback connection can be added that directs the connections back to the left side ofthe nodes. In Fig. 31, a multilayer ANN is represented, with 3120 and 3140 representing the first layer, 3110, 3125 and 3150 representing the second, hidden, layer, and 3130 representing the output layer. In this example, the ANN structure is a multilayer perceptron (MLP). The connection weights are illustrated in numerical terms in this figure, with the bottom part having higher numbers than the upper part. There are a number of types of neural networks that may be useful for various functions of a distributed mobile multirobotic system, including the MLP illustrated here, the Hop field Network, the Hebbian Network, the Boltzmann Network, the Bayesian Network, the evolutionary ANN (neuroevolution) and the recurrent Net. These types of ANNs can be classified as having feed-forward recall or feedback recall, being deterministic or indeterministic and, finally, possessing supervised learning or unsupervised learning.
[0293] In Fig. 32, genetic programming is used to calculate initial connection weights. The GP randomly generates a population, computes the fitness of its members, generates a new population by performing a crossover ofthe first generation and adding random mutations and, finally, seeks to identify the fitness of specific members of this most recent population by comparing the best fit members with the criteria to satisfy the problem of identifying the initial weight ofthe connection. Fig. 33 shows how genetic programming is applied to an indeterministic ANN.
[0294] In phase one, the multi-layer ANN has inputs that register higher relative numbers at the top (connections between 3315 and 3310, between 3310 and 3320 and between 3325 and 3320) ofthe network than at the bottom (connections between 3315 and 3325, between 3325 and 3330 and between 3325 and 3320) ofthe network. As the network grows, shown in phase two, it emphasizes growth at the top, where there is significantly higher activity, and adds nodes at 3340 and 3350, while lower positioned nodes at 3355, 3360 and 3365 become inactive. Fig. 34 shows the automatic generation of a new node (3450) and a new connection (between 3420 and 3450) through a mutating process. This process of mutation and growth through node (and connection) addition(s) provides an evolutionary model of ANN change called neuroevolution. GP calculates both the addition ofthe node, the addition ofthe connection and the connection weights. In addition, GP can simultaneously calculate the node/connection additions, connection weight changes and the architecture ofthe evolving ANN particularly in a distributed network using parallel computation techniques. Fig. 35 illustrates an evolutionary ANN indeterministic feed forward progression from the first phase to the second phase. In the second phase, new nodes (3560 and 3580) and their connections are added, while less active nodes (3550 and 3570) and their connections are made inactive. In this way, ANNs constantly "rewire" the network towards more productive nodes.
[0295] Fig. 36 shows a 3-2-1 multilayer network in which connection weights are calculated by genetic algorithms. The algorithms are represented as binary units in order to calculate the connection weights. The network is trained by fine-tuning the connection weights through a process of optimization that the successive generations of genetic algorithms perform.
[0296] Fuzzy logic is a method to provide new approaches to computing that includes terms like "maybe," "possibly" and other partial and soft descriptions. Also called soft computing, FL represents a departure from traditional hard computing with mutually exclusive logic. FL uses statistical methods to compute solutions to complex real world problems. FL is applied to ANN to produce complex adaptive networks. Fig. 37 describes a FL module. A sensor provides crisp data input (3710) to a fuzzifier module (3720), which is fed random mutations (3770) and proceeds with the fuzzification process. At this point, a fuzzy analysis proceeds in a fuzzy inference engine (3730) that operates according to fuzzy rules (3780) which are themselves adapted (3790). After the fuzzy analysis process, the defuzzification of data occurs in the defuzzifier module (3740) where crisp data is output (3750) and presented to actuators (3760) for functional performance. This process is similar to a the process a signal undergoes in conversion from an analogue waveform to a digital mode by way of an analogue-digital (A to D) converter, or, contrarily, from a digital to an analogue signal by way of a digital-to-analogue converter (DAC).
[0297] Fig. 38 shows a neuro fuzzy controller with two input variables and three rules. The input variables Al (3810) and A2 (3820) provide connections to the rule base RI (3830), R2 (3840) and R3 (3850), which then provide an output at X (3860). Fig. 39 shows a five layer evolving fuzzy neural network, with the input layer (3910), the fuzzification layer (3920), the rule node layer (3930), the decision layer (3940) and the output layer (3950). A more complex ANN architecture is described in Fig. 40. In this figure, an adaptive network based fuzzy inference system is shown in which inputs are presented to the initial presentation layer, which is shown here in a parallel configuration, with R and S nodes. A training process occurs in the multilayer network (4040) that contains hidden layers. The outputs of this training process are fed to the consequent parameters (4050) that then lead to outputs.
[0298] A multilayer neural fuzzy inference network is illustrated in Fig. 41. The first layer (4110) generates the offspring (4115), which produce neural nodes at level three (4120) that are evaluated for fitness at level four (4125). The nodes breed a new generation with inactive nodes (4130) at level five. The surviving nodes (4135) again breed a new population of nodes that result in two active members (4140) in layer seven. The successful mating of these nodes yields an output node (4145) at level eight. By breeding successive generations of successful populations, and by training these successful populations, the network is self- organizing and adaptive to its environment.
[0299] Fig. 42 shows a dynamic evolving fuzzy neural network. With five layers, including an input layer (4250), a fuzzy quantification layer (4255), an evolving rule nodes layer (4260), a weighted least square estimator layer (4265) and an output layer (4270). This model shows a complex synthesis of simpler ANN representations.
[0300] One ofthe advantages of applying evolutionary computation to ANN is that such advanced computing can be performed more efficiently by using parallel approaches to break down a problem into smaller parts so that a larger number of computer processors may solve the problem simultaneously. In this way, multiple MRAs may work on a problem together in order to accomplish the task in real time. One application of this approach is in the fitness evaluation part ofthe genetic algorithm population production process. The problem of identifying the successful candidates in a population can be performed, and expedited, by using parallel processing.
[0301] In another example of using parallel processing to accelerate computation problem solving, consider the problem of adding a neural node. By using parallel computation processes, not only can the neural node be added, but the connections to it can be added simultaneously; in addition, the architecture ofthe network can be configured and reconfigured in real time, as new training models are considered and tested. The application of parallel algorithms to evolutionary computation, and, in turn, the application of EC (both GA and GP) to ANN provide increasingly efficient approaches for use in a distributed mobile MRS. MRAs share ANN computation in a flexible way as illustrated in Fig. 43. Not only are the ANN not limited to the computation of a single MRA, but MRAs (4310 and 4320) may share ANN computation resources between them. This distributed manifestation of parallel computation shows a flexible and extensible model in which the sharing of resources results in increasingly efficient capabilities.
[0302] The application of EC and ANN to a distributed mobile MRS involves several important areas, including learning, traimng, adaptation and prediction. In order for MRAs to interact with an uncertain, and changing, environment, it must leam, predict and adapt. While EC is useful to train ANNs, it is the general learning capabilities that are regarded as an outcome of this training process that ANNs ultimately provide to MRAs and to the MRS that is critical to the effective real time adaptation needed by the system. Many ofthe problems that a mobile MRS encounters involve evolving solutions, adaptive behavior patterns, complex predictive scenario modeling and self-organized processes. These problems are solvable by applying EC and ANN models.
[0303] An example ofthe application of EC and ANN to an MRS is the modeling of game theoretic interactions. A particular strategy may be evolved for a particular player based on a basic rule pattern selection organized by a multilayered feed forward perceptron. Each layer performs a calculation ofthe weights of inputs, connections and biases. A random number of nodes is selected in the multilayer network, with a random number of offspring replicated from each parent and randomly mutated. A number of rules of game moves are identified and consistently applied. Each network generation is evaluated for accurate effectiveness of achieving a successful game move. The network is trained and retrained with full information, i this way, the learning process is refined so that each player is able to optimally move according to the rules.
[0304] This straightforward application of EC and ANN to a game theoretic modeling problem is relevant for a distributed mobile MRS because the present system uses simulations to model action. The simulations, which are discussed below in Figs. 68-73, can be either present-time-based or may be based on future scenarios. Since in the case ofthe MRS, multiple MRAs provide sensor data inputs into the system and multiple MRAs provide computation resources, the complexity ofthe game theoretic interaction increases with the size ofthe network. Only EC and ANN, along with parallel computation of a mobile grid computing system, is able to calculate the increasingly complex problem solving algorithms necessary to organize such a model. The main systemic unit that is able to organize such a complex architecture is the intelligent mobile software agent (IMSA) operating within a multi agent system (MAS). IMSAs, introduced in figure 13, are discussed in figures 44 through 49.
[0305] LMSA dynamics within the MAS are discussed in Fig. 44 in the context of
MRA interactions. MRA 1 (4410) launches a collaboration agent that is received by MRA 2 (4420) and collaboration between the two MRAs is initiated. A search agent is launched from MRA 1 to search databases in MRA 3 (4430) and MRA 4 (4440). A negotiation between MRA 3 and MRA 4 occurs by using intelligent negotiation agents (LNAs). LNAs are further discussed in figures 48 and 50 through 58 below. Finally, analytical agents are launched by MRA 4 to MRA 2 and MRA 3 in order to analyze a specific problem. Fig. 45 shows IMSA relations between MRAs. MRAs are able to communicate with each other about complex tasks simultaneously by using IMSA specialist agent roles. There are number of specific types of IMSAs, including analytical agents, search agents, collaboration agents and negotiation agents. Figs. 46 to 48 briefly describe analytical, search and negotiation agents.
[0306] In Fig. 46, analytical agents are described. After an MRA identifies a problem
(4610), it generates an analytical agent (4620). However, the process of initiating the AA begins with the generation of a search agent, which is sent to multiple MRAs' databases with an initial query (4630). The search agent reports back to the initiating MRA with the priorities of data in MRA databases (4640). The AA is then sent to the MRA in the order of priority sequence (4650) revealed by the search. The AA analyzes the problem using specific methods (4660) detailed at 4670 including MVA, regression analysis, pattern analysis, trend analysis and hybrid analyses. The AA develops solution options to the problem (4680) and shares the results with relevant MRAs (4690).
[0307] Search agents are described in Fig. 47. An MRA generates a search agent (4710) to query distributed MRA databases (4720). The search agent receives initial feedback from databases regarding initial query (4730), refines the query with specific databases (4740), evolves search parameters (4750) and seeks specific data sets among databases (4760). The search agent finds data sets as a result ofthe refined search (4770) and retrieves them for the MRA.
[0308] The general negotiation process is described with reference to intelligent negotiation agents (LNAs) in a distributed network in Fig. 48. The initiator INA meta- agent (4810) begins the process by launching initiator INA micro-agents to several other MRAs. LNA micro-agent 1 is launched to a negotiation session at LNA 2's location (4820), LNA micro-agent 2 is launched to a negotiation session at LNA 3's location (4825) and LNA micro- agent 3 is launched to a negotiation session at INA 4's location (4830). Each respective negotiation session occurs at each INA's location within its MRA (2, 3 and 4, respectively). The initiator INA interacts with LNAs at the various remote MRA locations (or at its home location) (4850), while a winner is determined at its home location (4855). Mutual agreement is reached, in this case between INA 3 and the initiator LNA (4860), while sessions are closed between the INA 2 and INA 4 negotiations (4865) and the overall negotiation process is closed (4870).
[0309] Fig. 49 describes an LMSA intercommunication with messenger sub-agents.
Once an MRA makes a decision (4910), the content ofthe decision is translated into specific instructions of action (4920) and the MRA creates messenger sub-agents (4930). The MRA launches the messenger sub-agents to other MRAs (4940), which then deliver the message with the instructions to the MRAs (4950).
[0310] Because LNAs are used in a critical way in a distributed mobile multi-robotic system, they are further developed in figures 50 through 58, including a description ofthe INA architecture, pre-negotiation process, INA logistics, negotiation process in a distributed network, multi-lateral negotiation process, multivariate negotiation factors, winner detennination process, argumentation process and opposing LNA strategies.
[0311] LNAs work by negotiating between at least two MRAs. LNAs use argumentation methods to negotiate by presenting arguments with variable weights. INAs also negotiate about the best simulation to use in a specific situation. In general, LNAs use multi-lateral and multivariate negotiation in order to come to agreement between noncooperating MRAs. In the case of competitive MRAs that negotiate for a compromise, problems are solved using group problem-solving and analytical techniques. Solutions to complex MRA group problems include the optimal or a temporary choice between solution options. Group problem solving is discussed in figures 59 to 67. In all cases, Al is used in order to facilitate the negotiation and problem solving processes.
[0312] In Fig. 50, the main INA architecture is described. Four LNAs, including an initiator LNA (5020) and INA 2 (5010), INA 3 (5015) and LNA 4 (5025) enter into a pre- negotiation session (5030), which is discussed more fully in figure 51 below. After pre- negotiation, all INAs negotiate in a first session between the initiator INA and the several INAs (5040), but stops negotiating with LNA 4 (5045). While the initiator LNA continues to negotiate with LNA 2 and LNA 3 in session two (5050), it eventually stops the negotiation process with LNA 2 (5070). However, the initiator LNA continues to negotiate with LNA 3 in session 3 (5060) where it reaches agreement (5080) and closes the session (5090).
[0313] Referring to Fig. 51 , the pre-negotiation process is described. After an initiator INA requests negotiation terms (5110), an INA micro-agent is launched (5120) and the initiator LNA moves to other locations in order to communicate with other INAs (5130). Several INAs, designated as SI (5140), S2 (5145 and Sn (5150) enter into a pre-negotiation process with the initiator LNA over parameters ofthe interaction session, including location(s), protocols, rules and methods (5160). If they do not agree on the negotiation parameters, they continue to interact until they do agree on these issues. The LNAs agree on the rules of negotiations, the number of negotiation sessions and so on, based on the constraints (5170) and the initiator INA proceeds to the negotiation sessions with the other LNAs based on these pre-negotiation protocols, rules and methods (5180).
[0314] In reference to Fig. 52, INA logistics are described. After initiating the session (5210), agents are generated and identified by codes (5215). The initial agent interaction protocols are generated (5220) in order for the agents to establish a common communication methodology. Such communication processes involves translation (5225) and synchronization (5230). Failure to synchronize communication leads to a termination at 5245. Once fully synchronized, LNAs may construct unique negotiation strategies using Al (5240) utilizing analytical agents (5235). At this point, agents signal the intention (5250) to negotiate with other agents. After signaling to other agents, LNAs send out communication streams (5255) to their home base, thereby constantly revealing to the home base their locations, status and plans. At this point, the initiator LNA enters into a pre-negotiation session with the selected LNAs (5260) and launches micro-agents to negotiate with INAs at different locations (5265). The LNAs then enter into the negotiation process (5270) and either cease negotiation (5275) or come to an agreement (5280). If they cease negotiation, the LNA settings are saved for later (5285) and the session closed. On the other hand, if there is agreement, the MRA functions are activated consistent with the agreement reached (5290).
[0315] Figs. 53A and 53B illustrate the negotiation process in a distributed system with mobility between INAs. The present example focuses on a one-to-one negotiation between an initiator LNA and INA 2. After an initiator LNA initiates a negotiation session with LNA 2 (5310), the INAs identify possible locations (5315) and specify agreed locations (5320) at which to negotiate. In the illustrated example, the initiator LNA moves to LNA 2's location (5323) with program code. LNA 2 identifies incoming initiator-LNA entry after activation and security protocol approval (5326) at INA 2's location.
[0316] The agents engage in (5330) and complete (5333) negotiation tasks, after which the initiator LNA notifies its home MRA of its remote location activities by sending a message (5336). After reviewing more tasks at the remote INA 2 location (5340), the initiator LNA either terminates (or returns home) (5343) or assesses additional tasks using internal database and analysis (5347), assessment (5350) and identification (5353) ofthe next location for task execution and moves to another location (5356).
[0317] After moving its program code (5360), the initiator LNA identifies a need for
Al computation (5363), requests Al computation resources at a specified location (5367), identifies available Al computation resources (5370) and messages a request for Al computation resources to be sent to a specific location (5373). The initiator INA receives (5377) and tests the Al computation resources at a specific negotiation site (5380). The negotiations are completed at the remote location (5385) and the initiator LNA returns home (5390).
[0318] As shown in this figure, though a one-to-one interactive negotiation is shown between an initiator LNA and another INA, an initiator LNA (or its micro-agents) may negotiate simultaneously with at least two INAs at two or more INA locations in another embodiment.
[0319] Fig. 54 shows a simultaneous multi-lateral negotiation process with multiple variables. At each phase in the process, denoted on the left column, INA 1 is in the position of negotiating with six LNAs, listed here as 2 through 7. In the first phase, after negotiation with LNA 2 in the first session, LNA 1 negotiates with LNA 3 in the second session. In the third session, LNA 1 negotiates simultaneously with LNA 2 and INA 3 on the second phase of negotiation with each. In session four, however, LNA 1 begins to negotiate with LNA 4, while it continues to negotiate with LNA 2 in a third phase. Similarly, in the fifth session, LNA 1 continues to negotiate with LNA 4 in a second phase, while it begins to negotiate with LNA 5 in a first phase. The sixth session continues this approach of continuing with LNA 5 in a second phase while it initiates a negotiation with INA 6, and so on in session seven. [0320] Fig. 55 shows multivariate negotiation factors in which, in the first phase,
MRA 1 negotiates over specific variables with MRA 2, rejecting successive possible variables until finally agreeing on, and thus selecting, "Z". In the second phase, MRA 2 negotiates over specific variables with MRA 3 in a similar way, also resulting in the agreement over, and selection of, "Z". This process of negotiating over a number of factors shows the key element of "convergence" to negotiation. By repeating this process a number of times, many INAs may agree with each other about numerous factors in a complex dynamic system.
[0321] Fig. 56 shows the tournament style winner determination process in a competitive LNA framework. Several LNAs (2 through 5) enter into a negotiation with an initiator LNA (5650) in phase one. The initiator INA agrees to narrow down the field to LNA 2 (5660) and INA 4 (5670) in phase two. Between these finalists, the initiator INA then selects the winner, LNA 4 (5690) in the third phase.
[0322] The argumentation process is shown in Fig. 57. During consecutive temporal phases of a negotiation process between MRA A and MRA B, several key factors are isolated and accepted by each MRA. First, negotiation variables are accepted by MRA B. Second, MRA A prunes out variables that it will not compromise on. Next, MRA B prunes out non- negotiable variables. Finally both MRAs determine the key variables that the will compromise on.
[0323] Negotiation is a process that fits into the overall game theoretic model that organizes competitive agents across limited goods. In this sense, negotiation involves agent strategies that anticipate opposing agent strategies. Fig. 58 shows the anticipation of opposing LNA strategies. After LNA1 presents an argument to LNA2 (5810), LNA 2 evaluates the argument (5830) using multi-variate analysis and regression analysis (5820). LNA2 anticipates LNA1 ' s strategy by examining the traj ectory of arguments (5840), which it performs by identifying cues to anticipate behavior in its environment (5850). INA 2 then presents a counter-argument to LNA 1 (5860). INA 1 anticipates INA 2's strategy by anticipating its possible argument scenarios (5870) and the LNAs eventually reach an agreement (5880).
[0324] Figs. 59 through 67 describe group problem solving.
[0325] In Fig. 59, problems are identified by MRAs and the collective agrees to narrow the focus ofthe problem. Any MRA in the group can identify a problem (5910), in sequence, such as "How to carry out a mission with other MRAs?" (5920), "How to combine with other MRAs for a common mission?" (5930), "How to target an object with a group of MRAs?" (5940) or other mission or goal based problems (5950). The group of MRAs prioritize problems by assigning values to each problem and ordering them by rank in real time (5960) so that potential solutions can be made in the ranking order (5970).
[0326] Solution options between MRAs are described in Fig. 60. A shared four- dimensional grid is created by MRAs in order to represent the framework of a potential field (6010). Simulation scenarios from the MRA group are tested in order to detect the best fitting solution for a specific option (6040) after analyses are performed on specific solution options by MRAs (6030). A competition is then established between various potential solutions for the best solution available (6050) and weights are attached to each solution option (6060) which allows the simulation scenario solution options to be ranked (6070).
[0327] Fig. 61 describes the solution option selection method developed and applied by MRAs. An MRA develops a benchmark of methods in order to select a simulation scenario (6110) and then applies an experimentation process to test possible solutions (6120). The shortest path option is selected as a default without environmental interaction (6125). But the MRAs interact with the environment (6160), a process that is informed by actual environmental change (6150). The MRAs receive the results ofthe environmental interaction (6170) and evaluate the results (6180). Each MRA has a distinct vantage and thus applies a unique analysis (6190). The MRAs prioritize the results by weighting them for probability of success and by ranking them in the order of highest probability (6130). The methods of solution selection are refined (6135) and a feedback loop is structured to apply continued experimentation, when combined with continued environmental interaction, in order to continue to refine the methods of solution selection. A winner is selected from the possible solution options (6140) and the optimal solution is selected for a possible scenario (6145).
[0328] There are times when an optimal solution to a problem is not possible. In these instances, the best we can hope to achieve is the best available solution in a specific circumstance. Fig. 62 describes this process of selecting the best available, not the optimum, solution, to a problem, while waiting for the most recent relevant information. The MRAs work together to establish a list of solution options (6210), which are filtered according to constraints (6220) by time, optimization, combinatorial optimization, accuracy, quality of information and by pruning out what is not probable (6230). The MRAs then apply solution option methods (6240) which are refined by interaction with the environment (6250). The MRAs either (1) undergo a convergence of agreement (6250), in which case they select a specific simulation scenario solution option (6275) and carry out a mission (6285), (2) partially agree with an overlap of interests within constraints (6260) and (3) temporarily agree (within constraints) (6265), in which cases they select merely the best available simulation scenario solution option (6280) and carry out a mission within these constraints (6290). On the other hand, the MRAs, may not agree at all (6270), in which event they must return back to the earlier phases ofthe process of filtering the solution options (at 6220). Fig. 63 shows an illustration of MRA group agreement.
[0329] In Fig. 63, part (A), three MRAs present arguments that are represented as small circles within the larger circles. The gray area that shows the overlap ofthe three MRAs signifies the common interest between the three. In the second diagram at (B), the best available optimum scenario is shown in the gray area with time constraints. The configuration of this optimum window of opportunity, because it is time constrained, changes with the changing circumstances ofthe environment.
[0330] Clearly, the time aspect ofthe decision process is important because perfect information is rarely available and because agents in a multi-robotic system that interact in uncertain and dynamic environments benefit from waiting for the latest available information before deciding to act. Fig. 64 shows the temporal aspect ofthe decision process, with the left column representing the temporal component, the second column representing the physical state ofthe multi-robotic system and the right column representing the analytical state ofthe multi-agent system. In the first line, past physical experiences influence past data flows, while past data flows affect future scenarios. Future scenarios affect present analysis and decision-making, which influence the selection of a preferred scenario of action. This section ofthe preferred scenario influences the present course of action. In this way, the analytical and physical states ofthe system have causal connections over time. These interconnections reveal the integration between the MAS and MRS.
[0331] The group problem solving process requires specific analytical methods, including multivariate analysis, regression analysis, trend analysis and pattern analysis, in order to select a successful candidate. Figs. 65, 66 and 67 describe these analytical tools.
[0332] In Fig. 65, multivariate analysis is applied to problem solving. A problem is forwarded to an MNA filter (6510), which strips the variables from the problem and analyzes each variable in isolation (6520). The MNA filtering process forwards the variable analysis procedure to multiple MRAs (6530) using parallel processing, where each MRA analyzes variables and compares this analysis with other MRA analyses (6540). The MRAs rank the multiple variables and share with the results between the MRAs (6550). The variables are evaluated in each solution option (6560) and the best available solution is selected from solution options (6570).
[0333] In Fig. 66, regression analysis is applied to the problem solving of conflicting
MRAs for a winner determination. The MRAs analyze a problem with a regression analysis filter (6610), sort through various variables (6620) and share the data between them (6630). Again, the MRAs divide the analysis between them in order to benefit from the advantages of parallel computation. The MRAs weight the variables by establishing priorities and comparing each variable with program parameters (6645). The MRAs evaluate the importance ofthe variables by comparing them with data sets in the distributed database (6650) and then rank the priorities of variables (6660) and apply the ranking ofthe problem variables to solution options (6670). MRAs select the best solution option by applying the program parameters (6680).
[0334] In Fig. 67, pattern analysis and trend analysis are applied to problem solving of conflicting MRAs for winner determination. Depending on which type of analysis is required, a problem is formulated (6710) and either pattern analysis (6720 and 6730)) or trend analysis (6725 and 6735) is applied. The pattern analysis approach analyses regularities in spatial coordinates using statistical methods (6740), while trend analysis analyses regularities in temporal coordinates using statistical methods (6745). In either case, each analysis is evaluated (6750), the results ranked (6760) and the analyses are applied to MRA decision logic (6770). The MRA group then makes a decision based on these analyses and formulates a plan (6780) that the group is able to activate (6790).
[0335] Much of the substance of the problem solving, and negotiation, processes underlying inter-MRA conflict involves simulations. Because MRAs are mechanical entities that assume physical shape and mobility in space and time, it is possible to model them by using simulations. The MRS may use a number of types of simulations, including cellular automata simulations, particle simulations and game theoretic simulations. All three main types of simulation add valuable qualities to the representation of complex activities in a mobile distributed multi-robotic system, including structuring the dynamics of aggregation processes. Figures 68 through 73 describe the cellular automata simulation of MRA group activities.
[0336] Cellular automata (CA) is a system of cells which are represented digitally as a binary unit or vacuum. As objects move through a grid, they fill up the space in the cell. If an object does not occupy a cell, it is empty. In this straightforward way, CA can simulate groups of objects in space and time. CA's may include two dimensional, three-dimensional or four-dimensional (i.e., including the time dimension) structures. Once including the time dimension, it is possible to model CA simulations. CA simulations are well suited to represent mobile distributed multi-robotic systems because the MRAs are seen as merely objects that move in space and time across a map in a nonoverlapping environment. Though the simulations may be complex, for instance, in modeling dynamic coalitions in adaptive sequences as they interact with a fast changing environment, their representation is critical in order to provide a mechanism for the self-organization ofthe MRS processes.
[0337] Fig. 68 shows the modeling of MRS activity with simulations in a situation assessment. As the illustration shows, a cubic space is occupied by mobile agents, represented here as A, B and C. In the case of this situation assessment, the map describes the change in spatial position ofthe agents from Al to A2 to A3 (6840), from Bl to B2 to B3 (6850) and from Cl to C2 to C3 (6860).
[0338] Fig. 69 describes synchronizing simulations within an MRA cluster. An MRA sensor detects other MRA locations (6910) and converts the analogue sensor data to a digital form (6920). The MRA data about other MRA positions is analyzed in real time to show phase state changes (6930) and a simulation is constructed to represent data about MRA position changes (6940). Each MRA continuously tracks all MRAs in the system in real time (6950) by using this approach and each MRA constructs a simulation to represent MRAs in the system (6960).
[0339] Fig. 70 describes a CA scenario option simulation. Two scenario options are presented for A and B. For scenario option A, MRA 1 (7010) and MRA 2 (7020) move across four phases to objects X and Y. For scenario option B, MRA 1 (7030) and MRA 2 (7040) move across the four phases towards objects X and Y but in a different path.
[0340] Fig. 71 describes a reversible, or deterministic, CA in which a simulation is constructed by projecting backwards from a goal. Though the scenario option representations look very similar to figure 70, the phasal process that is used is exactly opposite the causal approach. Rather, in this simulation model, the MRAs begin with the goal and project backwards. By using this reversible approach, the CA simulation is presented with a more goal-oriented solution.
[0341] Fig. 72 shows how adaptive geometric set theory is applied to an MRS. The three CA models of A (7210), B (7220) and C (7230) show three different sequences from one to three reflecting different positions. In the converged model (7240), a combination of the three models is reached which synthesizes the three by compromising the outcomes of B and C. Geometric set theory is useful to represent the overlap of aggregated sets.
[0342] Fig. 73 shows the selection of an optimal simulation as a (temporary) convergence of simulation scenarios. MRA 1 is represented by actual positions at 1 ' , 1 " , 1 ' " and 1"" (7320) while a possible scenario is represented by 1R", 1R'" and 1R'" (7310). Similarly, for MRA 2 (7330) and the possible simulation scenario (7340). Finally, the outcome for these sequences is a convergence of MRA 1 at 7325 and of MRA 2 at 7335.
[0343] Figs. 74 through 78 describe the aggregation process in a multi robotic system. Figures 75 through 84 describe the dynamic coalition (or reaggregation) process in a MRA and figures 85 through 88 show autonomous MRS self-organizing processes.
[0344] Fig. 74 describes the aggregation initiation process in which sets of MRAs form from a larger collective. The MRAs develop and present simulations (7410), test the simulations (7415), prune out the least useful simulations (7420), and compare the best simulations with the environment (7425) and with (updated) program parameters (7430). The best simulations (within constraints) are selected (7435) and converged (7450) in order to create overlap. From the converged simulations, a map is created (7455) and individual MRA locations are identified relative to their positions on the map (7460). The MRAs then move their physical locations in an efficient way according to the geometric location ofthe converged simulation map (7465).
[0345] The initiation of homogenous MRA group formation is described in Fig. 75.
In the first (top) section, an object X (7510) is confronted with seven similar MRAs (7520). After undergoing an aggregation initiation phase, the MRAs (7540) are shown in the second section as changing their position with regard to object X (7530) by moving towards the object. [0346] hi Fig. 76, the initiating process is shown involving common heterogeneous
MRAs. hi the first phase, an MRA with type "S" (7610) initiates a group of specialized MRAs (7620). In the second phase, the "S" MRAs (7630) concentrate in order to perform a specific task, while the other types of MRAs (7640) retain their positions. In this case, a particular type of specialized MRA is "picked out" in order to perform a specific function as a specialized unit.
[0347] In Fig. 77, a complementary heterogeneous MRS group formation initiation is described. In the first phase, the IVIRA with type "S" (7710) initiates a group of specialized MRAs in a similar was as with common heterogeneous MRAs. However, rather than attracting the same "S" type, it requests the "Y" and "T" types (7730) from the second column which leaves the other MRAs in their stable positions (7740). In this way, complementary specialists may work together as a team to perform complex functions in tandem.
[0348] The first phase of a demand-initiated environmental adaptation is described in Fig. 78. From the combination of static environmental data maps (7830) and actual environmental changes (7820), dynamic environment data maps (7825) are created. These maps inform past and present simulations (7850), which are analyzed (7840). The analysis is itself informed by learning methods (7810). Given the simulations and their analysis, negotiations occur between the MRAs (7855), which reach a decision, within limits (7865). This decision is also informed by limited (converged) scenario simulations (7870). Once a decision is made by MRAs, the selection is made about the specific form of aggregation to use (7875) and the actual special positions ofthe MRAs are changed in accordance with this new decision (7880).
[0349] In Figs. 79 to 84, dynamic coalitions, or re-aggregation processes, are discussed. In Fig. 79, the continuous MRA group composition reconfiguration process is described. In the first phase, a group is concentrated (7910) that includes MRAs 1, 2, 3 and 4. In the second phase, a new grouping is organized (7930) that includes MRAs 3, 4, 5 and 6. Finally, in the final phase, yet another grouping is organized to include MRAs 4, 5, 6, 7 and 8. The movement through the system from the left part ofthe group to the right part ofthe group illustrates the changing interaction response to the environment that requires the grouping to adapt to different sub-sets ofthe larger collective. [0350] In Fig. 80, the continuous reconfiguration of sub-networks is described. In this figure, the right column shows an object that the MRA group(s) on the left move towards. The first phase ofthe process is identified in the right column. In the first part of the process, at 8005, the first sub-set ofthe collective moves towards the object. In a second phase, the MRAs reconstitute the configuration ofthe MRA grouping (8010) and move toward the object. In a later phase, in the middle map, a larger initial grouping, including six MRAs (8015) move toward the object, while a second grouping (8020) moves to the object later. This second group includes the overlapping two members of both groups. However, in the third part ofthe process, the demand for MRAs changes again from the second part ofthe process. In this case, five MRAs move toward the object, while a grouping including six MRAs (including the last three ofthe first phase) move towards the object. This figures illustrates the dynamic motion aspect ofthe aggregation process as coalitions are dynamically created and reconfigured.
[0351] Fig. 81 illustrates dynamic group behavior adaptation to environmental interaction. In the first phase, the first MRA grouping (8120) moves towards a group of objects (8130). In this first phase, one object is knocked out, represented by an X, but two MRAs are also removed. In the second phase, the reconstituted group of MRAs (8150), which includes the combination of 8120 and 8110, move towards two more objects in the group of objects (8160) and three MRAs are removed from action, as represented by an X's. In the third phase, the newly reconstituted MRA group (8170) that includes a combination of 8150 and 8140, move towards the three remaining objects (8180).
[0352] The parallel dynamic traveling salesman problem is described with cooperating autonomous agents in Fig. 82. After they receive a sensor data stream (8210), a group of MRAs collect environmental data by sharing sensor data (8240) and use the initial prioritization of environmental data consistent with program parameters (8250). As the environmental data changes (8270), an interaction between MRAs and the environment occurs (8820) which informs the MRA sensor data stream (8210). The environmental data changes (8270) also reprioritize the order of priorities with the latest information ofa changing environment (8280); this reprioritization ofthe order of priorities are largely based on the MRAs' prioritization of a physical sequence (8260) based on a reprioritization of MRA program parameters (8230). Once the reprioritization of priorities with the latest information (to accommodate a changing environment) occurs, the MRAs perform a physical sequence of actions in the order of priority (8290). This process involves a dynamic connection between the analytical functions ofthe MAS and the physical processes of sensor data gathering from multiple changing MRA positions that yields variable data inputs from a changing environment. Because the MRS is distributed, the use of parallel processing allows increasingly efficient processing of computation resources. Figure 64 also illustrates a data flow process that accommodates both physical state and analytical state dynamics across time.
[0353] Fig. 83 shows altruistic MRAs sacrificing themselves in order to acquire sensor information to increase the chances of overall mission success. The MRAs shown with an X, move toward the object (8320) and are knocked out. However, the information that is obtained in this gambit mission is then sent back to the collective so that they are better able to defeat the object.
[0354] The general dynamic coalition process is described in Fig. 84. After mission goals and parameters are established (8455), sensor data and various sources examine the terrain (8460). The simultaneous parallel computation by numerous agents is performed by sharing data and by dividing computation resources (8465). The sensor data is then evaluated by various MRAs (8470). Groups of MRAs begin to emerge by agreeing to aggregate (8475). Decisions are made to form smaller groups in order to meet evolving mission parameters and priorities. Specified MRAs update the navigation plans and activate the mission (8485). As the mission evolves, groups of MRAs are added or removed as needed, for instance if the opposition is particularly hostile (8490).
[0355] Fig. 85 describes the group coordination and obstacle avoidance process that is involved in autonomous MRS self-organizing processes. Obstacles X, Y and Z (8510) move towards MRAs A, B and C (8550) from their initial positions. As the objects get closer, at 8520, the MRAs detect the objects as obstacles, at 8540, and begin to avoid them by moving out ofthe trajectory ofthe moving objects (8530).
[0356] In Fig. 86, specific MRAs A (8610), B (8630) and C (8650) move towards specific objects X (862), Y (8640) and Z (8660), with A attacking Z, B attacking X and C attacking Y. This specialization of a self-organizing process is further developed in Fig. 87 as a specialized group of MRAs work together as a team. MRAs A (8710), B (8720), C (8780) and D (8790) move into positions 8730, 8740, 8760 and 8770, respectively, in a phase in the process towards assembling together at 8750. In this position, the specialized MRAs work together sharing specific functions for greater usefulness on a mission, hi Fig. 88, multi-functional MRAs are described in a self-organizing process. Whereas in figure 87, the MRAs are specialized, in figure 88, the MRAs have multiple functions that may switch in specific changing circumstances. As the figure shows, MRA A in position Al (8810) and MRA B in position Bl (8850) move towards object X (8830). As they move towards the object, the MRAs detect the need to change from one specialized function to another. At positions A2 (8820) and B2 (8840), the MRAs change their functional mode to a different specialty in order to be more effective in their mission against the object.
[0357] Figs. 89 through 99 describe specific applications ofthe present system.
There are three main categories of application, including (1) remote sensing (described in Figs. 89 to 92), (2) hazard management (described in figures 93 to 95) and (3) building processes (described in Figs. 96 to 99). Remote sensing activities that use an MRS include surveillance, reconnaissance, remote exploration, sentry activities and cinematography.
Hazard management activities include toxic site clean-up, oil spill and fire fighting activities.
Building processes include manufacturing production and assembly, road building and surgical activities.
[0358] In Fig. 89, surveillance and reconnaissance is described using multiple micro objects for sensing and tracking of a mobile object. As two MRAs, X (8910) and Y (8960) move in parallel tracks to positions X2 (8920) and Y2 (8970), respectively, they track object A (8940). As the object moves to position A2 (8950) and then to position A3 (8955), MRA X moves to position X3 (8930) and then to position X4 (8935), while MRA Y moves to position Y3 (8980) and then to position Y4 (8990) by using sensors and by tracking the object closely.
[0359] In Fig. 90, a remote exploration process is described in which the initial tracking of multiple objects is performed by multiple micro-MRAs. In this example, MRAl (9010) moves towards object RI to position X'. However, the object itself moves, from position R' to position R" and is followed by the MRA, which moves to position X". This process is repeated with MRA 2 (9020) tracking object R2 (9060) and with MRA 3 (9030) tracking object R3 (9070).
[0360] Fig. 91 describes sentry activity within limited perimeters defending multiple objects with a multiple number of MRAs. In this illustration, the MRAs are spaced evenly apart in order to occupy a constrained field around the perimeters of a field. [0361] The current system is also applicable to cinematography, wherein one mobile object (or cluster of mobile objects) are sensed and tracked with MRAs. This process is described in Fig. 92. MRA 1 (9210) and MRA 2 (9270) track object X (9240) as it moves to positions 9250 and 9260. MRA 1 tracks the object along a path to position 9220 and 9230, while MRA 2 tracks the object along a path to position 9280 and 9290. This process may be variable so that as the object stops to pause, the MRAs stop as well. In this case, the MRAs have automated digital photographic capabilities with on-board auto-focus zoom lenses and data storage. The MRAs can be used to track multiple objects as well. One MRA may track the object(s) in a close in view while the other MRA(s) may track the object(s) from a distance in order to obtain a different view ofthe same scene.
[0362] Fig. 93 describes a toxic site cleanup. In this case, a static cleanup occurs within land perimeters by multiple MRAs. In the first phase, A-type MRAs (9310) are used to confine a limited amount of toxic contamination (9320) in a specific physical space. The MRAs move by using a side-to-side sweeping approach. In the second phase, the spill (9340) has been reduced and the MRS calls in the B-type MRAs (9330) in order to continue to eliminate the contamination by using a similar sweeping technique. Finally, as the toxic spill (9620) is controlled in a finite space, the MRS calls in the C-type MRAs (9350) to complete the mop up operation.
[0363] In a similar way as cleaning up toxic spills on land, Fig. 94 describes a dynamic clean up of an oil spill within limited hydro perimeters by multiple MRAs. In the first phase, the oil spill (9420) is surrounded by MRAs (9410), which operate to limit the damage and remove the oil. In the second phase, the oil spill is rendered smaller (9440) and MRAs (9430) continue to operate to remove the oil by operating in specific "cells" that act to sweep up the spill. This process continues in the final phase in which the oil spill (9470) is confined and the final drops of oil are mopped up by the MRAs (9460).
[0364] Fig. 95 describes the automated fire fighting process in which dynamic interaction occurs with a complex environment by multiple MRAs. In the first phase ofthe process, MRAs (9510) are dropped to the fire (9520) on one facade only (because the fire is initially inaccessible on the other side). As the MRAs (9530) are able to surround the fire (9540), in the second phase, they seek to put it out by using several methods, including removing brush that is flammable, by pouring fire retardant in a line around the fire and by directly pouring water on the fire. The MRAs may be air launched or ground launched and retrieved. In the final phase, the fire is reduced (9560) and the MRAs (9550) complete the task of extinguishing the fire.
[0365] Fig. 96 describes the manufacturing production process in which an object is created by using multiple MRAs. MRAs A (9610), B (9640), C (9650) and D (9630) work together to create the object (9620). One way to do this is for each MRA to attach parts of the object together from different spatial positions.
[0366] Fig. 97 shows the assembly of an object by using MRAs to combine the parts.
At an assembly facility (9710), MRA A at position Al(9720) and MRA B at position Bl (9730) act to assemble objects. Rather than having a movable assembly line, in this case, the MRAs themselves move. MRA A moves to position A2 (9740) and MRA B moves to position B2 (9750) in order to complete the assembly task. This process of organization of assembly tasks provides the opportunity for specialized functional MRAs to work together as a team in order to assemble objects by combining parts more quickly.
[0367] Roads can be built by using multiple MRAs as illustrated in Fig. 98. MRA A (9810) and MRA B (9820) proceed to create a road by laying down asphalt along adjacent tracks.
[0368] Fig. 99 describes micro surgery using MRAs for trauma intervention and stabilization, ha this case, MRA A and MRA B guide themselves to the patient. Initially, the MRAs ascertain, by using sensors, the symptoms of trauma in order to identify problems. The MRAs then move to various positions on the patient in order to solve the problems. In the case of a wound, the MRA will seek to stop the bleeding by cauterizing the wound with a laser or by applying pressure. In the case of heart. stoppage, the MRA will administer an electric shock. By stabilizing a patient, the chances of recovery are dramatically higher.
[0369] It is understood that the examples and embodiments described herein are for illustrative purposes only and that various modifications or changes in light thereof will be suggested to persons skilled in the art and are to be included within the spirit and purview of this application and scope ofthe appended claims. All publications, patents, and patent applications cited herein are hereby incorporated by reference for all purposes in their entirety.

Claims

WHAT IS CLAIMED IS:
1. A multi-robotic system architecture having a plurality of system layers interconnected to one another, comprising: a first layer including a hybrid control system for a plurality of mobile robotic agents (MRAs); a second layer including a distributed mobile robotic system for the plurality of MRAs; a third layer including a grid computing architecture in a plurality of dynamic clusters; a fourth layer including a dynamic distributed object relational database management system; a fifth layer including an omni-nodal evolutionary artificial neural network; a sixth layer including a multi-agent system and a plurality of intelligent mobile software agents; a seventh layer including a plurality of cellular automata simulations; and an eighth layer including a plurality of functional applications.
PCT/US2003/026764 2002-08-21 2003-08-21 Organizing groups of self-configurable mobile robotic agents WO2004018158A2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
AU2003262893A AU2003262893A1 (en) 2002-08-21 2003-08-21 Organizing groups of self-configurable mobile robotic agents
JP2004531235A JP2005539296A (en) 2002-08-21 2003-08-21 System, method and apparatus for an organization of self-configurable mobile robot agents within a multi-robot system
EP03793423A EP1563348A2 (en) 2002-08-21 2003-08-21 Systems, methods and apparatus for organizing groups of self-configurable mobile robotic agents in a multi-robotic system

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US40494602P 2002-08-21 2002-08-21
US40494502P 2002-08-21 2002-08-21
US60/404,945 2002-08-21
US60/404,946 2002-08-21

Publications (2)

Publication Number Publication Date
WO2004018158A2 true WO2004018158A2 (en) 2004-03-04
WO2004018158A3 WO2004018158A3 (en) 2005-06-16

Family

ID=31949877

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2003/026764 WO2004018158A2 (en) 2002-08-21 2003-08-21 Organizing groups of self-configurable mobile robotic agents

Country Status (5)

Country Link
US (3) US6904335B2 (en)
EP (1) EP1563348A2 (en)
JP (1) JP2005539296A (en)
AU (1) AU2003262893A1 (en)
WO (1) WO2004018158A2 (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006113173A1 (en) * 2005-04-14 2006-10-26 Honeywell International Inc. Decentralized maneuver control in heterogeneous autonomous vehicle networks
US7765038B2 (en) * 2005-02-16 2010-07-27 Lockheed Martin Corporation Mission planning system for vehicles with varying levels of autonomy
US8838271B2 (en) 2010-10-14 2014-09-16 Indian Institute Of Science Detection of nuclear spills using swarm optimization algorithms
CN104076800A (en) * 2014-07-10 2014-10-01 哈尔滨工程大学 UUV group task coordination system and method
US20150111591A1 (en) * 2005-10-04 2015-04-23 Steven M. Hoffberg Multifactorial optimization system and method
WO2018211927A1 (en) * 2017-05-15 2018-11-22 Omron Corporation Control apparatus, control program, learning data creation method, and learning method
CN109031959A (en) * 2018-10-26 2018-12-18 黑龙江大学 A kind of non-uniform nonlinear system cooperative control method and control system with control parameter adaptive equalization
CN109079813A (en) * 2018-08-14 2018-12-25 重庆四通都成科技发展有限公司 Automobile Marketing service robot system and its application method
CN110825088A (en) * 2019-11-29 2020-02-21 燕山大学 Multi-view vision guiding ship body cleaning robot system and cleaning method
US10943273B2 (en) 2003-02-05 2021-03-09 The Hoffberg Family Trust 2004-1 System and method for determining contingent relevance
CN115220441A (en) * 2022-03-24 2022-10-21 华东师范大学 Unmanned trolley cluster task coordination method based on biological visual perception

Families Citing this family (336)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6876991B1 (en) 1999-11-08 2005-04-05 Collaborative Decision Platforms, Llc. System, method and computer program product for a collaborative decision platform
US20020138246A1 (en) * 2001-03-08 2002-09-26 Czora Gregory J. System and method for simulating conciousness
US7328196B2 (en) * 2003-12-31 2008-02-05 Vanderbilt University Architecture for multiple interacting robot intelligences
US7305371B2 (en) * 2001-07-06 2007-12-04 Newvectors Llc Swarming agents for distributed pattern detection and classification
US8983886B2 (en) * 2012-03-28 2015-03-17 Knowmtech, Llc Self-evolvable logic fabric
WO2004018158A2 (en) * 2002-08-21 2004-03-04 Neal Solomon Organizing groups of self-configurable mobile robotic agents
AT412678B (en) * 2002-09-30 2005-05-25 Gerhard Dr Kranner METHOD FOR COMPUTER-ASSISTED PREPARATION OF PROGNOSES FOR OPERATIONAL SYSTEMS AND SYSTEM FOR CREATING PROGNOSES FOR OPERATIONAL SYSTEMS
JP4007947B2 (en) * 2002-12-20 2007-11-14 シャープ株式会社 Group robot system, sensing robot included in group robot system, base station included in group robot system, and control robot included in group robot system
US20040156388A1 (en) * 2003-02-07 2004-08-12 Lockheed Martin Corporation System for maintaining quality of service
US7716061B2 (en) * 2003-03-27 2010-05-11 International Business Machines Corporation Method and apparatus for obtaining status information in a grid
WO2004090692A2 (en) * 2003-04-04 2004-10-21 Icosystem Corporation Methods and systems for interactive evolutionary computing (iec)
US7613553B1 (en) * 2003-07-31 2009-11-03 The United States Of America As Represented By The Secretary Of The Navy Unmanned vehicle control system
US7333960B2 (en) * 2003-08-01 2008-02-19 Icosystem Corporation Methods and systems for applying genetic operators to determine system conditions
US7356518B2 (en) * 2003-08-27 2008-04-08 Icosystem Corporation Methods and systems for multi-participant interactive evolutionary computing
US7467102B2 (en) * 2003-09-11 2008-12-16 International Business Machines Corporation Request type grid computing
US7689337B2 (en) * 2003-09-30 2010-03-30 Honda Motor Co., Ltd. Cooperative vehicle control system
EP1678563A1 (en) * 2003-10-29 2006-07-12 Siemens Aktiengesellschaft Method for the operation of a technical system
US7016297B2 (en) * 2003-12-10 2006-03-21 Clive K Tang Method and apparatus providing decentralized, goal-orientated adaptive learning in an adaptive orthogonal frequency division multiplex communication system
US7529722B2 (en) * 2003-12-22 2009-05-05 Dintecom, Inc. Automatic creation of neuro-fuzzy expert system from online anlytical processing (OLAP) tools
US7400108B2 (en) * 2004-04-15 2008-07-15 University Of Utah Research Foundation System and method for controlling modular robots
EP1782285A1 (en) * 2004-07-06 2007-05-09 Icosystem Corporation Methods and apparatus for query refinement using genetic algorithms
US7707220B2 (en) * 2004-07-06 2010-04-27 Icosystem Corporation Methods and apparatus for interactive searching techniques
US20060075335A1 (en) * 2004-10-01 2006-04-06 Tekflo, Inc. Temporal visualization algorithm for recognizing and optimizing organizational structure
US7341224B1 (en) * 2004-10-14 2008-03-11 Osann Jr Robert Miniature expendable surveillance balloon system
AU2005309571A1 (en) 2004-11-23 2006-06-01 S. C. Johnson & Son, Inc. Device and methods of providing air purification in combination with cleaning of surfaces
US7340441B1 (en) * 2004-12-17 2008-03-04 The Mathworks, Inc. Search directions in pattern search via rotation
US7236861B2 (en) * 2005-02-16 2007-06-26 Lockheed Martin Corporation Mission planning system with asynchronous request capability
US8078319B2 (en) * 2005-02-16 2011-12-13 Lockheed Martin Corporation Hierarchical contingency management system for mission planners
US7353034B2 (en) 2005-04-04 2008-04-01 X One, Inc. Location sharing and tracking using mobile phones or other wireless devices
US7533071B2 (en) * 2005-06-28 2009-05-12 Neurosciences Research Foundation, Inc. Neural modeling and brain-based devices using special purpose processor
US7627540B2 (en) * 2005-06-28 2009-12-01 Neurosciences Research Foundation, Inc. Addressing scheme for neural modeling and brain-based devices using special purpose processor
US20070022061A1 (en) * 2005-07-20 2007-01-25 Bbnt Solutions Llc Optimizing solutions using local virtual forces
US7765029B2 (en) * 2005-09-13 2010-07-27 Neurosciences Research Foundation, Inc. Hybrid control device
US8423323B2 (en) 2005-09-21 2013-04-16 Icosystem Corporation System and method for aiding product design and quantifying acceptance
TWI311291B (en) * 2005-10-24 2009-06-21 Inst Information Industr Auction negotiation systems, methods and storage medium
US20070156460A1 (en) * 2005-12-29 2007-07-05 Nair Ranjit R System having a locally interacting distributed joint equilibrium-based search for policies and global policy selection
US7711663B2 (en) * 2006-03-27 2010-05-04 Board Of Trustees Of Michigan State University Multi-layer development network having in-place learning
US8893111B2 (en) 2006-03-31 2014-11-18 The Invention Science Fund I, Llc Event evaluation using extrinsic state information
CA2659672A1 (en) * 2006-06-26 2008-01-03 Icosystem Corporation Methods and systems for interactive customization of avatars and other animate or inanimate items in video games
US8271132B2 (en) * 2008-03-13 2012-09-18 Battelle Energy Alliance, Llc System and method for seamless task-directed autonomy for robots
US8355818B2 (en) * 2009-09-03 2013-01-15 Battelle Energy Alliance, Llc Robots, systems, and methods for hazard evaluation and visualization
US8965578B2 (en) 2006-07-05 2015-02-24 Battelle Energy Alliance, Llc Real time explosive hazard information sensing, processing, and communication for autonomous operation
US8073564B2 (en) * 2006-07-05 2011-12-06 Battelle Energy Alliance, Llc Multi-robot control interface
US7974738B2 (en) 2006-07-05 2011-07-05 Battelle Energy Alliance, Llc Robotics virtual rail system and method
KR100905582B1 (en) * 2006-09-06 2009-07-02 삼성전자주식회사 System and method for creating robot genome
US7672922B2 (en) * 2006-11-06 2010-03-02 Boris Kaplan Pointer-oriented object acquisition method for abstract treatment of information of AI of AI of a cyborg or an android based on a natural language
US20080270097A1 (en) * 2006-11-13 2008-10-30 Solomon Research Llc System and methods for immunocomputing applied to collectives of nanorobots
KR100823739B1 (en) * 2006-12-08 2008-04-21 한국전자통신연구원 Surroundings mapping apparatus capable of applying quickly changed surroundings information in mobile robot and method thereof
WO2008083489A1 (en) * 2007-01-12 2008-07-17 Hansjorg Baltes Method and system for robot generation
DE102007014650B3 (en) 2007-03-27 2008-06-12 Siemens Ag Method for computerized processing of data detected in sensor network, involves establishing data in sensor node by multiple adjacent sensor nodes, by which each sensor node is assigned to neural area
US8086551B2 (en) 2007-04-16 2011-12-27 Blue Oak Mountain Technologies, Inc. Electronic system with simulated sense perception and method of providing simulated sense perception
US7966093B2 (en) * 2007-04-17 2011-06-21 Yefim Zhuk Adaptive mobile robot system with knowledge-driven architecture
US8069127B2 (en) * 2007-04-26 2011-11-29 21 Ct, Inc. Method and system for solving an optimization problem with dynamic constraints
US7940177B2 (en) * 2007-06-15 2011-05-10 The Johns Hopkins University System and methods for monitoring security zones
US20080320087A1 (en) * 2007-06-22 2008-12-25 Microsoft Corporation Swarm sensing and actuating
US8160847B2 (en) * 2007-07-07 2012-04-17 Neal Solomon Hybrid multi-layer artificial immune system
WO2009038772A2 (en) 2007-09-20 2009-03-26 Evolution Robotics Transferable intelligent control device
US8548777B2 (en) * 2007-09-28 2013-10-01 Rockwell Automation Technologies, Inc. Automated recommendations from simulation
US20090089029A1 (en) * 2007-09-28 2009-04-02 Rockwell Automation Technologies, Inc. Enhanced execution speed to improve simulation performance
US20090089234A1 (en) * 2007-09-28 2009-04-02 Rockwell Automation Technologies, Inc. Automated code generation for simulators
US8069021B2 (en) * 2007-09-28 2011-11-29 Rockwell Automation Technologies, Inc. Distributed simulation and synchronization
US20090089031A1 (en) * 2007-09-28 2009-04-02 Rockwell Automation Technologies, Inc. Integrated simulation of controllers and devices
US7801710B2 (en) * 2007-09-28 2010-09-21 Rockwell Automation Technologies, Inc. Simulation controls for model variability and randomness
EP2051151B1 (en) * 2007-10-15 2011-06-29 Saab Ab Method and apparatus for generating at least one voted flight trajectory of a vehicle
US7840506B1 (en) 2008-01-09 2010-11-23 The United States Of America As Represented By The Secretary Of The Navy System and method for geodesic data mining
JP4465012B2 (en) * 2008-01-31 2010-05-19 ファナック株式会社 Machining line system that cooperates with the loading / unloading robot
JP4978494B2 (en) * 2008-02-07 2012-07-18 トヨタ自動車株式会社 Autonomous mobile body and control method thereof
US8915692B2 (en) 2008-02-21 2014-12-23 Harvest Automation, Inc. Adaptable container handling system
WO2009111062A1 (en) * 2008-03-05 2009-09-11 Brown University Improved techniques for stochastic combinatorial optimization
US8571745B2 (en) * 2008-04-10 2013-10-29 Robert Todd Pack Advanced behavior engine
JP4577422B2 (en) * 2008-07-14 2010-11-10 ソニー株式会社 Information processing system and information processing method, robot control system and control method, and computer program
KR101021836B1 (en) * 2008-10-09 2011-03-17 한국전자통신연구원 System for cooperation of multiple mobile robot using dynamic behavior binding and method thereof
US8326780B2 (en) * 2008-10-14 2012-12-04 Honda Motor Co., Ltd. Smoothed sarsa: reinforcement learning for robot delivery tasks
US9171266B2 (en) * 2008-10-15 2015-10-27 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Power provision based on self-sacrificing craft
US8250143B2 (en) 2008-12-10 2012-08-21 International Business Machines Corporation Network driven actuator mapping agent and bus and method of use
JP4730440B2 (en) * 2009-01-01 2011-07-20 ソニー株式会社 Trajectory planning apparatus, trajectory planning method, and computer program
US20100180711A1 (en) 2009-01-19 2010-07-22 Comau, Inc. Robotic end effector system and method
US9541505B2 (en) 2009-02-17 2017-01-10 The Boeing Company Automated postflight troubleshooting sensor array
US9418496B2 (en) * 2009-02-17 2016-08-16 The Boeing Company Automated postflight troubleshooting
US8364189B2 (en) * 2009-02-24 2013-01-29 Caterpillar Inc. Fleet communication network
US8812154B2 (en) * 2009-03-16 2014-08-19 The Boeing Company Autonomous inspection and maintenance
WO2010107872A2 (en) * 2009-03-17 2010-09-23 Comau, Inc. Industrial communication system and method
EP2246755A1 (en) * 2009-04-22 2010-11-03 Powitec Intelligent Technologies GmbH Control loop
DE102009018991A1 (en) * 2009-05-01 2010-11-04 Airbus Operations Gmbh Device for the spatial alignment of at least two subassembly components and method
US20100289753A1 (en) * 2009-05-14 2010-11-18 Microsoft Corporation Adjusting organization of media content on display
EP2581797B1 (en) * 2009-05-15 2021-08-18 Samsung Electronics Co., Ltd. Beacon collision avoidance method for a mobile robot system
US10188958B2 (en) 2009-05-28 2019-01-29 Anki, Inc. Automated detection of surface layout
DK2435149T3 (en) 2009-05-28 2015-09-21 Anki Inc Distributed system for autonomous control of toy cars
US9155961B2 (en) 2009-05-28 2015-10-13 Anki, Inc. Mobile agents for manipulating, moving, and/or reorienting components
US8882560B2 (en) 2009-05-28 2014-11-11 Anki, Inc. Integration of a robotic system with one or more mobile computing devices
US9046892B2 (en) * 2009-06-05 2015-06-02 The Boeing Company Supervision and control of heterogeneous autonomous operations
US20100312387A1 (en) * 2009-06-05 2010-12-09 The Boeing Company Supervision and Control of Heterogeneous Autonomous Operations
US8774970B2 (en) 2009-06-11 2014-07-08 S.C. Johnson & Son, Inc. Trainable multi-mode floor cleaning device
US9137847B1 (en) * 2009-06-29 2015-09-15 The Boeing Company Application of potential field methods in directing agents
US9131529B1 (en) 2009-06-29 2015-09-08 The Boeing Company System and method for demand driven network topology management
US8738832B1 (en) 2009-06-30 2014-05-27 International Business Machines Corporation Microgrid apparatus with central area containing plural processors communicating via respective wireless connection point or bi-directional bus with module in docking bays defined by radial arms
US7962425B1 (en) 2009-07-06 2011-06-14 International Business Machines Corporation Method and system for responding to an alert
US9154362B1 (en) 2009-08-14 2015-10-06 International Business Machines Corporation Macro grid governance and communication
US7979578B2 (en) 2009-09-02 2011-07-12 International Business Machines Corporation Dynamic and evolutionary placement in an event-driven component-oriented network data processing system
US10810519B2 (en) * 2009-10-05 2020-10-20 The Boeing Company Hierarchical mission management
US8664911B1 (en) 2009-10-30 2014-03-04 International Business Machines Corporation Micro grid bridge structures
JP5398489B2 (en) * 2009-11-20 2014-01-29 村田機械株式会社 Autonomous mobile object and its control method
US9164512B2 (en) * 2009-11-27 2015-10-20 Toyota Jidosha Kabushiki Kaisha Autonomous moving body and control method thereof
US8635015B2 (en) * 2009-12-17 2014-01-21 Deere & Company Enhanced visual landmark for localization
US8224516B2 (en) * 2009-12-17 2012-07-17 Deere & Company System and method for area coverage using sector decomposition
US20110153338A1 (en) * 2009-12-17 2011-06-23 Noel Wayne Anderson System and method for deploying portable landmarks
CA2785558C (en) * 2010-01-12 2016-05-10 Comau, Inc. Distributed control system
US8565914B2 (en) * 2010-01-18 2013-10-22 Vistaprint Technologies Limited Method and system for automatically tracking packages in automated packaging system
US8035337B1 (en) 2010-02-03 2011-10-11 International Business Machines Corporation Micro grid sensor and actuator apparatus
US8180999B1 (en) 2010-02-03 2012-05-15 International Business Machines Corporation Vertically tiered horizontally expandable micro grid system with bridge docking bay contoured by radial arms extending from central hub for coupling processor or power hubs
DE102010010875A1 (en) * 2010-03-10 2011-09-15 Siemens Aktiengesellschaft Method for monitoring the spatial environment of a movable device, in particular a medical device
US8773289B2 (en) 2010-03-24 2014-07-08 The Boeing Company Runway condition monitoring
US9305263B2 (en) 2010-06-30 2016-04-05 Microsoft Technology Licensing, Llc Combining human and machine intelligence to solve tasks with crowd sourcing
US9906838B2 (en) 2010-07-12 2018-02-27 Time Warner Cable Enterprises Llc Apparatus and methods for content delivery and message exchange across multiple content delivery networks
US8855812B2 (en) * 2010-07-23 2014-10-07 Chetan Kapoor System and method for robot safety and collision avoidance
US8599044B2 (en) 2010-08-11 2013-12-03 The Boeing Company System and method to assess and report a health of a tire
US8712634B2 (en) 2010-08-11 2014-04-29 The Boeing Company System and method to assess and report the health of landing gear related components
US8982207B2 (en) 2010-10-04 2015-03-17 The Boeing Company Automated visual inspection system
US8694296B2 (en) * 2010-10-22 2014-04-08 Agile Planet, Inc. Method and apparatus for integrated simulation
US8775341B1 (en) 2010-10-26 2014-07-08 Michael Lamport Commons Intelligent control with hierarchical stacked neural networks
US9015093B1 (en) * 2010-10-26 2015-04-21 Michael Lamport Commons Intelligent control with hierarchical stacked neural networks
KR101222624B1 (en) * 2010-12-17 2013-01-16 한국생산기술연구원 Swarm robot and sweeping method using swarm robot
US20190188122A1 (en) * 2017-12-20 2019-06-20 Rainforest Qa, Inc. Electronic product testing systems
US8856764B2 (en) * 2011-01-25 2014-10-07 International Business Machines Corporation Distributed static analysis of computer software applications
US8909762B2 (en) * 2011-03-17 2014-12-09 Hewlett-Packard Development Company, L.P. Network system management
US8935035B1 (en) * 2011-03-31 2015-01-13 The United States Of America As Represented By The Secretary Of The Army Advanced optimization framework for air-ground persistent surveillance using unmanned vehicles
US8639644B1 (en) * 2011-05-06 2014-01-28 Google Inc. Shared robot knowledge base for use with cloud computing system
US9566710B2 (en) 2011-06-02 2017-02-14 Brain Corporation Apparatus and methods for operating robotic devices using selective state space training
US8938328B2 (en) * 2011-08-08 2015-01-20 Panasonic Intellectual Property Management Co., Ltd. Electric vehicle and method of controlling the same
US9147173B2 (en) 2011-10-31 2015-09-29 Harvest Automation, Inc. Methods and systems for automated transportation of items between variable endpoints
US8775337B2 (en) 2011-12-19 2014-07-08 Microsoft Corporation Virtual sensor development
US8820678B2 (en) * 2012-01-09 2014-09-02 Google Inc. Relative positioning of balloons with altitude control and wind data
US8937410B2 (en) 2012-01-17 2015-01-20 Harvest Automation, Inc. Emergency stop method and system for autonomous mobile robots
US8874266B1 (en) * 2012-01-19 2014-10-28 Google Inc. Enhancing sensor data by coordinating and/or correlating data attributes
US8428777B1 (en) 2012-02-07 2013-04-23 Google Inc. Methods and systems for distributing tasks among robotic devices
KR20130093399A (en) * 2012-02-14 2013-08-22 한국전자통신연구원 Autonomous robot and method for controlling thereof
US8874360B2 (en) 2012-03-09 2014-10-28 Proxy Technologies Inc. Autonomous vehicle and method for coordinating the paths of multiple autonomous vehicles
US8788121B2 (en) 2012-03-09 2014-07-22 Proxy Technologies, Inc. Autonomous vehicle and method for coordinating the paths of multiple autonomous vehicles
US10054933B2 (en) * 2012-03-27 2018-08-21 Sirqul, Inc. Controlling distributed device operations
US8924011B2 (en) * 2012-04-03 2014-12-30 Knu-Industry Cooperation Foundation Intelligent robot apparatus responsive to environmental change and method of controlling and reconfiguring intelligent robot apparatus
US8447419B1 (en) 2012-05-02 2013-05-21 Ether Dynamics Corporation Pseudo-genetic meta-knowledge artificial intelligence systems and methods
US9208432B2 (en) * 2012-06-01 2015-12-08 Brain Corporation Neural network learning and collaboration apparatus and methods
US9796095B1 (en) * 2012-08-15 2017-10-24 Hanson Robokind And Intelligent Bots, Llc System and method for controlling intelligent animated characters
US9186793B1 (en) 2012-08-31 2015-11-17 Brain Corporation Apparatus and methods for controlling attention of a robot
US9117185B2 (en) 2012-09-19 2015-08-25 The Boeing Company Forestry management system
US9251698B2 (en) 2012-09-19 2016-02-02 The Boeing Company Forest sensor deployment and monitoring system
US9665101B1 (en) * 2012-09-28 2017-05-30 Waymo Llc Methods and systems for transportation to destinations by a self-driving vehicle
US20140272907A1 (en) * 2013-03-15 2014-09-18 First Principles, Inc. Method for communicating and ascertaining material
US9764468B2 (en) 2013-03-15 2017-09-19 Brain Corporation Adaptive predictor apparatus and methods
US8996177B2 (en) 2013-03-15 2015-03-31 Brain Corporation Robotic training apparatus and methods
KR20150119249A (en) * 2013-03-15 2015-10-23 에이&피 테크놀로지, 인코포레이티드 rapidly configurable braiding machine
US9733638B2 (en) 2013-04-05 2017-08-15 Symbotic, LLC Automated storage and retrieval system and control system thereof
US11334069B1 (en) 2013-04-22 2022-05-17 National Technology & Engineering Solutions Of Sandia, Llc Systems, methods and computer program products for collaborative agent control
US10168674B1 (en) * 2013-04-22 2019-01-01 National Technology & Engineering Solutions Of Sandia, Llc System and method for operator control of heterogeneous unmanned system teams
US9242372B2 (en) 2013-05-31 2016-01-26 Brain Corporation Adaptive robotic interface apparatus and methods
US9314924B1 (en) 2013-06-14 2016-04-19 Brain Corporation Predictive robotic controller apparatus and methods
US9792546B2 (en) 2013-06-14 2017-10-17 Brain Corporation Hierarchical robotic controller apparatus and methods
US9384443B2 (en) 2013-06-14 2016-07-05 Brain Corporation Robotic training apparatus and methods
US9436909B2 (en) 2013-06-19 2016-09-06 Brain Corporation Increased dynamic range artificial neuron network apparatus and methods
US9557740B2 (en) 2013-07-02 2017-01-31 David Crawley Autonomous mobile platform for service applications
US9296101B2 (en) 2013-09-27 2016-03-29 Brain Corporation Robotic control arbitration apparatus and methods
US9579789B2 (en) 2013-09-27 2017-02-28 Brain Corporation Apparatus and methods for training of robotic control arbitration
CN103529842B (en) * 2013-10-17 2016-06-29 哈尔滨工程大学 A kind of ship's fix control method based on asymptotic guiding
US9371622B2 (en) * 2013-10-22 2016-06-21 James Connor Buckley System and method to automatically determine irregular polygon for environmental hazard containment modules
US9597797B2 (en) 2013-11-01 2017-03-21 Brain Corporation Apparatus and methods for haptic training of robots
US9463571B2 (en) 2013-11-01 2016-10-11 Brian Corporation Apparatus and methods for online training of robots
US9248569B2 (en) 2013-11-22 2016-02-02 Brain Corporation Discrepancy detection apparatus and methods for machine learning
US9358685B2 (en) 2014-02-03 2016-06-07 Brain Corporation Apparatus and methods for control of robot actions based on corrective user inputs
US9201426B1 (en) 2014-02-19 2015-12-01 Google Inc. Reverse iteration of planning data for system control
US9364950B2 (en) 2014-03-13 2016-06-14 Brain Corporation Trainable modular robotic methods
US9987743B2 (en) 2014-03-13 2018-06-05 Brain Corporation Trainable modular robotic apparatus and methods
US9533413B2 (en) 2014-03-13 2017-01-03 Brain Corporation Trainable modular robotic apparatus and methods
US9446512B1 (en) * 2014-03-26 2016-09-20 Stc.Unm iAnt swarm robotic platform and evolutionary algorithms
CN103970135A (en) * 2014-04-22 2014-08-06 重庆邮电大学 Multi-mobile-robot cooperation positioning method based on filtering of MAPSO particle optimization filtering
US9346167B2 (en) 2014-04-29 2016-05-24 Brain Corporation Trainable convolutional network apparatus and methods for operating a robotic vehicle
US9958862B2 (en) 2014-05-08 2018-05-01 Yaskawa America, Inc. Intuitive motion coordinate system for controlling an industrial robot
WO2015175440A1 (en) 2014-05-12 2015-11-19 Unmanned Innovation, Inc. Unmanned aerial vehicle authorization and geofence envelope determination
US9273981B1 (en) 2014-05-12 2016-03-01 Unmanned Innovation, Inc. Distributed unmanned aerial vehicle architecture
US20150356452A1 (en) * 2014-06-10 2015-12-10 Axon Al, LLC Distributed dynamic detection of signatures from entity indicators
CN104166750B (en) * 2014-06-18 2017-07-11 南京邮电大学 Robocup based on weighting synergetic rescues collaboration method
US10754352B1 (en) * 2014-06-25 2020-08-25 Santa Clara University Multi-robot gradient based adaptive navigation system
US10363656B1 (en) * 2014-06-25 2019-07-30 Santa Clara University Multi-robot gradient based adaptive navigation system
US20160031081A1 (en) * 2014-08-01 2016-02-04 Brian David Johnson Systems and methods for the modular configuration of robots
ES2558732B2 (en) * 2014-08-05 2016-11-14 Universidad De Alicante System and method for autonomous flight planning
US9840007B1 (en) 2014-08-25 2017-12-12 X Development Llc Robotic operation libraries
US9659274B2 (en) * 2014-08-25 2017-05-23 GPS of Things, Inc. Inventory tracking and management
EP3186790A4 (en) * 2014-08-27 2019-05-01 Ponsse Oyj Method, arrangement and user interface for presenting data describing forest work unit operation
US9630318B2 (en) 2014-10-02 2017-04-25 Brain Corporation Feature detection apparatus and methods for training of robotic navigation
RS56492B1 (en) 2014-10-23 2018-01-31 Comau Spa System for monitoring and controlling an industrial plant
KR102246362B1 (en) * 2014-10-24 2021-04-28 삼성전자주식회사 Method and apparatus for wireless grid-computing
US9426946B2 (en) 2014-12-02 2016-08-30 Brain Corporation Computerized learning landscaping apparatus and methods
US9928474B1 (en) 2014-12-12 2018-03-27 Amazon Technologies, Inc. Mobile base utilizing transportation units for delivering items
WO2016106715A1 (en) * 2014-12-31 2016-07-07 SZ DJI Technology Co., Ltd. Selective processing of sensor data
US9996369B2 (en) 2015-01-05 2018-06-12 Anki, Inc. Adaptive data analytics service
US11400595B2 (en) * 2015-01-06 2022-08-02 Nexus Robotics Llc Robotic platform with area cleaning mode
US9823070B2 (en) * 2015-01-11 2017-11-21 Kenneth Dean Stephens, Jr. Remote reconnaissance for space exploration
US9717387B1 (en) 2015-02-26 2017-08-01 Brain Corporation Apparatus and methods for programming and training of robotic household appliances
CN104794281B (en) * 2015-04-22 2017-10-17 大连理工大学 A kind of libration point Spacecraft formation reconstructing method based on Adaptive proxy model
US10060056B1 (en) 2015-05-04 2018-08-28 A&P Technology, Inc. Interlocking braided structures
US9840003B2 (en) 2015-06-24 2017-12-12 Brain Corporation Apparatus and methods for safe navigation of robotic devices
US9745060B2 (en) * 2015-07-17 2017-08-29 Topcon Positioning Systems, Inc. Agricultural crop analysis drone
US10166680B2 (en) * 2015-07-31 2019-01-01 Heinz Hemken Autonomous robot using data captured from a living subject
DE102015216272A1 (en) * 2015-08-26 2017-03-02 Airbus Operations Gmbh Modular robot kit, swarm of modularized robots, and task accomplishment by a swarm of modularized robots
US10310518B2 (en) 2015-09-09 2019-06-04 Apium Inc. Swarm autopilot
US10838376B2 (en) * 2015-09-10 2020-11-17 I.Systems Automação Industrial S.A Method of generating fuzzy knowledge base for a programmable fuzzy controller
US10321663B2 (en) 2015-09-24 2019-06-18 Digi-Star, Llc Agricultural drone for use in livestock monitoring
US10231441B2 (en) 2015-09-24 2019-03-19 Digi-Star, Llc Agricultural drone for use in livestock feeding
US11072067B2 (en) * 2015-11-16 2021-07-27 Kindred Systems Inc. Systems, devices, and methods for distributed artificial neural network computation
US9760088B2 (en) 2016-02-09 2017-09-12 King Fahd University Of Petroleum And Minerals Method for cooperatively deploying robots and cooperative system of robots
US10924881B2 (en) * 2016-03-03 2021-02-16 Husqvarna Ab Device for determining construction device and worker position
US10249197B2 (en) * 2016-03-28 2019-04-02 General Electric Company Method and system for mission planning via formal verification and supervisory controller synthesis
US10314224B2 (en) 2016-03-30 2019-06-11 Autonomous Solutions, Inc. Multiple harvester planner
CN105843227B (en) * 2016-04-15 2018-10-23 上海大学 A kind of multi-robot Cooperation of task based access control closeness dynamic adjustment surrounds and seize method for allocating tasks
US10379218B1 (en) 2016-04-19 2019-08-13 Scientific Innovations, Inc. Self-locating system and methods for multistatic active coherent sonar
DE102016206781A1 (en) * 2016-04-21 2017-10-26 Kuka Roboter Gmbh A method of controlling a plurality of mobile driverless manipulator systems
US20200348662A1 (en) * 2016-05-09 2020-11-05 Strong Force Iot Portfolio 2016, Llc Platform for facilitating development of intelligence in an industrial internet of things system
WO2017197018A2 (en) 2016-05-10 2017-11-16 Trustees Of Tufts College Systems and methods enabling online one-shot learning and generalization by intelligent systems of task-relevant features and transfer to a cohort of intelligent systems
US10241514B2 (en) 2016-05-11 2019-03-26 Brain Corporation Systems and methods for initializing a robot to autonomously travel a trained route
US9987752B2 (en) 2016-06-10 2018-06-05 Brain Corporation Systems and methods for automatic detection of spills
US11092446B2 (en) 2016-06-14 2021-08-17 Motional Ad Llc Route planning for an autonomous vehicle
US10309792B2 (en) 2016-06-14 2019-06-04 nuTonomy Inc. Route planning for an autonomous vehicle
US10126136B2 (en) 2016-06-14 2018-11-13 nuTonomy Inc. Route planning for an autonomous vehicle
US10282849B2 (en) 2016-06-17 2019-05-07 Brain Corporation Systems and methods for predictive/reconstructive visual object tracker
US10016896B2 (en) 2016-06-30 2018-07-10 Brain Corporation Systems and methods for robotic behavior around moving bodies
EP3479182A4 (en) * 2016-07-01 2019-07-24 nuTonomy Inc. Affecting functions of a vehicle based on function-related information about its environment
US10829116B2 (en) 2016-07-01 2020-11-10 nuTonomy Inc. Affecting functions of a vehicle based on function-related information about its environment
US10058995B1 (en) * 2016-07-08 2018-08-28 X Development Llc Operating multiple testing robots based on robot instructions and/or environmental parameters received in a request
GB2552019B (en) * 2016-07-08 2020-01-08 Rolls Royce Plc Methods and apparatus for controlling at least one of a first robot and a second robot to collaborate within a system
US10372127B2 (en) * 2016-07-18 2019-08-06 International Business Machines Corporation Drone and drone-based system and methods for helping users assemble an object
US10216188B2 (en) * 2016-07-25 2019-02-26 Amazon Technologies, Inc. Autonomous ground vehicles based at delivery locations
US10539967B2 (en) 2016-08-23 2020-01-21 King Fahd University Of Petroleum And Minerals GPS-free robots
US10248120B1 (en) 2016-09-16 2019-04-02 Amazon Technologies, Inc. Navigable path networks for autonomous vehicles
US10245993B1 (en) 2016-09-29 2019-04-02 Amazon Technologies, Inc. Modular autonomous ground vehicles
US10241516B1 (en) 2016-09-29 2019-03-26 Amazon Technologies, Inc. Autonomous ground vehicles deployed from facilities
US10303171B1 (en) 2016-09-29 2019-05-28 Amazon Technologies, Inc. Autonomous ground vehicles providing ordered items in pickup areas
US10222798B1 (en) 2016-09-29 2019-03-05 Amazon Technologies, Inc. Autonomous ground vehicles congregating in meeting areas
CN106334283A (en) * 2016-10-10 2017-01-18 南京工程学院 Fire-fighting and rescue robot system and control method
WO2018075611A1 (en) 2016-10-18 2018-04-26 Apium Inc. Swarm communication method and device
US10987804B2 (en) * 2016-10-19 2021-04-27 Fuji Xerox Co., Ltd. Robot device and non-transitory computer readable medium
US10681513B2 (en) 2016-10-20 2020-06-09 nuTonomy Inc. Identifying a stopping place for an autonomous vehicle
US10331129B2 (en) 2016-10-20 2019-06-25 nuTonomy Inc. Identifying a stopping place for an autonomous vehicle
US10857994B2 (en) 2016-10-20 2020-12-08 Motional Ad Llc Identifying a stopping place for an autonomous vehicle
US10473470B2 (en) 2016-10-20 2019-11-12 nuTonomy Inc. Identifying a stopping place for an autonomous vehicle
US10274325B2 (en) 2016-11-01 2019-04-30 Brain Corporation Systems and methods for robotic mapping
US10001780B2 (en) 2016-11-02 2018-06-19 Brain Corporation Systems and methods for dynamic route planning in autonomous navigation
US10723018B2 (en) 2016-11-28 2020-07-28 Brain Corporation Systems and methods for remote operating and/or monitoring of a robot
US10310499B1 (en) 2016-12-23 2019-06-04 Amazon Technologies, Inc. Distributed production of items from locally sourced materials using autonomous vehicles
US10308430B1 (en) 2016-12-23 2019-06-04 Amazon Technologies, Inc. Distribution and retrieval of inventory and materials using autonomous vehicles
US10310500B1 (en) 2016-12-23 2019-06-04 Amazon Technologies, Inc. Automated access to secure facilities using autonomous vehicles
US10324466B2 (en) * 2017-01-27 2019-06-18 International Business Machines Corporation Personality sharing among drone swarm
US10377040B2 (en) 2017-02-02 2019-08-13 Brain Corporation Systems and methods for assisting a robotic apparatus
US10852730B2 (en) 2017-02-08 2020-12-01 Brain Corporation Systems and methods for robotic mobile platforms
PL3366409T3 (en) 2017-02-23 2019-12-31 Comau S.P.A. Articulated robot carrying an electric resistance welding head with electrodes located on the same side ; corresponding method of resistance electric welding on a component to be welded
US10265844B2 (en) * 2017-03-24 2019-04-23 International Business Machines Corporation Creating assembly plans based on triggering events
US10409291B2 (en) * 2017-03-27 2019-09-10 International Business Machines Corporation Teaming in swarm intelligent robot sets
US10293485B2 (en) 2017-03-30 2019-05-21 Brain Corporation Systems and methods for robotic path planning
WO2018184193A1 (en) * 2017-04-07 2018-10-11 Intel Corporation Advanced artificial intelligence agent for modeling physical interactions
US20180326581A1 (en) * 2017-05-11 2018-11-15 King Fahd University Of Petroleum And Minerals System and method for auction-based and adaptive multi-threshold multi-agent task allocation
US10802453B2 (en) 2017-06-02 2020-10-13 Bank Of America Corporation Robotics process automation macro bot
WO2019018337A1 (en) 2017-07-20 2019-01-24 Walmart Apollo, Llc Task management of autonomous product delivery vehicles
KR101924729B1 (en) * 2017-08-02 2018-12-03 서울대학교산학협력단 System and method for providing total logistic using drone
US10486305B2 (en) * 2017-09-01 2019-11-26 Robert Bosch Gmbh Method, computer program and system for controlling a plurality of robots, and computer-readable storage medium
US10102659B1 (en) 2017-09-18 2018-10-16 Nicholas T. Hariton Systems and methods for utilizing a device as a marker for augmented reality content
US10831199B2 (en) 2017-09-29 2020-11-10 Deere & Company Using unmanned aerial vehicles to provide ground control points for aerial survey
US10105601B1 (en) 2017-10-27 2018-10-23 Nicholas T. Hariton Systems and methods for rendering a virtual content object in an augmented reality environment
US20190188269A1 (en) * 2017-12-14 2019-06-20 Honeywell International Inc. Providing bots for industrial processes
DE102017223717B4 (en) * 2017-12-22 2019-07-18 Robert Bosch Gmbh Method for operating a robot in a multi-agent system, robot and multi-agent system
US11475291B2 (en) 2017-12-27 2022-10-18 X Development Llc Sharing learned information among robots
US11237877B2 (en) * 2017-12-27 2022-02-01 Intel Corporation Robot swarm propagation using virtual partitions
US10730181B1 (en) 2017-12-27 2020-08-04 X Development Llc Enhancing robot learning
US11017317B2 (en) 2017-12-27 2021-05-25 X Development Llc Evaluating robot learning
US10775774B2 (en) * 2017-12-28 2020-09-15 Intel Corporation Systems, apparatus, and methods for robot swarm coordination
CN108181928B (en) * 2017-12-29 2021-04-30 山东大学 Multi-robot formation traceability control method for sea surface oil spill
US10636188B2 (en) 2018-02-09 2020-04-28 Nicholas T. Hariton Systems and methods for utilizing a living entity as a marker for augmented reality content
US20200410377A1 (en) * 2018-03-05 2020-12-31 Trustees Of Tufts College Systems and methods for synergistic sharing of architectural components of intelligent agents
US11829886B2 (en) 2018-03-07 2023-11-28 International Business Machines Corporation Epistemic and aleatoric deep plasticity based on sound feedback
CN108710718B (en) * 2018-03-30 2022-05-17 国网安徽省电力有限公司 Mountain fire spreading simulation method and system based on cellular automaton
US10198871B1 (en) 2018-04-27 2019-02-05 Nicholas T. Hariton Systems and methods for generating and facilitating access to a personalized augmented rendering of a user
IT201800005091A1 (en) 2018-05-04 2019-11-04 "Procedure for monitoring the operating status of a processing station, its monitoring system and IT product"
DE102018207539A1 (en) * 2018-05-15 2019-11-21 Robert Bosch Gmbh Method for operating a robot in a multi-agent system, robot and multi-agent system
US11307968B2 (en) 2018-05-24 2022-04-19 The Calany Holding S. À R.L. System and method for developing, testing and deploying digital reality applications into the real world via a virtual world
US11079897B2 (en) 2018-05-24 2021-08-03 The Calany Holding S. À R.L. Two-way real-time 3D interactive operations of real-time 3D virtual objects within a real-time 3D virtual world representing the real world
CN108830483B (en) * 2018-06-15 2021-11-23 桂林电子科技大学 Task planning method for multi-agent system
US10953541B2 (en) * 2018-07-31 2021-03-23 At&T Intellectual Property I, L.P. Providing logistical support for robots
US20200167631A1 (en) * 2018-09-19 2020-05-28 Az, Llc Human-Robots: The New Specie
US11235460B2 (en) * 2018-09-24 2022-02-01 Salesforce.Com, Inc. Techniques and architectures for managing heterogeneous robots to manage deliveries
US20200110560A1 (en) * 2018-10-09 2020-04-09 Nicholas T. Hariton Systems and methods for interfacing with a non-human entity based on user interaction with an augmented reality environment
US11161245B2 (en) * 2018-10-25 2021-11-02 Wells Fargo Bank, N.A. Systems and methods for secure locker feeders
CN109343535A (en) * 2018-11-16 2019-02-15 广东嘉腾机器人自动化有限公司 A kind of control method and device of AGV trolley
TWI680364B (en) * 2018-11-27 2019-12-21 財團法人資訊工業策進會 Coach apparatus and cooperative operation controlling method for coach-driven multi-robot cooperative operation system
US11131992B2 (en) 2018-11-30 2021-09-28 Denso International America, Inc. Multi-level collaborative control system with dual neural network planning for autonomous vehicle control in a noisy environment
US11392130B1 (en) 2018-12-12 2022-07-19 Amazon Technologies, Inc. Selecting delivery modes and delivery areas using autonomous ground vehicles
CN109725532B (en) * 2018-12-24 2021-11-16 杭州电子科技大学 Relative distance control and self-adaptive correction method applied to multiple intelligent agents
US11504607B2 (en) 2019-02-05 2022-11-22 Deep Innovations Ltd. System and method for using a camera unit for the pool cleaning robot for safety monitoring and augmented reality games
DE102019203214B4 (en) * 2019-03-08 2022-01-20 Robert Bosch Gmbh Method of operating a robot in a multi-agent system, Robot and multi-agent system
US10586396B1 (en) 2019-04-30 2020-03-10 Nicholas T. Hariton Systems, methods, and storage media for conveying virtual content in an augmented reality environment
US20200359550A1 (en) * 2019-05-13 2020-11-19 Bao Tran Farm ecosystem
US11115468B2 (en) 2019-05-23 2021-09-07 The Calany Holding S. À R.L. Live management of real world via a persistent virtual world system
US11153162B2 (en) 2019-05-31 2021-10-19 Raytheon Company Communications network including intelligent network service manager
WO2020244778A1 (en) * 2019-06-07 2020-12-10 Bystronic Laser Ag Sorting system, mobile robot, method for operating a sorting system, computer program product and computer-readable medium
CN112100798A (en) 2019-06-18 2020-12-18 明日基金知识产权控股有限公司 System and method for deploying virtual copies of real-world elements into persistent virtual world systems
CN112100284A (en) 2019-06-18 2020-12-18 明日基金知识产权控股有限公司 Interacting with real world objects and corresponding databases through virtual twin reality
CN112130549A (en) * 2019-06-24 2020-12-25 上海大学 Unmanned ship formation control system with variable working modes and control method
CN110442134B (en) * 2019-08-01 2020-11-10 北京理工大学 Multi-agent cluster control method based on double-layer network
JP7308104B2 (en) * 2019-08-30 2023-07-13 三菱重工業株式会社 Unmanned aircraft cooperative system, unmanned aircraft cooperative processing method and program
US11385636B2 (en) 2019-08-30 2022-07-12 Kabushiki Kaisha Toshiba System and method for transporting inventory items in logistic facilities
US11958183B2 (en) 2019-09-19 2024-04-16 The Research Foundation For The State University Of New York Negotiation-based human-robot collaboration via augmented reality
US10796562B1 (en) 2019-09-26 2020-10-06 Amazon Technologies, Inc. Autonomous home security devices
US11260970B2 (en) 2019-09-26 2022-03-01 Amazon Technologies, Inc. Autonomous home security devices
US11182611B2 (en) 2019-10-11 2021-11-23 International Business Machines Corporation Fire detection via remote sensing and mobile sensors
US11738461B2 (en) * 2019-10-18 2023-08-29 Off-World, Inc. Systems and methods for industrial robotics
US11709121B2 (en) * 2019-11-18 2023-07-25 Spraying Systems Co. Machine learning-based particle-laden flow field characterization
US11465617B2 (en) 2019-11-19 2022-10-11 Ford Global Technologies, Llc Vehicle path planning
US11586705B2 (en) 2019-12-02 2023-02-21 International Business Machines Corporation Deep contour-correlated forecasting
CN111369038B (en) * 2020-02-25 2022-06-14 华南理工大学 Multi-target dynamic scheduling method based on genetic algorithm
US11440193B2 (en) 2020-03-30 2022-09-13 Wipro Limited Method, device, and system for managing collaboration amongst robots
CN111476998B (en) * 2020-04-13 2022-04-29 新石器慧成(北京)科技有限公司 Movable traffic command device and unmanned vehicle
BR112022023623A2 (en) * 2020-05-19 2023-02-07 Tata Consultancy Services Ltd SYSTEM AND METHOD FOR DEVELOPMENT AND IMPLEMENTATION OF SELF-ORGANISABLE CYBERPHYSICAL SYSTEMS FOR MANUFACTURING INDUSTRIES
US11436534B2 (en) 2020-05-20 2022-09-06 Bank Of America Corporation Distributed artificial intelligence model
US11727306B2 (en) 2020-05-20 2023-08-15 Bank Of America Corporation Distributed artificial intelligence model with deception nodes
US20210390402A1 (en) * 2020-06-11 2021-12-16 Sony Group Corporation Updating a neural network model on a computation device
CN112083738B (en) * 2020-09-24 2021-09-21 北京航空航天大学 Multi-agent cluster control method inspired by physics and chemistry
JP7334702B2 (en) * 2020-10-02 2023-08-29 トヨタ自動車株式会社 Service management device
US11809910B2 (en) 2020-10-14 2023-11-07 Bank Of America Corporation System and method for dynamically resizing computational infrastructure to accommodate unexpected demands
CN113033756B (en) * 2021-03-25 2022-09-16 重庆大学 Multi-agent control method based on target-oriented aggregation strategy
US11478927B1 (en) * 2021-04-01 2022-10-25 Giant.Ai, Inc. Hybrid computing architectures with specialized processors to encode/decode latent representations for controlling dynamic mechanical systems
DE112022001174T5 (en) * 2021-04-27 2024-01-18 Nvidia Corporation STRATEGY LEVELS FOR MACHINE CONTROL
CN113359626B (en) * 2021-05-21 2022-06-24 中国地质大学(武汉) Finite time hierarchical control method for multi-robot system
CN113741281A (en) * 2021-09-05 2021-12-03 西北工业大学 Method for realizing software architecture of open reconfigurable flight control system
CN113858200B (en) * 2021-09-29 2022-10-28 长春师范大学 Group robot control method for improving multi-universe inspired by foraging behavior of slime mold
CN118541309A (en) 2021-11-17 2024-08-23 H2快速帆船有限公司 Improved systems, methods, and apparatus for manufacturing airships using robotics
CN114326633B (en) * 2021-12-31 2024-04-09 四川红华实业有限公司 Automatic control method for feeding process
CN115840365A (en) * 2022-12-15 2023-03-24 江苏理工学院 PSO (particle swarm optimization) based fuzzy MRAS (fuzzy-parameter-analysis-system) permanent magnet synchronous electric spindle speed sensorless control method and control system
CN116679742B (en) * 2023-04-11 2024-04-02 中国人民解放军海军航空大学 Multi-six-degree-of-freedom aircraft collaborative combat decision-making method
CN117213501B (en) * 2023-11-09 2024-02-02 成都工业职业技术学院 Robot obstacle avoidance planning method based on distributed model prediction
CN117590751B (en) * 2023-12-28 2024-03-22 深圳市德威胜潜水工程有限公司 Underwater environment monitoring method and system based on underwater robot

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5652489A (en) 1994-08-26 1997-07-29 Minolta Co., Ltd. Mobile robot control system
US5819008A (en) 1995-10-18 1998-10-06 Rikagaku Kenkyusho Mobile robot sensor system
US5825981A (en) 1996-03-11 1998-10-20 Komatsu Ltd. Robot system and robot control device
US20010047231A1 (en) 1998-12-29 2001-11-29 Friendly Robotics Ltd. Method for operating a robot
US20020013641A1 (en) 2000-07-25 2002-01-31 Illah Nourbakhsh Socially interactive autonomous robot
US6374155B1 (en) 1999-11-24 2002-04-16 Personal Robotics, Inc. Autonomous multi-platform robot system

Family Cites Families (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5210821A (en) * 1988-03-28 1993-05-11 Nissan Motor Company Control for a group of robots
US5020411A (en) * 1989-03-06 1991-06-04 Larry Rowan Mobile assault logistic kinetmatic engagement device
US7274332B1 (en) * 1996-09-09 2007-09-25 Tracbeam Llc Multiple evaluators for evaluation of a purality of conditions
GB2337386B (en) * 1996-09-09 2001-04-04 Dennis J Dupray Location of a mobile station
US6249252B1 (en) * 1996-09-09 2001-06-19 Tracbeam Llc Wireless location using multiple location estimators
JP2000112707A (en) * 1998-04-07 2000-04-21 Hewlett Packard Co <Hp> Peripheral component monitoring device
US6415274B1 (en) * 1999-06-24 2002-07-02 Sandia Corporation Alpha-beta coordination method for collective search
JP2002160185A (en) * 2000-03-31 2002-06-04 Sony Corp Robot device, behavior controlling method for robot device and device and method for detecting external force
US6507771B2 (en) * 2000-07-10 2003-01-14 Hrl Laboratories Method and apparatus for controlling the movement of a plurality of agents
AU2001281038A1 (en) * 2000-08-01 2002-02-13 Hrl Laboratories, Llc Method and apparatus for signaling among a plurality of agents
US6687571B1 (en) * 2001-04-24 2004-02-03 Sandia Corporation Cooperating mobile robots
US6408226B1 (en) * 2001-04-24 2002-06-18 Sandia Corporation Cooperative system and method using mobile robots for testing a cooperative search controller
US6584375B2 (en) * 2001-05-04 2003-06-24 Intellibot, Llc System for a retail environment
US6636781B1 (en) * 2001-05-22 2003-10-21 University Of Southern California Distributed control and coordination of autonomous agents in a dynamic, reconfigurable system
WO2003029922A2 (en) * 2001-10-01 2003-04-10 Kline & Walker, Llc Pfn/trac system faa upgrades for accountable remote and robotics control
US7085694B2 (en) * 2001-10-22 2006-08-01 Sandia Corporation Apparatus and method for interaction phenomena with world modules in data-flow-based simulation
US7117067B2 (en) * 2002-04-16 2006-10-03 Irobot Corporation System and methods for adaptive control of robotic devices
US20040068416A1 (en) * 2002-04-22 2004-04-08 Neal Solomon System, method and apparatus for implementing a mobile sensor network
US20040030571A1 (en) * 2002-04-22 2004-02-12 Neal Solomon System, method and apparatus for automated collective mobile robotic vehicles used in remote sensing surveillance
US20040030448A1 (en) * 2002-04-22 2004-02-12 Neal Solomon System, methods and apparatus for managing external computation and sensor resources applied to mobile robotic network
US20040068415A1 (en) * 2002-04-22 2004-04-08 Neal Solomon System, methods and apparatus for coordination of and targeting for mobile robotic vehicles
US20040030570A1 (en) * 2002-04-22 2004-02-12 Neal Solomon System, methods and apparatus for leader-follower model of mobile robotic system aggregation
US20040068351A1 (en) * 2002-04-22 2004-04-08 Neal Solomon System, methods and apparatus for integrating behavior-based approach into hybrid control model for use with mobile robotic vehicles
US7047861B2 (en) * 2002-04-22 2006-05-23 Neal Solomon System, methods and apparatus for managing a weapon system
US7091852B2 (en) * 2002-07-02 2006-08-15 Tri-Sentinel, Inc. Emergency response personnel automated accountability system
WO2004018158A2 (en) * 2002-08-21 2004-03-04 Neal Solomon Organizing groups of self-configurable mobile robotic agents
DE102004026813A1 (en) * 2004-06-02 2005-12-29 Kuka Roboter Gmbh Method and device for controlling handling devices
EP1922593B1 (en) * 2005-09-06 2010-02-24 Siemens Industry, Inc. Application of microsystems for a building system employing a system knowledge base
US20080236656A1 (en) * 2006-11-13 2008-10-02 Solomon Research Llc Embedded transformable nanofilaments in multilayer crystalline space for photovoltaic cells and method of fabrication
US20080270097A1 (en) * 2006-11-13 2008-10-30 Solomon Research Llc System and methods for immunocomputing applied to collectives of nanorobots
US7921384B2 (en) * 2006-11-13 2011-04-05 Neal Solomon System, methods and apparatuses for integrated circuits for nanorobotics

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5652489A (en) 1994-08-26 1997-07-29 Minolta Co., Ltd. Mobile robot control system
US5819008A (en) 1995-10-18 1998-10-06 Rikagaku Kenkyusho Mobile robot sensor system
US5825981A (en) 1996-03-11 1998-10-20 Komatsu Ltd. Robot system and robot control device
US20010047231A1 (en) 1998-12-29 2001-11-29 Friendly Robotics Ltd. Method for operating a robot
US6374155B1 (en) 1999-11-24 2002-04-16 Personal Robotics, Inc. Autonomous multi-platform robot system
US20020013641A1 (en) 2000-07-25 2002-01-31 Illah Nourbakhsh Socially interactive autonomous robot

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11790413B2 (en) 2003-02-05 2023-10-17 Hoffberg Family Trust 2 System and method for communication
US10943273B2 (en) 2003-02-05 2021-03-09 The Hoffberg Family Trust 2004-1 System and method for determining contingent relevance
US7765038B2 (en) * 2005-02-16 2010-07-27 Lockheed Martin Corporation Mission planning system for vehicles with varying levels of autonomy
WO2006113173A1 (en) * 2005-04-14 2006-10-26 Honeywell International Inc. Decentralized maneuver control in heterogeneous autonomous vehicle networks
US20150111591A1 (en) * 2005-10-04 2015-04-23 Steven M. Hoffberg Multifactorial optimization system and method
USRE49334E1 (en) 2005-10-04 2022-12-13 Hoffberg Family Trust 2 Multifactorial optimization system and method
US9794797B2 (en) * 2005-10-04 2017-10-17 Steven M. Hoffberg Multifactorial optimization system and method
US10567975B2 (en) 2005-10-04 2020-02-18 Hoffberg Family Trust 2 Multifactorial optimization system and method
US8838271B2 (en) 2010-10-14 2014-09-16 Indian Institute Of Science Detection of nuclear spills using swarm optimization algorithms
CN104076800B (en) * 2014-07-10 2017-04-12 哈尔滨工程大学 UUV group task coordination system and method
CN104076800A (en) * 2014-07-10 2014-10-01 哈尔滨工程大学 UUV group task coordination system and method
WO2018211927A1 (en) * 2017-05-15 2018-11-22 Omron Corporation Control apparatus, control program, learning data creation method, and learning method
CN109079813A (en) * 2018-08-14 2018-12-25 重庆四通都成科技发展有限公司 Automobile Marketing service robot system and its application method
CN109031959A (en) * 2018-10-26 2018-12-18 黑龙江大学 A kind of non-uniform nonlinear system cooperative control method and control system with control parameter adaptive equalization
CN109031959B (en) * 2018-10-26 2021-09-07 黑龙江大学 Cooperative control method and system for non-uniform nonlinear system
CN110825088A (en) * 2019-11-29 2020-02-21 燕山大学 Multi-view vision guiding ship body cleaning robot system and cleaning method
CN110825088B (en) * 2019-11-29 2021-10-01 燕山大学 Multi-view vision guiding ship body cleaning robot system and cleaning method
CN115220441A (en) * 2022-03-24 2022-10-21 华东师范大学 Unmanned trolley cluster task coordination method based on biological visual perception

Also Published As

Publication number Publication date
JP2005539296A (en) 2005-12-22
US6904335B2 (en) 2005-06-07
US7343222B2 (en) 2008-03-11
AU2003262893A8 (en) 2004-03-11
US20050251291A1 (en) 2005-11-10
US20040162638A1 (en) 2004-08-19
US20100286824A1 (en) 2010-11-11
EP1563348A2 (en) 2005-08-17
AU2003262893A1 (en) 2004-03-11
WO2004018158A3 (en) 2005-06-16
US8112176B2 (en) 2012-02-07

Similar Documents

Publication Publication Date Title
US8112176B2 (en) System for self-organizing mobile robotic collectives
Albiero et al. Swarm robots in mechanized agricultural operations: A review about challenges for research
Duan et al. Bio-inspired computation in unmanned aerial vehicles
Chang et al. Coactive design of explainable agent-based task planning and deep reinforcement learning for human-UAVs teamwork
US7047861B2 (en) System, methods and apparatus for managing a weapon system
US20040030571A1 (en) System, method and apparatus for automated collective mobile robotic vehicles used in remote sensing surveillance
US20040068415A1 (en) System, methods and apparatus for coordination of and targeting for mobile robotic vehicles
US20040030570A1 (en) System, methods and apparatus for leader-follower model of mobile robotic system aggregation
US20040030448A1 (en) System, methods and apparatus for managing external computation and sensor resources applied to mobile robotic network
US20040068416A1 (en) System, method and apparatus for implementing a mobile sensor network
US20040068351A1 (en) System, methods and apparatus for integrating behavior-based approach into hybrid control model for use with mobile robotic vehicles
Zoghby et al. Chapter 8: Robot cooperation and swarm intelligence
Blais et al. Reinforcement learning for swarm robotics: An overview of applications, algorithms and simulators
Majid et al. Swarm robotics behaviors and tasks: a technical review
Puente-Castro et al. Q-learning based system for path planning with unmanned aerial vehicles swarms in obstacle environments
Mannan et al. Classical versus reinforcement learning algorithms for unmanned aerial vehicle network communication and coverage path planning: A systematic literature review
Quamar et al. Cooperative prey hunting for multi agent system designed using bio-inspired adaptation technique
Gauci Swarm robotic systems with minimal information processing
Skulimowski et al. A novel software architecture of anticipatory harvesting robot teams
Araújo et al. Cooperative observation of malicious targets in a 3d urban traffic environment using uavs
Bayona Latorre Comparative study of SAC and PPO in multi-agent reinforcement learning using unity ML-agents
Zhang et al. Learning Cooperative Policies with Graph Networks in Distributed Swarm Systems
Fontbonne Individual and group learning dynamics in evolutionary collective robotics
Lucas et al. A distributed control of movements and fuzzy logic-based task allocation for a swarm of autonomous agents
Thai What is a Swarm? A Framework for Understanding Swarms and their Applications

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 2004531235

Country of ref document: JP

WWE Wipo information: entry into national phase

Ref document number: 2003793423

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 2003793423

Country of ref document: EP