CN116394264B - Group coding impulse neural network-based multi-mechanical arm cooperative motion planning method - Google Patents

Group coding impulse neural network-based multi-mechanical arm cooperative motion planning method Download PDF

Info

Publication number
CN116394264B
CN116394264B CN202310664822.9A CN202310664822A CN116394264B CN 116394264 B CN116394264 B CN 116394264B CN 202310664822 A CN202310664822 A CN 202310664822A CN 116394264 B CN116394264 B CN 116394264B
Authority
CN
China
Prior art keywords
mechanical arm
neural network
group
mechanical
arm
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202310664822.9A
Other languages
Chinese (zh)
Other versions
CN116394264A (en
Inventor
孙长银
吴巧云
谭春雨
周云
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Anhui University
Original Assignee
Anhui University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Anhui University filed Critical Anhui University
Priority to CN202310664822.9A priority Critical patent/CN116394264B/en
Publication of CN116394264A publication Critical patent/CN116394264A/en
Application granted granted Critical
Publication of CN116394264B publication Critical patent/CN116394264B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • B25J9/1656Programme controls characterised by programming, planning systems for manipulators
    • B25J9/1664Programme controls characterised by programming, planning systems for manipulators characterised by motion, path, trajectory planning
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • B25J9/1679Programme controls characterised by the tasks executed
    • B25J9/1682Dual arm manipulator; Coordination of several manipulators
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/02Total factory control, e.g. smart factories, flexible manufacturing systems [FMS] or integrated manufacturing systems [IMS]

Landscapes

  • Engineering & Computer Science (AREA)
  • Robotics (AREA)
  • Mechanical Engineering (AREA)
  • Manipulator (AREA)
  • Image Analysis (AREA)

Abstract

The application discloses a multi-mechanical arm cooperative motion planning method based on a group coding impulse neural network, which comprises the following steps: any mechanical arm at any moment in motion sceneAnd acquiring state data of the peripheral related mechanical arms, respectively designing a group coding module, a pulse neural network module and a group decoding module, outputting the angle at which each joint of the mechanical arm should rotate at the next moment, and realizing the autonomous motion planning of the mechanical arm. According to the application, through collaborative motion planning of the multi-mechanical arm based on the group coding pulse neural network, each mechanical arm considers the motions of other mechanical arms around the mechanical arm, so that the collision between the mechanical arms can be reduced to a great extent.

Description

Group coding impulse neural network-based multi-mechanical arm cooperative motion planning method
Technical Field
The application relates to the technical field of mechanical arm motion planning, in particular to a multi-mechanical arm cooperative motion planning method based on a group coding pulse neural network.
Background
The cooperation of multiple mechanical arms means that multiple mechanical arms cooperate with each other in the same task to achieve higher efficiency and higher precision. In large scale manufacturing processes, such as automotive manufacturing, aerospace manufacturing, etc., multiple robotic arms are often required to work together to complete the various manufacturing steps. In the field of logistics and storage, the cooperative movement of multiple mechanical arms has become one of important means for improving production efficiency, reducing cost and improving quality. In addition, under some dangerous scenes, such as nuclear power stations, dangerous chemical plants and the like, the cooperation of the plurality of mechanical arms can be used for replacing manpower to finish some dangerous operations, so that the production efficiency is improved, and meanwhile, the safety of staff is also ensured.
However, the coordination and consistency of the motion trail of each robot are ensured for the cooperative work of the multiple mechanical arms, so that the task can be efficiently completed. The planning process of the motion trail is a key factor of the cooperative work of the mechanical arms. The motion planning algorithm based on sampling increases with the degree of freedom of the mechanical arm, the calculated amount of the motion planning algorithm increases exponentially, is not rapid and flexible, and lacks the capability of real-time response and dynamic planning. The learning-based motion planning algorithm needs to make trade-offs between a longer training phase and faster online execution, and is mostly focused on solving the motion problem of a single mechanical arm itself, which is insufficient to deal with the problem of co-cooperation between multiple mechanical arms. Existing learning-based multi-arm motion planning studies rely mostly on a centralized Inverse Kinematics (IK) solver, whose constraints must be learned from scratch if arms are added, removed or moved in a work scenario.
The pulsed neural network can be deployed on a neuromorphic processor, which is an energy efficient alternative to deep networks. In the emerging neuromorphic calculation model, memory and calculation are tightly integrated, and only when pulses occur, synaptic weights are added to neuron membrane potentials, so that a large number of floating point number high-precision multiplication operations in a traditional neural network are avoided, and therefore, energy consumption is extremely low in the system operation process. At present, the impulse neural network has been applied to the field of automatic driving environment sensing for detecting, classifying or tracking objects such as automobiles, pedestrians, road signs and the like, but the research of the impulse neural network on the problem of high-dimensional continuous motion planning of an industrial mechanical arm is quite limited, and is getting more attention and investment.
Disclosure of Invention
The application aims to provide a multi-mechanical arm cooperative motion planning method based on a group coding pulse neural network, which aims to solve the problems of low flexibility and high energy consumption in the conventional multi-mechanical arm motion planning.
In order to achieve the above purpose, the present application provides the following technical solutions: a multi-mechanical arm cooperative motion planning method based on a group coding impulse neural network comprises the following steps:
step one, any mechanical arm at any moment in a motion sceneAcquiring state data of related mechanical arms around the robot;
step two, designing a group coding module, and coding each dimension of the observed state data into a neuron group to obtain a pulse signal;
step three, designing a pulse neural network module to process pulse signals and output group activity pulses;
step four, designing a group decoding module, namely decoding group movement pulses into angles at which each joint of the mechanical arm should rotate at the next moment, and realizing autonomous movement planning of the mechanical arm;
and fifthly, optimizing parameters of a group coding module, a pulse neural network module and a group decoding module based on the deep reinforcement learning framework, and planning cooperative movement of a plurality of mechanical arms in a scene.
Preferably, the relevant mechanical arm is all the substrates and mechanical armsThe mechanical arm with the substrate distance within 1.0m comprises a mechanical arm per se, and is a mechanical arm +.>Is provided;
the status data of any one of the relevant mechanical arms includes: base pose, tail end pose, 10 joint positions, 6 joint angles and target tail end pose of the related mechanical arm;
the pose is a position and a pose of the European space, the position is represented by 3 dimension data, the pose is represented by a quaternion, namely 4 dimensions, and the state data of any one related mechanical arm is represented by (7+7+10×3+6+7=57) dimensions.
Preferably, step S2 further includes: for any mechanical armThe coding sequence and coding criteria of the mechanical arm related to the design.
Preferably, the coding sequence of the mechanical arm related to the design comprises the following steps:
substrate-to-robot based on related robotThe distance of the substrate is determined by the far and near sequencing related mechanical arms>Set a of robotic arms that are centered:
wherein P is the number of the mechanical arms, the value from 1 to P, P is the number of the related mechanical arms,the last mechanical arm in the set A is shown as the mechanical arm +.>Itself.
Preferably, the coding criteria for designing the mechanical arm related to the coding criteria include:
for any one relative mechanical armAt time->Status of->Is +.>Data of->,/>
Will firstTransformation into a population +.>Stimulation intensity of individual neurons:
wherein ,are trainable parameters; then based on the stimulus intensity->Generating pulse signalsThe generation rule is as follows:
wherein ,,/>is a weight coefficient, takes the value of 0.99, < >>Is an indication equation, each element of the input vector is processed to be greater than a threshold +.>When in use, thenThe corresponding element takes a value of 1, otherwise takes a value of 0, therefore, go through +.>And processing to obtain pulse signals.
Preferably, the impulse neural network module designs impulse neurons based on the circuit leakage-integration-discharge working principle.
Preferably, the impulse neuron is designed based on a circuit leakage-integration-discharge operation principle, and comprises:
firstly, determining an updating rule of any pulse neuron mode voltage as follows:
wherein ,representing the number of layers of the neural network in which the neurons are located; />Indicating the layer->A neuron; the neuron initial mode voltage is 0, namely: />
First, theLayer->The inputs to the individual impulse neurons are:
wherein n represents the thNumber of layer pulse neurons, firstThe input of the 1-layer neuron is the pulse signal output by the group coding module, (-) and>) Parameters representing neurons;
for related mechanical armsStatus data of->Layer->The output of the individual impulse neurons is:
preferably, the impulse neural network module consists of 3 impulse neural layers which are connected in sequence, and the number of impulse neurons of the impulse neural layers is 256, 128 and 10 respectively.
Preferably, the group motion pulse is weighted, and the pulse signal is decoded into a mechanical armThe real value of the angle by which the 6 joints of (a) should be turned at the next moment, i.e. +.>Angle value of +.>, wherein />Is a parameter of the group decoding module,/>Is a group activity pulse.
Preferably, in step S5, parameters of a group encoding module, a pulse neural network module and a group decoding module are optimized based on a deep reinforcement learning framework, so as to implement collaborative motion planning of a plurality of mechanical arms in a scene, including:
the strategy network in the deep reinforcement learning framework consists of a group coding module, a pulse neural network module and a group decoding module, and the value network consists of the group coding module, the pulse neural network module and a full-connection layer, wherein the output dimension of the full-connection layer is 1;
the reward functions in the training process of the deep reinforcement learning framework are:
wherein ,rewards representing the arrival of a single arm at the target end position, when the single arm arrives at the target end position,/->Otherwise->;/>Representing rewards for all the mechanical arms reaching the target end pose in the scene, and when all the mechanical arms reach the target end pose, the reward is +.>Otherwise->;/>Indicating a penalty for collision during the planning of the movement of the mechanical arm, when a collision occurs, the +.>Otherwise, let(s)>
Randomly initializing parameters of the whole reinforcement learning framework, updating the parameters by using an adaptive moment estimation optimization algorithm based on a loss function of the deep reinforcement learning framework until a reward function curve converges, and completing an online training process to obtain three optimized modules;
and deploying the optimized three modules to each mechanical arm in the scene to realize the cooperative motion planning of a plurality of mechanical arms in the scene.
Compared with the prior art, the application has the beneficial effects that:
the multi-mechanical arm cooperative motion planning method based on the group coding pulse neural network can realize the cooperative motion planning of any number of mechanical arms in any working scene, and compared with the traditional mechanical arm motion planning method based on learning, the method has the advantages that each mechanical arm considers the motions of other mechanical arms around the mechanical arm, so that the collision among the mechanical arms can be reduced to a great extent, and the cooperative motion planning of the multi-mechanical arms can be realized;
the application is based on a group coding pulse neural network, and only when pulses appear, the synaptic weight is added to the neuron membrane potential of the pulse neural network, so that a large number of floating point number high-precision multiplication operations in the traditional neural network are avoided, and the energy consumption is extremely low and the pulse neural network is easy to deploy into an embedded system in a robot arm in the running process of the system.
Drawings
FIG. 1 is a main flow chart of a multi-mechanical arm cooperative motion planning method based on a group coding impulse neural network provided by an embodiment of the application;
FIG. 2 is a diagram of a determination and mechanical arm in a multi-mechanical-arm collaborative motion planning method based on a group coding impulse neural network according to an embodiment of the present applicationSchematic diagram of a related mechanical arm;
fig. 3 is a schematic diagram of a pulse neural network module in a multi-mechanical arm cooperative motion planning method based on a group coding pulse neural network according to an embodiment of the present application;
fig. 4 is a schematic diagram of motion planning simulation of multiple mechanical arms in a multi-mechanical-arm collaborative motion planning method based on a group coding impulse neural network according to an embodiment of the present application.
Detailed Description
The following description of the embodiments of the present application will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present application, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the application without making any inventive effort, are intended to be within the scope of the application.
The main execution body of the method in this embodiment is a terminal, and the terminal may be a device such as a mobile phone, a tablet computer, a PDA, a notebook or a desktop, but of course, may be another device with a similar function, and this embodiment is not limited thereto.
Referring to fig. 1 to 4, the present application provides a multi-mechanical arm cooperative motion planning method based on a group coding impulse neural network, which is applied to multi-mechanical arm cooperative motion planning, and includes:
s1, any mechanical arm at any moment in a motion sceneAcquiring state data of related mechanical arms around the robot;
the correlation between the mechanical arms in S1 is determined by the distance between the substrates of the mechanical arms, and the correlated mechanical arms are all substrates and mechanical armsThe mechanical arm with the substrate distance within 1.0m comprises a mechanical arm per se, and is a mechanical arm +.>Is provided; referring to fig. 2, fig. 2 is a schematic diagram of a group coding impulse neural network-based multi-mechanical arm collaborative motion planning method according to an embodiment of the present applicationDetermining and arm->Schematic diagram of a related mechanical arm;
wherein, the state data of any one relevant mechanical arm comprises: base pose, tail end pose, 10 joint positions, 6 joint angles and target tail end pose of the related mechanical arm; the pose is the position and the pose of the European space, the position is represented by 3 dimension data, the pose is represented by quaternion, namely 4 dimensions, and the state data of any one related mechanical arm is represented by (7+7+10×3+6+7=57) dimensions.
S2, designing a group coding module, and coding each dimension of the observed state data into a neuron group to obtain a pulse signal;
wherein S2 further comprises: for any mechanical armThe coding sequence and coding criteria of the mechanical arm related to the design.
Specifically, the method comprises the following steps:
s21, designing the coding sequence of the mechanical arm related to the coding sequence, wherein the coding sequence comprises the following steps:
substrate-to-robot based on related robotThe distance of the substrate is determined by the far and near sequencing related mechanical arms>Set a of robotic arms that are centered:
wherein P is the number of the mechanical arms, the value from 1 to P, P is the number of the related mechanical arms,the last mechanical arm in the set A is shown as the mechanical arm +.>Itself.
S22, designing coding criteria of the mechanical arm related to the design criteria, wherein the coding criteria comprise:
for any one relative mechanical armAt time->Status of->Is +.>Data of->,/>
Will firstTransformation into a population +.>Stimulation intensity of individual neurons:
wherein ,are trainable parameters; then based on the stimulus intensity->Generating pulse signalsThe generation rule is as follows:
wherein ,,/>is a weight coefficient, takes the value of 0.99, < >>Is an indication equation, each element of the input vector is processed to be greater than a threshold +.>If so, the corresponding element takes a value of 1, otherwise takes a value of 0, thus, go through +.>And processing to obtain pulse signals.
S3, designing a pulse neural network module to process the pulse signals and output group activity pulses;
the impulse neural network module designs impulse neurons based on a circuit leakage-integration-discharge working principle.
Specifically, the method comprises the following steps:
s31, designing a pulse neuron based on a circuit leakage-integration-discharge working principle, comprising:
firstly, determining an updating rule of any pulse neuron mode voltage as follows:
wherein ,representing the number of layers of the neural network in which the neurons are located; />Indicating the layer->A neuron; the neuron initial mode voltage is 0, namely: />
First, theLayer->The inputs to the individual impulse neurons are:
wherein n represents the thThe number of the layer pulse neurons, namely the input of the 1 st layer neuron is the pulse signal output by the group coding module, (-) and%>) Parameters representing neurons;
for related mechanical armsStatus data of->Layer->The output of the individual impulse neurons is:
s32, constructing a pulse neural network module comprises the following steps:
the impulse neural network module consists of 3 impulse neural layers which are sequentially connected, and comprises a first impulse neural layer, a second impulse neural layer and a third impulse neural layer, wherein the number of impulse neurons of each impulse neural layer is 256, 128 and 10 respectively in sequence, please refer to fig. 3, and fig. 3 is a schematic diagram of the impulse neural network module in the multi-mechanical arm cooperative motion planning method based on the group coding impulse neural network provided by the embodiment of the application.
S4, designing a group decoding module to decode group activity pulses into mechanical armsThe angle that each joint should rotate at the next moment realizes the autonomous motion planning of the mechanical arm;
specifically, S4 includes: weighting group activity pulse, decoding pulse signal into mechanical armThe real value of the angle by which the 6 joints of (a) should be turned at the next moment, i.e. +.>Angle value of (2), wherein />Is a parameter of the group decoding module,/>Is a group activity pulse.
And S5, optimizing parameters of a group coding module, a pulse neural network module and a group decoding module based on a deep reinforcement learning framework (SAC), and planning cooperative movement of a plurality of mechanical arms in a scene.
Specifically, S5 includes:
s51, a strategy network in a deep reinforcement learning framework (SAC) consists of a group coding module, a pulse neural network module and a group decoding module, wherein the value network consists of the group coding module, the pulse neural network module and a full-connection layer, and the output dimension of the full-connection layer is 1;
s52, rewarding functions in the training process of the deep reinforcement learning frame (SAC) are as follows:
wherein ,rewards representing the arrival of a single arm at the target end position, when the single arm arrives at the target end position,/->Otherwise->;/>Representing rewards for all the mechanical arms reaching the target end pose in the scene, and when all the mechanical arms reach the target end pose, the reward is +.>Otherwise->;/>Indicating a penalty for collision during the planning of the movement of the mechanical arm, when a collision occurs, the +.>Otherwise, let(s)>
S53, randomly initializing parameters of the whole reinforcement learning frame (SAC), updating the parameters by using an adaptive moment estimation optimization algorithm (Adam) based on a loss function of the learning frame (SAC) until a reward function curve converges, and completing an online training process to obtain three optimized modules;
and S54, deploying the optimized three modules on each mechanical arm in the scene to realize the cooperative motion planning of a plurality of mechanical arms in the scene, wherein as shown in fig. 4, fig. 4 is a simulation schematic diagram of the motion planning of the three mechanical arms in the multi-mechanical-arm cooperative motion planning method based on the group coding pulse neural network, and the numbers of the mechanical arms are represented in the figure.
In this embodiment, each mechanical arm considers the movements of other mechanical arms around the mechanical arm, so that the collision between the mechanical arms can be reduced to a great extent, and thus the collaborative motion planning of multiple mechanical arms can be realized.
In a multi-mechanical-arm working scene, the target end pose of each mechanical arm is known, and by adopting the multi-mechanical-arm cooperative motion planning method based on the group coding pulse neural network, each mechanical arm in the scene can avoid mechanical arms of other motions, smoothly reaches the target pose, and the energy consumption in the whole motion planning process is lower. The multi-mechanical-arm cooperative motion planning method based on the group coding pulse neural network enables any number of mechanical arms in a working scene to reach a target pose under the condition of no collision, realizes co-cooperation among the multi-mechanical arms and ensures lower system energy consumption.
On the basis of the embodiment, the application further provides electronic equipment, which comprises:
the device comprises a processor and a memory, wherein the processor is in communication connection with the memory;
in this embodiment, the memory may be implemented in any suitable manner, for example: the memory can be read-only memory, mechanical hard disk, solid state hard disk, USB flash disk or the like; the memory is used for storing executable instructions executed by at least one processor;
in this embodiment, the processor may be implemented in any suitable manner, e.g., the processor may take the form of, for example, a microprocessor or processor and a computer-readable medium storing computer-readable program code (e.g., software or firmware) executable by the (micro) processor, logic gates, switches, application specific integrated circuits (Application Specific Integrated Circuit, ASIC), a programmable logic controller, and an embedded microcontroller, etc.; the processor is configured to execute executable instructions to implement a multi-manipulator collaborative motion planning method based on a group-encoded impulse neural network as described above.
On the basis of the embodiment, the application also provides a computer readable storage medium, wherein a computer program is stored in the computer readable storage medium, and the computer program realizes the multi-mechanical arm cooperative motion planning method based on the group coding pulse neural network when being executed by a processor.
Those of ordinary skill in the art will appreciate that the modules and method steps of the examples described in connection with the embodiments disclosed herein can be implemented as electronic hardware, or as a combination of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the solution. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application.
It will be clearly understood by those skilled in the art that, for convenience and brevity of description, specific working procedures of the apparatus, device and module described above may refer to corresponding procedures in the foregoing method embodiments, which are not repeated herein.
In the several embodiments provided by the present application, it should be understood that the disclosed apparatus, device and method may be implemented in other manners. For example, the apparatus embodiments described above are merely illustrative, e.g., the division of modules is merely a logical function division, and there may be additional divisions of actual implementation, e.g., multiple modules or units may be combined or integrated into another apparatus, or some features may be omitted, or not performed. Alternatively, the coupling or direct coupling or communication connection shown or discussed with each other may be an indirect coupling or communication connection via some interfaces, devices or apparatuses, which may be in electrical, mechanical or other form.
The modules illustrated as separate components may or may not be physically separate, and components shown as modules may or may not be physical units, may be located in one place, or may be distributed over multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
In addition, each functional module in each embodiment of the present application may be integrated into one processing module, or each module may exist alone physically, or two or more modules may be integrated into one module.
The functions, if implemented in the form of software functional modules and sold or used as a stand-alone product, may be stored on a computer readable storage medium. Based on this understanding, the technical solution of the present application may be embodied essentially or in a part contributing to the prior art or in a part of the technical solution in the form of a software product stored in a storage medium, comprising several instructions for causing a computer device (which may be a personal computer, a server, a network device, or the like) to perform all or part of the steps of the method according to the embodiments of the present application. And the aforementioned storage medium includes: a usb disk, a removable hard disk, a read-only memory server, a random access memory server, a magnetic disk or an optical disk, or other various media capable of storing program instructions.
In addition, it should be noted that the combination of the technical features described in the present application is not limited to the combination described in the claims or the combination described in the specific embodiments, and all the technical features described in the present application may be freely combined or combined in any manner unless contradiction occurs between them.
It should be noted that the above-mentioned embodiments are merely examples of the present application, and it is obvious that the present application is not limited to the above-mentioned embodiments, and many similar variations are possible. All modifications attainable or obvious from the present disclosure set forth herein should be deemed to be within the scope of the present disclosure.
The foregoing is merely illustrative of the preferred embodiments of the present application, and is not intended to limit the scope of the present application. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present application should be included in the protection scope of the present application.

Claims (8)

1. A multi-mechanical arm cooperative motion planning method based on a group coding impulse neural network is characterized by comprising the following steps:
s1, any mechanical arm at any moment in a motion scene k Acquiring state data of related mechanical arms around the robot;
s2, designing a group coding module, and coding each dimension of the observed state data into a neuron group to obtain a pulse signal;
s3, designing a pulse neural network module to process the pulse signals and output group activity pulses;
the impulse neural network module designs impulse neurons based on a circuit leakage-integration-discharge working principle, and specifically comprises the following components:
firstly, determining an updating rule of any pulse neuron mode voltage as follows:
wherein, P is the number of the mechanical arms, the value from 1 to P, and P is the number of the related mechanical arms; θ (·) is an indication equation, processing each element of the input vector, greater than a threshold V th When the corresponding element takes a value of 1, otherwise, takes a value of 0; v (V) th Is the pulse neuron mode voltage threshold; l represents the number of layers of the neural network in which the neurons are located; m represents the mth neuron of the layer; the neuron initial mode voltage is 0, namely:
the inputs to the mth impulse neuron of layer l are:
wherein n represents the number of the first-1 layer pulse neurons, the input of the first layer neurons is the pulse signal output by the group coding module,parameters representing neurons;
arm for related mechanical arm (p) The output of the mth pulse neuron of the first layer is:
s4, designing a group decoding module, and decoding group activity pulses into angles at which each joint of the mechanical arm should rotate at the next moment, so as to realize autonomous motion planning of the mechanical arm;
and S5, optimizing parameters of a group coding module, a pulse neural network module and a group decoding module based on the deep reinforcement learning framework, and planning cooperative movement of a plurality of mechanical arms in a scene.
2. Group-encoded impulse-based neural network of claim 1The multi-mechanical arm cooperative motion planning method is characterized in that the related mechanical arms are all substrates and mechanical arm k A robotic arm having a base distance in the range of 1.0m and including the robotic arm itself;
the status data of any one of the relevant mechanical arms includes: base pose, tail end pose, 10 joint positions, 6 joint angles and target tail end pose of the related mechanical arm;
the pose is a position and a pose of the European space, the position is represented by 3 dimension data, the pose is represented by a quaternion, namely 4 dimensions, and the state data of any one related mechanical arm is represented by 7+7+10×3+6+7=57 dimensions.
3. The method for planning the coordinated motion of multiple mechanical arms based on the group-encoded impulse neural network according to claim 1, wherein in the step S2, the method further comprises: arm for any mechanical arm k The coding sequence and coding criteria of the mechanical arm related to the design.
4. A multi-robot collaborative motion planning method based on a group coded impulse neural network according to claim 3, wherein the design and coding sequence of its associated robots comprises:
based on the distance from the base of the relevant mechanical arm to the base of the mechanical arm, the relevant mechanical arm is sequenced from far to near to obtain arm k Set a of robotic arms that are centered:
wherein P is the number of the mechanical arms, and can be from 1 to P, P is the number of the relevant mechanical arms, arm (p) Representing the last mechanical arm in the set A as the arm k Itself.
5. A multi-robot collaborative motion planning method based on a group coded impulse neural network according to claim 3, wherein the design and coding criteria of its associated robot comprises:
for any one relative mechanical arm (p) E A at time t i State of (2)Data of any one dimension j
Will firstConversion to stimulation intensity of L neurons in a population:
wherein ,are trainable parameters; then based on the stimulus intensity->Generating a pulse signal +.>The generation rule is as follows:
wherein ,alpha is a weight coefficient, the value is 0.99, theta (·) is an indication equation, each element of the input vector is processed to be greater than a threshold value V th When the pulse signal is processed, the corresponding element takes a value of 1, otherwise takes a value of 0, so that the pulse signal can be obtained through theta (·) processing.
6. The group coding impulse neural network-based multi-mechanical arm collaborative motion planning method according to claim 1, wherein the impulse neural network module is composed of 3 impulse neural layers which are connected in sequence, and the number of impulse neurons of the impulse neural layers is 256, 128 and 10 respectively in sequence.
7. The method for planning the coordinated motion of multiple mechanical arms based on the group-encoded impulse neural network according to claim 1, wherein the group-active impulses are weighted to decode impulse signals into mechanical arm k The real angle value of the 6 joints to be rotated at the next moment, i.e. the angle value a for any joint q q =W q ·δ+b q, wherein (Wq ,b q ) Is a parameter of the group decoding module, and δ is the group activity pulse.
8. The method for planning the coordinated motion of multiple mechanical arms based on the group coding impulse neural network according to claim 1, wherein the optimizing parameters of the group coding module, the impulse neural network module and the group decoding module based on the deep reinforcement learning framework in the step S5, to implement the coordinated motion planning of multiple mechanical arms in a scene, includes:
the strategy network in the deep reinforcement learning framework consists of a group coding module, a pulse neural network module and a group decoding module, and the value network consists of the group coding module, the pulse neural network module and a full-connection layer, wherein the output dimension of the full-connection layer is 1;
the reward function in the training process of the deep reinforcement learning framework is as follows:
r t =r g +r all +r c
wherein ,rg Representing rewards for reaching a target end pose by a single robotic arm, r when the single robotic arm reaches the target end pose g =0.01, otherwise r g =0;r all Representing rewards for all mechanical arms reaching the target end pose in a scene, and r is when all mechanical arms reach the target end pose all =1.0, otherwise r all =0;r c Indicating punishment of collision in the process of planning movement of mechanical arm, and when collision occurs, r c = -0.05, otherwise, r c =0;
Randomly initializing parameters of the whole reinforcement learning framework, updating the parameters by using an adaptive moment estimation optimization algorithm based on a loss function of the deep reinforcement learning framework until a reward function curve converges, and completing an online training process to obtain three optimized modules;
and deploying the optimized three modules to each mechanical arm in the scene to realize the cooperative motion planning of a plurality of mechanical arms in the scene.
CN202310664822.9A 2023-06-07 2023-06-07 Group coding impulse neural network-based multi-mechanical arm cooperative motion planning method Active CN116394264B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310664822.9A CN116394264B (en) 2023-06-07 2023-06-07 Group coding impulse neural network-based multi-mechanical arm cooperative motion planning method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310664822.9A CN116394264B (en) 2023-06-07 2023-06-07 Group coding impulse neural network-based multi-mechanical arm cooperative motion planning method

Publications (2)

Publication Number Publication Date
CN116394264A CN116394264A (en) 2023-07-07
CN116394264B true CN116394264B (en) 2023-08-25

Family

ID=87009048

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310664822.9A Active CN116394264B (en) 2023-06-07 2023-06-07 Group coding impulse neural network-based multi-mechanical arm cooperative motion planning method

Country Status (1)

Country Link
CN (1) CN116394264B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116989800B (en) * 2023-09-27 2023-12-15 安徽大学 Mobile robot visual navigation decision-making method based on pulse reinforcement learning

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111612136A (en) * 2020-05-25 2020-09-01 之江实验室 Neural morphology visual target classification method and system
CN111923039A (en) * 2020-07-14 2020-11-13 西北工业大学 Redundant mechanical arm path planning method based on reinforcement learning
CN112699956A (en) * 2021-01-08 2021-04-23 西安交通大学 Neural morphology visual target classification method based on improved impulse neural network
CN114372563A (en) * 2022-01-10 2022-04-19 四川大学 Robot control method and system based on hybrid pulse reinforcement learning network structure
CN115438776A (en) * 2022-10-10 2022-12-06 西安交通大学 Coding method for impulse neural network
CN115870988A (en) * 2022-12-29 2023-03-31 西安交通大学 Multi-agent cooperative control mechanical arm automatic filling motion planning method
CN115908896A (en) * 2022-10-28 2023-04-04 北京理工大学 Image identification system based on impulse neural network with self-attention mechanism

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180137408A1 (en) * 2015-05-22 2018-05-17 Socopra Sciences Et Genie S.E.C. Method and system for event-based neural networks

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111612136A (en) * 2020-05-25 2020-09-01 之江实验室 Neural morphology visual target classification method and system
CN111923039A (en) * 2020-07-14 2020-11-13 西北工业大学 Redundant mechanical arm path planning method based on reinforcement learning
CN112699956A (en) * 2021-01-08 2021-04-23 西安交通大学 Neural morphology visual target classification method based on improved impulse neural network
CN114372563A (en) * 2022-01-10 2022-04-19 四川大学 Robot control method and system based on hybrid pulse reinforcement learning network structure
CN115438776A (en) * 2022-10-10 2022-12-06 西安交通大学 Coding method for impulse neural network
CN115908896A (en) * 2022-10-28 2023-04-04 北京理工大学 Image identification system based on impulse neural network with self-attention mechanism
CN115870988A (en) * 2022-12-29 2023-03-31 西安交通大学 Multi-agent cooperative control mechanical arm automatic filling motion planning method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于脉冲神经网络的目标检测技术研究;韩雯慧;《基于脉冲神经网络的目标检测技术研究》;全文 *

Also Published As

Publication number Publication date
CN116394264A (en) 2023-07-07

Similar Documents

Publication Publication Date Title
Karkus et al. Qmdp-net: Deep learning for planning under partial observability
US9367798B2 (en) Spiking neuron network adaptive control apparatus and methods
US9189730B1 (en) Modulated stochasticity spiking neuron network controller apparatus and methods
WO2019091020A1 (en) Weight data storage method, and neural network processor based on method
US20200139540A1 (en) Reduced degree of freedom robotic controller apparatus and methods
US9082079B1 (en) Proportional-integral-derivative controller effecting expansion kernels comprising a plurality of spiking neurons associated with a plurality of receptive fields
US9008840B1 (en) Apparatus and methods for reinforcement-guided supervised learning
US20170203437A1 (en) Apparatus and methods for operating robotic devices using selective state space training
CN116394264B (en) Group coding impulse neural network-based multi-mechanical arm cooperative motion planning method
US20150127149A1 (en) Apparatus and methods for online training of robots
WO2013184688A1 (en) Stochastic apparatus and methods for implementing generalized learning rules
Jalali et al. Optimal autonomous driving through deep imitation learning and neuroevolution
Chao et al. Generative adversarial nets in robotic Chinese calligraphy
US11279025B2 (en) Apparatus and methods for operating robotic devices using selective state space training
Vuković et al. Trajectory learning and reproduction for differential drive mobile robots based on GMM/HMM and dynamic time warping using learning from demonstration framework
CN114169394A (en) Multi-variable time series prediction method for multi-scale adaptive graph learning
Fan et al. Learning resilient behaviors for navigation under uncertainty
CN114077258A (en) Unmanned ship pose control method based on reinforcement learning PPO2 algorithm
Zhang et al. Auto-conditioned recurrent mixture density networks for learning generalizable robot skills
CN113743287B (en) Robot self-adaptive grabbing control method and system based on impulse neural network
Liu et al. Robotic cognitive behavior control based on biology-inspired episodic memory
Sykora et al. Artificial Neural Networks in Educational Process
Murata et al. Learning and recognition of multiple fluctuating temporal patterns using S-CTRNN
Yamada et al. Attractor representations of language-behavior structure in a recurrent neural network for human-robot interaction
Li et al. Achieving different levels of adaptability for human robot collaboration utilizing a neuro-dynamical system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant