CN101685309A - Method for controlling multi-robot coordinated formation - Google Patents
Method for controlling multi-robot coordinated formation Download PDFInfo
- Publication number
- CN101685309A CN101685309A CN200810222774A CN200810222774A CN101685309A CN 101685309 A CN101685309 A CN 101685309A CN 200810222774 A CN200810222774 A CN 200810222774A CN 200810222774 A CN200810222774 A CN 200810222774A CN 101685309 A CN101685309 A CN 101685309A
- Authority
- CN
- China
- Prior art keywords
- robot
- pilot
- sampling instant
- pmax
- follow
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000015572 biosynthetic process Effects 0.000 title claims abstract description 46
- 238000000034 method Methods 0.000 title claims abstract description 37
- 238000005070 sampling Methods 0.000 claims abstract description 50
- 238000004891 communication Methods 0.000 claims abstract description 8
- 238000006243 chemical reaction Methods 0.000 claims abstract description 4
- 230000000007 visual effect Effects 0.000 claims description 7
- 239000003086 colorant Substances 0.000 claims description 3
- 238000010586 diagram Methods 0.000 description 3
- 230000000694 effects Effects 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 238000009825 accumulation Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000004904 shortening Methods 0.000 description 1
Images
Abstract
The invention provides a method for controlling multi-robot coordinated formation, which is formation control realized by coordinating an accompanying robot with a pilot robot through recurrence predication. The accompanying robot acquires distances and observed azimuths d <k-s>, phi <k-s> (s=0,..., Np max relatively to the pilot robot at current sampling time k and previous Np max times of sampling time by combination with vision and coded disc information; reoccurring to acquire the position L<K>(xL<k>, yL<k>) (see in formula 2) of the pilot robot under a local coordinate system (see formula1) of the accompanying robot at the sampling time k and k-Np max and the position Of<k> (xof<k>, yof<k>) of a photocenter Copt (see formula 3) of a forward camera of the accompanying robot at the sampling time k through conversion of the local coordinate system by means of coded disc information; calculating an angle psiij<k> between the predicted moving direction of the pilot robot and the direction from the pilot robot to the Copt so as to combine an expected angle psiijd, distance Lij<k> (namely d<k>) between the accompanying robot and the pilot robot and the expected distance Lijd, thus combining vision information, and carrying out motion decision to acquire speed vf and a turn angle thetaf. The method is suitable for conditions of communication failures, and provides technical support for application of multi-robot in military and other fields.
Description
Technical field
The present invention relates to the Robotics field, particularly a kind of method for controlling multi-robot coordinated formation is used for multi-robot system.
Background technology
The control of multirobot formation, the team that refers to a plurality of robots composition under the condition of constraint that conforms, keeps predetermined geometric relationship each other in the process of executing the task.The multirobot formation has broad application prospects at aspects such as military affairs, Aero-Space, and its realization helps shortening the task executions time, improves the overall performance of system.Formation control method commonly used has pilotage people's follower method, based on the method and the virtual architecture method of behavior, a lot of research work are devoted to these methods are studied and improved to improve system's operation stability and environmental suitability.
For a system that forms by N robot, a kind of thinking is that the formation control problem that formation control problem is decomposed into two robots of N-1 group is realized, one in two robots of every group are pilot robot for following robot, one, and they are devoted to keep predefined geometric configuration.By the formation control in twos of robot in the system, present the shape of coordination on the whole.In order to realize formation control, follow the relevant information that robot need obtain pilot robot, the means of the information of obtaining commonly used are communications.Consider the problems such as error accumulation of communication network delay, data-bag lost and even communication failures and topworks, design a kind ofly do not have communication, the method by the perceptual inference of its pilot robot being realized coordinate formation control is significant.
Summary of the invention
The purpose of this invention is to provide a kind of method for controlling multi-robot coordinated formation, be based on the method for perceptual inference, make that following robot relies on perceptual inference to realize and the formation control of pilot robot, reaches gratifying effect.
For achieving the above object, technical solution of the present invention is:
A kind of method for controlling multi-robot coordinated formation comprises step:
A) follow robot according to the colour code tube on the pilot robot,, obtain current sampling instant k and N before in conjunction with vision and code-disc information
PmaxInferior sampling instant is with respect to the distance and the observed azimuth d of pilot robot
K-s,
B) follow robot by code-disc information, by the local coordinate system conversion, recursion obtains sampling instant k and k-N
P maxThe time pilot robot following the robot local coordinate system
Under position L
k(x
L k, y
L k),
And the photocentre C that follows robot forward direction video camera during current sampling instant k
Opt Position O under the coordinate system
f k(x
Of k, y
Of k);
C) position-based L
k,
And O
f k, follow robot and calculate the pilot robot direction of motion of prediction and pilot robot to C
OptAngle ψ between the direction
Ij k
D) in conjunction with angle ψ
Ij kAnd expected angle ψ
Ijd, follow the distance L between robot and the pilot robot
Ij k(be d
k) and desired distance L
Ijd, comprehensive visual information, the decision-making of moving obtains speed v
fAnd rotational angle theta
f, realize following the control of robot speed's size and Orientation.
Described method for controlling multi-robot coordinated formation, the colour code tube on its pilot robot is the cylindrical tube of hollow, is combined up and down by at least two kinds of colors, is put in the robot, the center of its center and robot is consistent.
Described method for controlling multi-robot coordinated formation, the control of its formation are to realize there is not communication between the robot on the basis of robot to the pilot robot perceptual inference following.
Described method for controlling multi-robot coordinated formation, its described B) follows the robot local coordinate system in the step
Be and sampling instant k-N
PmaxCorresponding, its initial point
Get forward direction video camera photocentre C
OptThe position at place,
Direction and sampling instant k-N
PmaxThe time robot motion's direction be consistent.
Described method for controlling multi-robot coordinated formation, its described C) following the prediction of robot to pilot robot direction of motion in the step, is according to current sampling instant k and this moment N before
PmaxThe pilot robot positional information of inferior sampling instant draws, and the pilot robot positional information of these two sampling instants all is based upon the N that follows before the sampling instant k of robot
PmaxUnder the pairing local coordinate system of inferior sampling instant.
Described method for controlling multi-robot coordinated formation, its described D) follows distance L between robot and the pilot robot in the step
Ij kBe d
k
Described method for controlling multi-robot coordinated formation, its described N
Pmax〉=1.
Method for controlling multi-robot coordinated formation of the present invention, the situation of suitable communication failures is for the application of multirobot at aspects such as military affairs provides technical support.
Description of drawings
Fig. 1 is the control block diagram of a kind of method for controlling multi-robot coordinated formation of the present invention;
Fig. 2 is two robot formation control charts;
Fig. 3 is the recursion prognostic chart of a kind of method for controlling multi-robot coordinated formation of the present invention;
Fig. 4 be a kind of method for controlling multi-robot coordinated formation of the present invention follow robot motion's decision diagram;
Fig. 5 is with pilot robot of the inventive method and the movement locus of following robot.
Embodiment
The invention provides a kind of method for controlling multi-robot coordinated formation.Follow robot in conjunction with vision and self code-disc information, obtain current sampling instant k and N before
PmaxInferior sampling instant is with respect to the distance and the observed bearing angle information of pilot robot.By code-disc information, by the local coordinate system conversion, recursion obtains the N before sampling instant k and this moment
PmaxPilot robot during inferior sampling instant is being followed the robot local coordinate system (with sampling instant k-N
PmaxCorresponding) under the position, and the position of photocentre under this local coordinate system of following robot forward direction video camera (its optical axis direction is consistent with robot motion's direction) during current sampling instant k.Based on above-mentioned position, follow robot and calculate the pilot robot direction of motion of prediction and pilot robot to the angle between the photocentre direction of following robot forward direction video camera, and then in conjunction with pilot robot direction of motion and pilot robot to the expected angle between the photocentre direction of following robot forward direction video camera and follow robot and pilot robot between distance and desired distance, comprehensive visual information, the decision-making of moving, obtain following the speed and the corner of robot, realize following the control of robot speed's size and Orientation.The control block diagram of method for controlling multi-robot coordinated formation as shown in Figure 1, wherein, d
K-s,
Be current sampling instant k and N before
PmaxInferior sampling instant is with respect to the distance and the observed azimuth of pilot robot, L
k(x
L k, y
L k) and
Be respectively sampling instant k and k-N
PmaxThe time pilot robot exist
Coordinate system (is followed robot and sampling instant k-N
PmaxCorresponding local coordinate system) position under, O
f k(x
Of k, y
Of k) the photocentre of following robot forward direction video camera during for current sampling instant k exists
Position under the coordinate system, ψ
Ij kAnd ψ
IjdBe respectively the pilot robot direction of motion of prediction and pilot robot to angle and expected angle between the photocentre direction of following robot forward direction video camera, L
Ij k(be d
k) and L
IjdBe respectively the distance and the desired distance of following between robot and the pilot robot, v
fAnd θ
fBe respectively the speed and the corner of following robot.
Fig. 2 is two robot formation control chart, wherein R
iAnd R
jRobot and pilot robot are followed in expression respectively, and l is the distance between the robot two driving wheel centers, and XOY is a world coordinate system, X
fO
fY
fBe selected as R
iLocal coordinate system, O
fGet forward direction video camera C
fPhotocentre C
OptThe position at place, d
cBe O
fAnd the distance between the robot center, Y
fDirection and robot motion's direction be consistent.Work as R
iRecursion dopes pilot robot R
jMotion the time, based on L
Ijdψ
Ijd, obtain expecting formation point P
Fd, along with R
iMarch on towards P
Fd, L
Ij→ L
Ijd, ψ
Ij→ ψ
Ijd, finally set up a kind of coordination formation relation, wherein L
Ij, L
IjdBe respectively R
iWith R
jBetween distance and desired distance, ψ
Ij(ψ
Ij∈ [0,2 π)), ψ
IjdBe respectively R
jDirection of motion and R
jTo R
iAngle and expected angle between the photocentre direction of forward direction video camera.
1, pilot robot location estimation:
Pilot robot is equipped with a colour code tube, and it is the cylindrical tube of a hollow, is combined up and down by at least two kinds of colors, and the colour code tube is put on the pilot robot, and its center and robot center are consistent.Follow robot by the visual identity of colour code tube being finished identification,, and then obtain the relative positioning information of estimation in conjunction with vision calibration to pilot robot.Make (d
k,
) be distance and the observed azimuth of sampling instant k with respect to pilot robot.
The definition (u v) is that image coordinate is fastened a bit, (x, y, z) be its former be the coordinate system w of initial point to the video camera photocentre
CameraOn coordinate, have:
Wherein M is the intrinsic parameter matrix of video camera, is obtained by camera calibration.
Note (u
1, v
1) and (u
2, v
2) be respectively the point of colour code tube top and bottom, (x
1, y
1, z
1) and (x
2, y
2, z
2) be that they are at w
CameraOn coordinate, u
1=u
2, z
1=z
2, y
2-y
1Be the true altitude of colour code tube, then d
kCan estimate as follows:
α in the formula
yObtain v by M
2-v
1Be that the pixels tall of colour code tube in image is poor.
Note (u
T, v
T) be the center of colour code tube, u
d, θ
vBe respectively the width of image and the width in the visual field, then
Can estimate as follows:
Influenced by the external environment condition or the visual field etc., follow robot and utilize forward direction video camera C
fMay cannot see pilot robot, need estimate the pilot robot position according to code-disc and former visual information this moment.If follow robot through continuous N
CsInferior sampling does not still observe pilot robot, abandons estimating; See pilot robot in case follow robot by other video camera, its overriding concern be to make the forward direction video camera find pilot robot as early as possible by rotation, at this moment, it abandons estimating.
Make (d
K-1,
) expression previous sampling instant pilot robot positional information, D
l, D
rExpression left and right wheels of the last one-period distance of travelling respectively then has:
Wherein, α
K-1The robot corner of representing last one-period, D
K-1Be the robot center distance of travelling of last one-period.
When following the robot circular motion, have:
2, prediction pilot robot direction of motion:
In order to predict the direction of motion of pilot robot, follow robot and need obtain current sampling instant k and N before
Pmax(N
Pmax〉=1) inferior sampling instant is with respect to the distance and the observed azimuth d of pilot robot
K-s,
(s=0 ..., N
Pmax) and relevant code-disc information.
Be defined as and follow robot and sampling instant k-N
PmaxCorresponding local coordinate system, note L
K-s(x
L K-s, y
L K-s) (s=0 ..., N
Pmax) be sampling instant k ..., k-N
PmaxThe time pilot robot exist
Position under the coordinate system, O
f K-s(x
Of K-s, y
Of K-s) (s=0 ..., N
Pmax) be sampling instant k ..., k-N
PmaxThe time the photocentre C that follows robot forward direction video camera
Opt Position under the coordinate system, as shown in Figure 3.
Wherein,
S
(k-s-1) (k-s)The distance of between sampling instant k-s-1 and k-s, travelling when robot rectilinear motion (type=1) is followed in expression, θ
(k-s-1) (k-s)And r
(k-s-1) (k-s)Central angle and the radius of representing the circular arc of correspondence when following robot does circular motion (type=0) from sampling instant k-s-1 to k-s respectively.
In conjunction with distance and observed azimuth, follow robot and estimate that pilot robot exists with respect to pilot robot
Position L under the coordinate system
K-s(x
L K-s, y
L K-s) (s=0 ..., N
Pmax) as follows:
Based on
With L
k, follow robot and obtain
Coordinate system is the pilot robot direction of motion of prediction down, its direction angle alpha
K+1For:
3, calculate pilot robot direction of motion and pilot robot to the photocentre C that follows robot forward direction video camera
OptAngle ψ between the direction
Ij k:
Follow the photocentre C of robot forward direction video camera
Opt Position O under the coordinate system
f K-s(x
Of K-s, y
Of K-s) (s=0 ..., N
Pmax) be calculated as follows:
Based on O
f kAnd L
k, obtain β
k(see figure 3):
So the pilot robot direction of motion that can predict and pilot robot are to the photocentre C that follows robot forward direction video camera
OptAngle ψ between the direction
Ij kAs follows:
Consider the inexactness of sensing data, ψ
Ij kBe limited in one and expected angle ψ
Ij dIn the relevant interval, [ψ for example
Ij d-ψ
Ij d-, ψ
Ij d+ ψ
Ij D-u].
4, motion decision-making:
By ψ
Ij kAnd L
Ij k(be d
k), in conjunction with the ψ of expectation
Ij dAnd L
Ij d, draw the desirable formation point P that follows robot
Fd k, as shown in Figure 4.d
P kBe O
f kAnd P
Fd kBetween distance, γ
1Be O
f kL
kWith O
f kP
Fd kBetween angle, then:
Note
For following the relative distance between robot and the pilot robot and the deviation of desired distance,
For pilot robot direction of motion and pilot robot to the angle between the photocentre direction of following robot forward direction video camera and the deviation of expected angle.Work as e
Fld≤ d
FlAnd e
Fl ψ≤ ψ
Fl, follow robot and think that formation realizes stop motion; Work as e
Fld≤ d
FlAnd e
Fl ψ>ψ
Fl, follow robot and fall back up to e along its opposite direction to the pilot robot direction
FldSurpass d
FlTill; Work as e
Fld>d
FlAnd e
Fl ψ≤ ψ
Fl, follow robot with the line of pilot robot on move; Work as e
Fld>d
FlAnd e
Fl ψ>ψ
Fl, its speed v
fAnd rotational angle theta
fFollowing calculating:
When follow robot unpredictable, when estimating pilot robot, see pilot robot if follow the non-forward direction video camera of robot, it should rotate and make the forward direction video camera find pilot robot as early as possible, if do not find pilot robot, its meeting rotary search, if following robot forward direction video camera sees pilot robot, it can move to pilot robot under the restriction of desired distance.
Embodiment
Method for controlling multi-robot coordinated formation of the present invention is applied in the formation control of two robots, and robot adopts the AIM intelligent robot of Institute of Automation Research of CAS's development.Each robot all is furnished with ccd video camera and code-disc, about 60 ° of single camera field range.Follow robot in the experiment and adopt the forward direction video camera C consistent with direction of motion
fTo pilot robot discern, location estimation; The effect of other video camera is to identify and makes C after the pilot robot as early as possible
fFind pilot robot.v
max=0.2m/s,θ
max=20°,N
cs=5,N
pmax=5,
l=0.27m,d
c=5cm,d
fl=0.2m,ψ
fl=20°。Follow the initial distance 1.54m between robot and the pilot robot, 189 ° of initial angles, formation requires that desired distance is 1.2m between two robots, and expected angle is 210 °.Adopt method provided by the present invention, can satisfy the demands, Fig. 5 has provided pilotage people and follower's's (pilot robot and abbreviation of following robot) movement locus, S
lAnd S
fThe starting point of representing pilotage people and follower respectively, G
lAnd G
fThe terminating point of representing pilotage people and follower respectively.
Claims (7)
1. a method for controlling multi-robot coordinated formation is used for multi-robot system; It is characterized in that, comprise step:
A) follow robot according to the colour code tube on the pilot robot,, obtain in conjunction with vision and code-disc information
Current sampling instant k and N before
PmaxInferior sampling instant is with respect to the distance and the observed azimuth d of pilot robot
K-s,
(s=0 ..., N
Pmax);
B) follow robot by code-disc information, by the local coordinate system conversion, recursion obtains sampling instant k and k-N
PmaxThe time pilot robot following the robot local coordinate system
Under position L
k(x
L k, y
L k),
And the photocentre C that follows robot forward direction video camera during current sampling instant k
Opt Position O under the coordinate system
f k(x
Of k, y
Of k);
C) position-based L
k,
And O
f k, follow robot and calculate the pilot robot direction of motion of prediction and pilot robot to C
OptAngle ψ between the direction
Ij k
D) in conjunction with angle ψ
Ij kAnd expected angle ψ
Ijd, follow the distance L between robot and the pilot robot
Ij kAnd desired distance L
Ijd, comprehensive visual information, the decision-making of moving obtains speed v
fAnd rotational angle theta
f, realize following the control of robot speed's size and Orientation.
2. method for controlling multi-robot coordinated formation as claimed in claim 1, it is characterized in that the colour code tube on the pilot robot is the cylindrical tube of hollow, is combined up and down by at least two kinds of colors, be put in the robot, the center of its center and robot is consistent.
3. method for controlling multi-robot coordinated formation as claimed in claim 1 is characterized in that, formation control is to realize there is not communication between the robot on the basis of robot to the pilot robot perceptual inference following.
4. method for controlling multi-robot coordinated formation as claimed in claim 1 is characterized in that, described B) follow the robot local coordinate system in the step
Be and sampling instant k-N
PmaxCorresponding, its initial point
Get forward direction video camera photocentre C
OptThe position at place,
Direction and sampling instant k-N
PmaxThe time robot motion's direction be consistent.
5. method for controlling multi-robot coordinated formation as claimed in claim 1 is characterized in that, described C) follow the prediction of robot in the step to pilot robot direction of motion, be according to current sampling instant k and this moment N before
PmaxThe pilot robot positional information of inferior sampling instant draws, and the pilot robot positional information of these two sampling instants all is based upon the N that follows before the sampling instant k of robot
PmaxUnder the pairing local coordinate system of inferior sampling instant.
6. method for controlling multi-robot coordinated formation as claimed in claim 1 is characterized in that, described D) follow distance L between robot and the pilot robot in the step
Ij kBe d
k
7. as claim 1,4 or 5 described method for controlling multi-robot coordinated formation, it is characterized in that described N
Pmax〉=1.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2008102227743A CN101685309B (en) | 2008-09-24 | 2008-09-24 | Method for controlling multi-robot coordinated formation |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2008102227743A CN101685309B (en) | 2008-09-24 | 2008-09-24 | Method for controlling multi-robot coordinated formation |
Publications (2)
Publication Number | Publication Date |
---|---|
CN101685309A true CN101685309A (en) | 2010-03-31 |
CN101685309B CN101685309B (en) | 2011-06-08 |
Family
ID=42048501
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN2008102227743A Expired - Fee Related CN101685309B (en) | 2008-09-24 | 2008-09-24 | Method for controlling multi-robot coordinated formation |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN101685309B (en) |
Cited By (30)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101825901A (en) * | 2010-03-31 | 2010-09-08 | 北京航空航天大学 | Multi-agent robot cooperative control method based on artificial physics method |
CN101630162B (en) * | 2008-07-16 | 2011-06-08 | 中国科学院自动化研究所 | Local following control method of multiple mobile robots |
CN102169346A (en) * | 2011-02-18 | 2011-08-31 | 杭州电子科技大学 | Intelligent control method for coordinating multiple-robot system |
CN103220778A (en) * | 2013-03-11 | 2013-07-24 | 哈尔滨工业大学 | Mobile node formation change method based on wireless sensor networks and achieving device thereof |
CN103576692A (en) * | 2013-11-07 | 2014-02-12 | 哈尔滨工程大学 | Method for achieving coordinated flight of multiple unmanned aerial vehicles |
CN103592945A (en) * | 2013-11-22 | 2014-02-19 | 中国科学院自动化研究所 | Robot clockwise motion control method |
CN104865826A (en) * | 2015-03-19 | 2015-08-26 | 杭州电子科技大学 | Cooperative predictive control-based multi-robot environment monitoring method |
CN105867428A (en) * | 2016-05-18 | 2016-08-17 | 南京航空航天大学 | Formation changing method for multi-robot system on basis of multiple movement models and multi-view geometry |
CN106020178A (en) * | 2016-08-01 | 2016-10-12 | 天津理工大学 | Autonomous spherical amphibious multi-robot communication system and working method thereof |
CN106143657A (en) * | 2015-03-13 | 2016-11-23 | 西北农林科技大学 | A kind of crawler type detection multi-robot system |
CN106444789A (en) * | 2016-12-03 | 2017-02-22 | 河池学院 | Multi-robot formation control system based on wireless sensor network |
CN106881716A (en) * | 2017-02-21 | 2017-06-23 | 深圳市锐曼智能装备有限公司 | Human body follower method and system based on 3D cameras robot |
CN107085432A (en) * | 2017-06-22 | 2017-08-22 | 星际(重庆)智能装备技术研究院有限公司 | A kind of target trajectory tracking of mobile robot |
CN107273850A (en) * | 2017-06-15 | 2017-10-20 | 上海工程技术大学 | A kind of autonomous follower method based on mobile robot |
CN107943071A (en) * | 2017-11-03 | 2018-04-20 | 中国科学院自动化研究所 | The formation of unmanned vehicle keeps control method and system |
CN108121341A (en) * | 2017-10-31 | 2018-06-05 | 深圳市博鑫创科科技有限公司 | The automatic parking scheme and system of a kind of balance car fleet |
CN108428059A (en) * | 2018-03-27 | 2018-08-21 | 昆明理工大学 | A kind of detecting robot of pipe queue forms and develops method |
CN108664024A (en) * | 2018-05-08 | 2018-10-16 | 清华大学 | The motion planning and Cooperative Localization Method and device that unmanned vehicle network is formed into columns |
CN108983818A (en) * | 2018-08-10 | 2018-12-11 | 电子科技大学 | Unmanned plane order switching method based on virtual architecture |
CN109144047A (en) * | 2018-06-12 | 2019-01-04 | 南京理工大学 | A kind of finite time formation control method based on multi-agent system |
CN109634310A (en) * | 2019-03-01 | 2019-04-16 | 郑州大学 | One kind is based on adaptive multirobot optimization formation control method and system |
CN110058593A (en) * | 2019-04-26 | 2019-07-26 | 国网浙江省电力有限公司电力科学研究院 | Multiple agent approach to formation control and system under a kind of interference environment |
CN110162094A (en) * | 2019-06-13 | 2019-08-23 | 中国人民解放军军事科学院国防科技创新研究院 | A kind of close/intra control method of view-based access control model metrical information |
CN110244772A (en) * | 2019-06-18 | 2019-09-17 | 中国科学院上海微系统与信息技术研究所 | The navigator's system for tracking and navigator's follow-up control method of mobile robot |
CN110780325A (en) * | 2019-08-23 | 2020-02-11 | 腾讯科技(深圳)有限公司 | Method and device for positioning moving object and electronic equipment |
CN111208830A (en) * | 2020-02-23 | 2020-05-29 | 陕西理工大学 | Three-closed-loop formation track tracking control method for wheeled mobile robot |
CN111650938A (en) * | 2020-06-08 | 2020-09-11 | 西交利物浦大学 | AGV control method and device |
CN112148021A (en) * | 2020-09-10 | 2020-12-29 | 北京理工大学 | Multi-agent formation transformation method based on binary tree topological structure and specific rules |
CN112494034A (en) * | 2020-11-30 | 2021-03-16 | 重庆优乃特医疗器械有限责任公司 | Data processing and analyzing system and method based on 3D posture detection and analysis |
CN112650214A (en) * | 2020-11-20 | 2021-04-13 | 上海航天控制技术研究所 | Formation control method for dynamic formation of cluster system |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105527960A (en) * | 2015-12-18 | 2016-04-27 | 燕山大学 | Mobile robot formation control method based on leader-follow |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN100393486C (en) * | 2004-07-13 | 2008-06-11 | 中国科学院自动化研究所 | Method and apparatus for quick tracing based on object surface color |
US7211980B1 (en) * | 2006-07-05 | 2007-05-01 | Battelle Energy Alliance, Llc | Robotic follow system and method |
CN101630162B (en) * | 2008-07-16 | 2011-06-08 | 中国科学院自动化研究所 | Local following control method of multiple mobile robots |
-
2008
- 2008-09-24 CN CN2008102227743A patent/CN101685309B/en not_active Expired - Fee Related
Cited By (48)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101630162B (en) * | 2008-07-16 | 2011-06-08 | 中国科学院自动化研究所 | Local following control method of multiple mobile robots |
CN101825901A (en) * | 2010-03-31 | 2010-09-08 | 北京航空航天大学 | Multi-agent robot cooperative control method based on artificial physics method |
CN102169346A (en) * | 2011-02-18 | 2011-08-31 | 杭州电子科技大学 | Intelligent control method for coordinating multiple-robot system |
CN102169346B (en) * | 2011-02-18 | 2013-07-17 | 杭州电子科技大学 | Intelligent control method for coordinating multiple-robot system |
CN103220778A (en) * | 2013-03-11 | 2013-07-24 | 哈尔滨工业大学 | Mobile node formation change method based on wireless sensor networks and achieving device thereof |
CN103220778B (en) * | 2013-03-11 | 2015-08-19 | 哈尔滨工业大学 | A kind of mobile node order switching method based on wireless sensor network and implement device |
CN103576692A (en) * | 2013-11-07 | 2014-02-12 | 哈尔滨工程大学 | Method for achieving coordinated flight of multiple unmanned aerial vehicles |
CN103592945B (en) * | 2013-11-22 | 2016-08-17 | 中国科学院自动化研究所 | A kind of robot clockwise motion control method |
CN103592945A (en) * | 2013-11-22 | 2014-02-19 | 中国科学院自动化研究所 | Robot clockwise motion control method |
CN106143657A (en) * | 2015-03-13 | 2016-11-23 | 西北农林科技大学 | A kind of crawler type detection multi-robot system |
CN104865826B (en) * | 2015-03-19 | 2017-09-26 | 杭州电子科技大学 | A kind of multirobot method of environmental monitoring controlled based on collaborative forecasting |
CN104865826A (en) * | 2015-03-19 | 2015-08-26 | 杭州电子科技大学 | Cooperative predictive control-based multi-robot environment monitoring method |
CN105867428A (en) * | 2016-05-18 | 2016-08-17 | 南京航空航天大学 | Formation changing method for multi-robot system on basis of multiple movement models and multi-view geometry |
CN106020178A (en) * | 2016-08-01 | 2016-10-12 | 天津理工大学 | Autonomous spherical amphibious multi-robot communication system and working method thereof |
CN106020178B (en) * | 2016-08-01 | 2023-04-18 | 天津理工大学 | Autonomous spherical amphibious multi-robot communication system and working method thereof |
CN106444789A (en) * | 2016-12-03 | 2017-02-22 | 河池学院 | Multi-robot formation control system based on wireless sensor network |
CN106881716A (en) * | 2017-02-21 | 2017-06-23 | 深圳市锐曼智能装备有限公司 | Human body follower method and system based on 3D cameras robot |
CN107273850A (en) * | 2017-06-15 | 2017-10-20 | 上海工程技术大学 | A kind of autonomous follower method based on mobile robot |
CN107085432A (en) * | 2017-06-22 | 2017-08-22 | 星际(重庆)智能装备技术研究院有限公司 | A kind of target trajectory tracking of mobile robot |
CN107085432B (en) * | 2017-06-22 | 2020-03-31 | 迪比(重庆)智能科技研究院有限公司 | Target track tracking method of mobile robot |
CN108121341A (en) * | 2017-10-31 | 2018-06-05 | 深圳市博鑫创科科技有限公司 | The automatic parking scheme and system of a kind of balance car fleet |
CN107943071A (en) * | 2017-11-03 | 2018-04-20 | 中国科学院自动化研究所 | The formation of unmanned vehicle keeps control method and system |
CN107943071B (en) * | 2017-11-03 | 2020-02-07 | 中国科学院自动化研究所 | Formation maintaining control method and system for unmanned vehicle |
CN108428059B (en) * | 2018-03-27 | 2021-07-16 | 昆明理工大学 | Pipeline detection robot queue forming and evolution method |
CN108428059A (en) * | 2018-03-27 | 2018-08-21 | 昆明理工大学 | A kind of detecting robot of pipe queue forms and develops method |
CN108664024A (en) * | 2018-05-08 | 2018-10-16 | 清华大学 | The motion planning and Cooperative Localization Method and device that unmanned vehicle network is formed into columns |
CN108664024B (en) * | 2018-05-08 | 2020-09-25 | 清华大学 | Motion planning and cooperative positioning method and device for unmanned vehicle network formation |
CN109144047A (en) * | 2018-06-12 | 2019-01-04 | 南京理工大学 | A kind of finite time formation control method based on multi-agent system |
CN109144047B (en) * | 2018-06-12 | 2021-07-06 | 南京理工大学 | Finite time formation control method based on multi-agent system |
CN108983818A (en) * | 2018-08-10 | 2018-12-11 | 电子科技大学 | Unmanned plane order switching method based on virtual architecture |
CN109634310A (en) * | 2019-03-01 | 2019-04-16 | 郑州大学 | One kind is based on adaptive multirobot optimization formation control method and system |
CN109634310B (en) * | 2019-03-01 | 2021-08-20 | 郑州大学 | Self-adaptive multi-robot-based optimized formation control method and system |
CN110058593A (en) * | 2019-04-26 | 2019-07-26 | 国网浙江省电力有限公司电力科学研究院 | Multiple agent approach to formation control and system under a kind of interference environment |
CN110058593B (en) * | 2019-04-26 | 2022-09-02 | 国网浙江省电力有限公司电力科学研究院 | Multi-agent formation control method and system in interference environment |
CN110162094B (en) * | 2019-06-13 | 2022-02-22 | 中国人民解放军军事科学院国防科技创新研究院 | Intensive formation control method based on visual measurement information |
CN110162094A (en) * | 2019-06-13 | 2019-08-23 | 中国人民解放军军事科学院国防科技创新研究院 | A kind of close/intra control method of view-based access control model metrical information |
CN110244772A (en) * | 2019-06-18 | 2019-09-17 | 中国科学院上海微系统与信息技术研究所 | The navigator's system for tracking and navigator's follow-up control method of mobile robot |
CN110780325A (en) * | 2019-08-23 | 2020-02-11 | 腾讯科技(深圳)有限公司 | Method and device for positioning moving object and electronic equipment |
CN110780325B (en) * | 2019-08-23 | 2022-07-19 | 腾讯科技(深圳)有限公司 | Method and device for positioning moving object and electronic equipment |
CN111208830B (en) * | 2020-02-23 | 2023-04-25 | 陕西理工大学 | Three-closed-loop formation track tracking control method for wheeled mobile robot |
CN111208830A (en) * | 2020-02-23 | 2020-05-29 | 陕西理工大学 | Three-closed-loop formation track tracking control method for wheeled mobile robot |
CN111650938B (en) * | 2020-06-08 | 2023-09-26 | 西交利物浦大学 | AGV control method and device |
CN111650938A (en) * | 2020-06-08 | 2020-09-11 | 西交利物浦大学 | AGV control method and device |
CN112148021A (en) * | 2020-09-10 | 2020-12-29 | 北京理工大学 | Multi-agent formation transformation method based on binary tree topological structure and specific rules |
CN112650214B (en) * | 2020-11-20 | 2022-12-23 | 上海航天控制技术研究所 | Formation control method for dynamic formation of cluster system |
CN112650214A (en) * | 2020-11-20 | 2021-04-13 | 上海航天控制技术研究所 | Formation control method for dynamic formation of cluster system |
CN112494034B (en) * | 2020-11-30 | 2023-01-17 | 重庆优乃特医疗器械有限责任公司 | Data processing and analyzing system and method based on 3D posture detection and analysis |
CN112494034A (en) * | 2020-11-30 | 2021-03-16 | 重庆优乃特医疗器械有限责任公司 | Data processing and analyzing system and method based on 3D posture detection and analysis |
Also Published As
Publication number | Publication date |
---|---|
CN101685309B (en) | 2011-06-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN101685309B (en) | Method for controlling multi-robot coordinated formation | |
CN110989526B (en) | double-AGV cooperative carrying control system and method | |
CN109032138B (en) | Consistency algorithm-based multi-robot formation control system and method | |
Saska et al. | Coordination and navigation of heterogeneous UAVs-UGVs teams localized by a hawk-eye approach | |
CN111766879A (en) | Intelligent vehicle formation system based on autonomous collaborative navigation | |
CN103454919B (en) | The control method of the kinetic control system of mobile robot in intelligent space | |
Bürki et al. | Appearance-based landmark selection for efficient long-term visual localization | |
US20220114897A1 (en) | Method for feasibility evaluation of UAV digital twin based on vicon motion capture system | |
CN103885449A (en) | Intelligent visual tracking wheeled robot based on multiple sensors and control method thereof | |
CN108897215B (en) | Multi-ocean-robot collaborative annular scanning method based on distributed model predictive control | |
CN114721275B (en) | Visual servo robot self-adaptive tracking control method based on preset performance | |
Miao et al. | Low-complexity leader-following formation control of mobile robots using only FOV-constrained visual feedback | |
Liu et al. | Hercules: An autonomous logistic vehicle for contact-less goods transportation during the COVID-19 outbreak | |
Wanasinghe et al. | Distributed collaborative localization for a heterogeneous multi-robot system | |
CN110333723B (en) | Unmanned ship collaborative formation method based on dual communication equipment | |
CN113618741B (en) | Robot joint pose optimization method, robot control method and robot | |
CN113835434B (en) | Distributed multi-unmanned ship self-adaptive neural network formation control method | |
CN110967017A (en) | Cooperative positioning method for rigid body cooperative transportation of double mobile robots | |
CN113009922B (en) | Scheduling management method for robot walking path | |
Pacheco et al. | Control signal weighting for robot formation | |
Naglak et al. | Backseat control of sandshark auv using ros on raspberrypi | |
KR101981641B1 (en) | Method and system for formation control of multiple mobile robots | |
Zhao et al. | Formation control of networked mobile robots with unknown reference orientation | |
CN111708010B (en) | Mobile equipment positioning method, device and system and mobile equipment | |
CN111221318A (en) | Multi-robot state estimation method based on model predictive control algorithm |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20110608 Termination date: 20210924 |