CN103878770B - Robot for space vision time delay error compensating method based on velocity estimation - Google Patents

Robot for space vision time delay error compensating method based on velocity estimation Download PDF

Info

Publication number
CN103878770B
CN103878770B CN201410138351.9A CN201410138351A CN103878770B CN 103878770 B CN103878770 B CN 103878770B CN 201410138351 A CN201410138351 A CN 201410138351A CN 103878770 B CN103878770 B CN 103878770B
Authority
CN
China
Prior art keywords
theta
centerdot
robot
space
time delay
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201410138351.9A
Other languages
Chinese (zh)
Other versions
CN103878770A (en
Inventor
王滨
李振宇
刘宏
赵京东
李志奇
王志超
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin Institute of Technology
Original Assignee
Harbin Institute of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin Institute of Technology filed Critical Harbin Institute of Technology
Priority to CN201410138351.9A priority Critical patent/CN103878770B/en
Publication of CN103878770A publication Critical patent/CN103878770A/en
Application granted granted Critical
Publication of CN103878770B publication Critical patent/CN103878770B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • B25J9/1656Programme controls characterised by programming, planning systems for manipulators
    • B25J9/1671Programme controls characterised by programming, planning systems for manipulators characterised by simulation, either to verify existing program or to create and verify new program, CAD/CAM oriented, graphic oriented programming systems

Abstract

Robot for space vision time delay error compensating method based on velocity estimation, solves Free-floating underchassis space robot vision and measures the compensation problem of time delay error.Including determining the vision time delay of system, set up the mathematical relationship between the vision measurement data of band time delay and the true relative pose of physics;Vision measurement data according to band time delay and the joint instruction of mechanical arm, estimate current robot for space tip speed;Design error controller, reduces the estimation difference of robot for space tip speed;According to correction rear space robot end's Velocity Estimation, the current vision measurement data with time delay are compensated, obtains the vision measurement data through overcompensation.The present invention utilizes historical measurement data to merge joint angle speed command estimation current spatial robot end's speed, and design estimation difference controller is to reduce the error of velocity estimation;Realize the accurate vision time delay of Free-floating underchassis space robot to compensate, be advantageously implemented robot for space and complete accurate operation task accurately.

Description

Robot for space vision time delay error compensating method based on velocity estimation
Technical field
The present invention relates to a kind of robot for space vision measurement time delay error compensation method.
Background technology
When robot for space (including carrier spacecraft and mechanical arm) under floating pedestal performs space tasks, faced Individual practical problem is the vision measurement system on its mechanical arm, carries out generally requiring during Vision information processing when consuming longer Between, one reason for this is that the Vision information processing method that robot for space is used is more complicated, separately owing to illumination condition is complicated The processor computing capability that one reason can be used for robot for space Vision information processing is the most limited, owing to vision is believed Contradiction between breath method complexity and Vision information processing equipment computing capability result in vision measurement information often to be had Bigger time delay, shows and will appear as having bigger measurement error on videogrammetry system.Due to Slight measurement errors Exist, make robot for space when tasks such as performance objective captures, its performance accuracy can be affected.In order to enable robot for space More rapid, smoothly achieve fine service role in-orbit, need to consider space machine in Space Robot System design process Device people's vision time delay compensation of error problem.
Some have been had to can be used for the prediction of ground fixed pedestal robot vision time delay error process and estimation side at present Method.Least-squares estimation is the minimum criterion of the quadratic sum with error, according to unknown parameter in observation data estimation linear model A kind of basic parameter method of estimation.Its basic ideas be select estimator make model output with actual measurement output difference square With minimize.This mode seeking error sum of squares can avoid positive negative error to offset, and is easy to Mathematical treatment.Kalman Filtering is a kind of linear minimum-variance estimation, can estimate past and the current state of signal, even can estimate state in the future. Kalman prediction model estimates desired signal by method from the measured value relevant with being extracted signal.Wherein estimated Meter signal be the random response caused by white-noise excitation, driving source and response between system equation oneself know, measurement amount and quilt Also oneself knows functional relationship between estimator.Smith predictor is usually applied to ground robot vision time delay and processes.Smith The action principle of prediction device is by introducing time delay rectification link, compensating with the vision data measuring time delay, thus Outside pure time delay in system is partially separated whole closed loop system, thus improve the control performance of whole control system.From The action principle of Smith predictor is it can be seen that the basis of Smith predictor design is must assure that and can predict accurately Go out controlled device dynamic characteristic under given input state, the time delay part in metrical information can be compensated, therefore Smith predictor is highly dependent on sets up accurate mathematical model.Robot for space under floating pedestal is with ground robot not With, the pedestal do not fixed due to robot for space, the motion of mechanical arm is coupled with the motion of carrier spacecraft, any The motion of space manipulator all can change position and the attitude of carrier spacecraft, and the change of carrier Space Vehicle position attitude also can Affecting the location of space manipulator in turn, therefore the dynamics calculation of robot for space is extremely complex.Moreover, along with load The continuous consumption of body spacecraft fuel, the quality of carrier spacecraft, centroid position and inertia matrix are all constantly in change, accurately Mathematical model is difficult to obtain, and therefore classical prediction and method of estimation are difficult to obtain well effect in robot for space is applied Really.
Summary of the invention
The present invention provides a kind of Free-floating underchassis space robot vision time delay error compensation method based on velocity estimation, To reduce the Slight measurement errors that robot for space visual system produces because time delay exists, beneficially robot for space is based on benefit Repay after-vision information and complete fine space tasks.
Robot for space vision time delay error compensating method based on velocity estimation is completed by following steps:
Step one, determine the vision time delay of visual system according to the vision processing algorithm used and the hardware applied, Set up the mathematical relationship between the vision measurement data of band time delay and the true relative pose of physics;Step 2, basis are with time delay The joint instruction of vision measurement data and mechanical arm, estimates the tip speed (see formula (26)) of current spatial robot;Step 3, Design error controller, reduces the error of robot for space tip speed valuation, obtains correcting rear space robot end's speed Valuation (see formula (30));Step 4, according to correction rear space robot end's Velocity Estimation, currently vision with time delay is surveyed Amount data compensate, and obtain the vision measurement data (see formula (32)) through overcompensation.
The method have the advantages that the inventive method passes through robot for space mathematics model analysis, give profit Merge joint angle speed command by historical measurement data and estimate current spatial robot end's speed.The present invention utilizes space machine The measurement data of the band time delay that human visual system obtains and the instruction of joint of mechanical arm angular velocity, estimate current spatial robot end Speed, simultaneously design estimation difference controller, reduce the error of robot for space tip speed valuation, according to empty after being corrected Between robot end's Velocity Estimation, target is regarded by the Free-floating underchassis space robot that can effectively compensate for causing due to time delay Feel measurement error, it is ensured that robot for space completes accurate operation task the most accurately.
The present invention is not required to set up the accurate mathematical model of robot for space, and the mechanical arm utilizing mathematics model analysis to obtain is true Real tip speed and the simplification relation of band time delay tip speed, carry out tip speed estimation, and use the method for closed loop control to carry The high precision of robot for space vision time delay error compensation;This method avoid the dynamics calculation of complexity, calculate simple, just Compensate in realizing the accurate vision time delay of Free-floating underchassis space robot, be advantageously implemented robot for space and complete accurately Accurate operation task, can be used for robot for space safeguard in-orbit, space junk cleaning, the space application such as survey of deep space.
Accompanying drawing explanation
Fig. 1 is the vision time delay error control system block diagram of the present invention;Fig. 2 is the flow chart of the present invention;Fig. 3 is the present invention Space Robot System and each coordinate system schematic diagram;Fig. 4 be compensate estimation data and raw measurement data that after-vision measures and True position versus figure relatively, wherein: Fig. 4 a is to compensate estimation data and raw measurement data and the true phase that after-vision is measured To the position comparison diagram in x-axis direction, Fig. 4 b is to compensate estimation data and raw measurement data and the true phase that after-vision is measured To the position comparison diagram in y-axis direction, Fig. 4 c is for compensating estimation data and raw measurement data that after-vision measures and truly The position comparison diagram in z-axis direction relatively;
Fig. 5 is the contrast compensating after-vision measurement data estimation difference with not compensating anterior optic measurement data time delay error Figure, wherein: Fig. 5 a is for compensating after-vision measurement data estimation difference and not compensating anterior optic measurement data time delay error in x-axis side To comparison diagram, Fig. 5 b is vision measurement data estimation error and do not compensate anterior optic measurement data time delay error in y-axis direction Comparison diagram, Fig. 5 c is vision measurement data estimation error and do not compensate anterior optic measurement data time delay error in z-axis direction Comparison diagram.
Detailed description of the invention
Detailed description of the invention one: combining Fig. 1, Fig. 2 and Fig. 3 and present embodiment is described, present embodiment is complete by following steps Become: one, determine the time delay of visual system according to the vision processing algorithm used and the hardware applied, set up regarding of band time delay Mathematical relationship between feel measurement data and the true relative pose of physics (measuring camera and target);Two, according to regarding with time delay Feel that the joint of measurement data and mechanical arm instructs, estimate the tip speed of current spatial robot;Three, design error controller, Reduce the error of robot for space tip speed valuation, obtain correcting rear space robot end's Velocity Estimation;Four, according to correction The current vision measurement data with time delay are compensated, obtain through overcompensation by rear space robot end's Velocity Estimation Vision measurement data.
Detailed description of the invention two: the present embodiment difference from detailed description of the invention one is: present embodiment is in step The time delay determining visual system according to the vision processing algorithm used and the hardware applied described in one, sets up band time delay Mathematical relationship between vision measurement data and the true relative pose of physics (measuring camera and target) by: according to the vision used Information processing method and the hardware platform applied determine that the time delay that whole vision measurement link causes is m cycle, if system Time a length of t in each cycles, then the vision time delay T of Space Robot SystemdFor:
Td=m × ts (1)
Vision measurement information D in definition space robot k momentv(k) and actual distance information Dr(k) be
Dv(k)=[xv(k) yv(k) zv(k) αv(k) βv(k) γv(k)]T (2)
Dr(k)=[xr(k) yr(k) zr(k) αr(k) βr(k) γr(k)]T (3)
Wherein k is any time, and x (k), y (k), z (k) are relative position information, and α (k), β (k), γ (k) are for describing phase Eulerian angles to pose, robot for space vision measurement information Dv(k) and actual distance information DrK the relation between () is
Dr(k)=Dv(k+m) (4)
K represents that any time, m represent time delay.
If current time is N, vision measurement information is Dv(N), then may know that Dr(N-m) the actual distance information before is all Can be directly obtained by the vision measurement data of last time.Definition true velocity information Vr(k) be
V r ( k ) = x · r ( k ) y · r ( k ) z · r ( k ) α · r ( k ) β · r ( k ) γ · r ( k ) T - - - ( 5 )
Can be by DrAnd D (k+1)rK () is calculated
V r ( k ) = D r ( k + 1 ) - D r ( k ) t s - - - ( 6 )
Therefore true velocity information V of robot for spacerAnd V (N-m-1)r(N-m-1) the true velocity information before also may be used To pass through to be calculated, the most real range information Dk(N) it is
D r ( N ) = D v ( N ) + t s × Σ i = 1 m V r ( N - i ) - - - ( 7 )
Detailed description of the invention three: the present embodiment difference from detailed description of the invention two is: present embodiment is in step Instructing according to the vision measurement data of time delay and the joint of mechanical arm described in two, estimates the end of current spatial robot Speed, its process is:
Utilize the vision measurement data with time delay, available true velocity sequenceValuation sequenceThen can Negotiation speed valuation sequenceTo current vision measurement information Dv(N) compensating, formula (7) can It is rewritten as:
D ~ r ( N ) = D v ( N ) + t s × Σ i = 1 m V ~ r ( N - i ) - - - ( 8 )
Thus obtain the estimated value of current actual distance information
Below by combining with the joint instruction of mechanical arm with the robot for space vision measurement data of time delay, construct The valuation of the robot for space tip speed of current time.
Robot for space owing to being under zero gravity state meets the conservation of momentum and the conservation of angular momentum, can set up according to this Mathematical relationship between robot for space joint motions state and end-effector generalized velocity, meets equation:
v e ω e = J g ( Ψ 0 , Θ , m i , I i ) Θ · = J g _ v J g _ ω Θ · - - - ( 9 )
Wherein parameter matrix JgIt is exactly the broad sense Jacobian matrix of robot for space, is inertia matrix Ii, mass parameter miWith Joint of robot angle Θ, carrier spacecraft attitude Ψ0Function,It is the joint angle speed of robot.Wherein
v e = x · r y · r z · r T - - - ( 10 )
ω e = 0 - sinα r cosα r cosβ r 0 cosα r sinα r cosβ r 1 0 - sinβ r α · r β · r γ · r - - - ( 11 )
veAnd ωeIt is linear velocity and the angular velocity of robot for space end effector respectively,WithFor space machine The differential of robot end attitude Eulerian angles.
Generalized Jacobian is launched, it is considered to the end generalized velocity of robot for space
v e ω e = Σ i = 1 n J g i θ i - - - ( 12 )
Wherein JgiIt is the i-th row of broad sense Jacobian matrix.
Because each kinesiology of robot for space, the actual value of kinetic parameter are included in the measurement data of band time delay, fixed It is engraved in the neighborhood of current time N, by J during justice Mgi(M) expand into:
J g i ( M ) = J g i ( N ) + Σ i = 1 n Σ j = 1 ∞ 1 j ! ∂ j J g i ( N ) ∂ θ i j [ θ i ( M ) - θ i ( N ) ] j + E J i ′ ( M ) - - - ( 13 )
Wherein E'Ji(M) for the error caused because of carrier position variation.Only retain first order therein, be further deformed into
J g i ( M ) = J g i ( N ) + Σ i = 1 n ∂ J g i ( N ) ∂ θ i [ θ i ( M ) - θ i ( N ) ] + E J i ( M ) E J i ( M ) = Σ i = 1 n Σ j = 2 ∞ 1 j ! ∂ j J g i ( N ) ∂ θ i j [ θ i ( M ) - θ i ( N ) ] j + E J i ′ ( M ) - - - ( 14 )
Now the robot for space tip speed in M moment can be expressed as
v e ( M ) ω e ( M ) = Σ i = 1 n J g i ( M ) θ · i ( M ) = J g ( N ) θ · ( M ) + ΔJ N ( M ) θ · ( M ) + E J ( M ) θ · ( M ) - - - ( 15 )
Wherein
ΔJ N ( M ) = ∂ J g 1 ( N ) ∂ θ 1 [ θ 1 ( M ) - θ 1 ( N ) ] ... ∂ J g n ( N ) ∂ θ n [ θ n ( M ) - θ n ( N ) ] - - - ( 16 )
The robot for space joint angle speed in M moment is expressed as
θ · ( M ) = θ · ( M ) · θ · ( N ) | θ · ( N ) | 2 2 θ · ( N ) + E θ ( M ) - - - ( 17 )
Wherein
E θ ( M ) = θ · ( M ) - θ · ( M ) · θ · ( N ) | θ · ( N ) | 2 2 θ · ( N ) - - - ( 18 )
When moving due to robot for space, joint angle velocity variations is slow, so formula (17) Section 1 is that the M moment is empty Between the major part of joint of robot angular velocity, Eθ(M) it is the remainder relevant to acceleration;
Bring formula (17) into formula (15), obtain the robot for space tip speed in M moment and be expressed as:
v e ( M ) ω e ( M ) = Σ i = 1 n J g i ( M ) θ · i ( M ) = θ · ( M ) · θ · ( N ) | θ · ( N ) | 2 2 J g ( N ) θ · ( N ) + ΔJ N ( M ) θ · ( M ) · θ · ( N ) | θ · ( N ) | 2 2 θ · ( N ) + E V ( M ) = θ · ( M ) · θ · ( N ) | θ · ( N ) | 2 2 v e ( M ) ω e ( M ) + ΔJ N ( M ) θ · ( M ) · θ · ( N ) | θ · ( N ) | 2 2 θ · ( N ) + E V ( M ) - - - ( 19 )
Wherein
E V ( M ) = J g ( N ) E θ ( M ) + Δ J ( M ) E θ ( M ) + E J ( M ) θ · ( M ) - - - ( 20 )
Current time is n-hour, and vision measurement time delay is P sampling period, according to formula (19), obtains the end in N-P moment Speed is:
v e ( N - P ) ω e ( N - P ) = θ · ( N - P ) · θ · ( N ) | θ · ( N ) | 2 2 v e ( N ) ω e ( N ) + ΔJ N ( N - P ) θ · ( N - P ) · θ · ( N ) | θ · ( N ) | 2 2 θ · ( N ) + E V ( N - P ) - - - ( 21 )
According to formula (19), obtaining the tip speed in N-P-1 moment is:
v e ( N - P - 1 ) ω e ( N - P - 1 ) = θ · ( N - P - 1 ) · θ · ( N ) | θ · ( N ) | 2 2 v e ( N ) ω e ( N ) + ΔJ N ( N - P - 1 ) θ · ( N - P - 1 ) · θ · ( N ) | θ · ( N ) | 2 2 θ · ( N ) + E V ( N - P - 1 ) - - - ( 22 )
And the N-P-1 moment is for the target rough estimate matrix Δ J of n-hourNIt is represented by:
ΔJ N ( N - P - 1 ) = ∂ J g 1 ( N ) ∂ θ 1 [ θ 1 ( N - P - 1 ) - θ 1 ( N ) ] ... ∂ J g n ( N ) ∂ θ n [ θ n ( N - P - 1 ) - θ n ( N ) ] = [ θ 1 ( N - P - 1 ) - θ 1 ( N ) ] · [ θ 1 ( N - P ) - θ 1 ( N ) ] | [ θ 1 ( N - P ) - θ 1 ( N ) ] | 2 2 - - - ( 23 )
Wherein TΔJFor remainder.Can obtain current end speed according to above conclusion is
v e ( N ) ω e ( N ) = - β α 1 ( 1 - β ) v e ( N - P ) ω e ( N - P ) + 1 α 2 ( 1 - β ) v e ( N - P - 1 ) ω e ( N - P - 1 ) + Δ E - - - ( 24 )
Wherein
β = [ θ 1 ( N - P - 1 ) - θ 1 ( N ) ] · [ θ 1 ( N - P ) - θ 1 ( N ) ] | [ θ 1 ( N - P ) - θ 1 ( N ) ] | 2 2 α 1 = θ · ( N - P ) · θ · ( N ) | θ · ( N ) | 2 2 α 2 = θ · ( N - P - 1 ) · θ · ( N ) | θ · ( N ) | 2 2 - - - ( 25 )
For convenience, α can be claimed1、α2For second-order linearity estimation coefficient, β is state difference item.Ignore high-order error delta E, this Time tip speed can be with approximate calculation
v ~ e ( N ) ω ~ e ( N ) = - β α 1 ( 1 - β ) v e ( N - P ) ω e ( N - P ) + 1 α 2 ( 1 - β ) v e ( N - P - 1 ) ω e ( N - P - 1 ) - - - ( 26 )
Estimation difference now is
Δ E = - β α 1 ( 1 - β ) E V ( N - P ) + 1 α 2 ( 1 - β ) [ T Δ J θ · ( N ) + E V ( N - P - 1 ) ] - - - ( 27 )
Estimating speed now may be calculated
V ~ r ( k ) = x · r ( k ) y · r ( k ) z · r ( k ) α · r ( k ) β · r ( k ) γ · r ( k ) T = [ v ~ e ( k ) 0 - sinα r cosα r cosβ r 0 cosα r sinα r cosβ r 1 0 - sinβ r - 1 ω ~ e ( k ) ] - - - ( 28 )
In the case of, robot for space motion less in measurement time delay is relatively stable, error delta E directly carrying out estimating should The least, and estimation difference Δ E is likely to bigger in other cases.For the special circumstances that estimation difference is bigger, design one Error controller, carries out real-time correction to estimated value, thus improves estimated accuracy.
Detailed description of the invention four: the present embodiment difference from detailed description of the invention three is: present embodiment is in step Error controller described in three is designed as: the process of the Velocity Estimation sequence described in step 2 is regarded as surpassing with error Prosomite:
P ( s ) = e ~ st s = e st s [ 1 + H Δ ( s ) ] - - - ( 29 )
WhereinFor the error transfer function of differentiation element, the transmission letter of vision time delay error control system output Number is:
Y ( s ) = e - st s e ~ st s + G ( s ) e - st s 1 + G ( s ) e - st s V r ( s ) = V r ( s ) + H Δ ( s ) 1 + G ( s ) e - st s V r ( s ) - - - ( 30 )
Wherein VrS () is the actual value of robot end's speed, Y (s) is robot end's Velocity Estimation, G (s) after correction For error controller.
The estimation difference transmission function of system is:
E ( s ) = - H Δ ( s ) e - st s - G ( s ) e - 2 st s 1 + G ( s ) e - st s V r ( s ) - - - ( 31 )
By POLE PLACEMENT USING, design makes error controller G (s) stable for E (s), and estimation difference E (t) is the most gradually received Holding back, after correction, robot end's Velocity Estimation gradually levels off to speed actual value.
Detailed description of the invention five: the present embodiment difference from detailed description of the invention four is: present embodiment is in step Currently vision measurement data with time delay are entered by described according to correction rear space robot end's Velocity Estimation described in four Row compensates, through the vision measurement data of overcompensationIt is calculated as follows:
D ~ r ( N ) = D v ( N ) + t s × Σ i = 1 m Y ( N - i ) - - - ( 32 )
Wherein Y (N-i) is correction rear space robot end's Velocity Estimation that step 4 draws, when so far completing vision Prolong compensation of error.
The content not being described in detail in description of the invention belongs to prior art known to professional and technical personnel in the field.
Embodiment
In conjunction with Fig. 1, Fig. 2, Fig. 3, the present embodiment being described, experimental system is made up of robot for space, target satellite, space machine Device people includes sixdegree-of-freedom simulation and carrier spacecraft, and mechanical arm tail end installs camera.Machinery arm straps camera is followed the tracks of close to mesh Mark star, target satellite moves along mechanical arm tail end Z-direction with the speed of 4mm/s.
For setting up analogue system, carrier spacecraft and the kinesiology of sixdegree-of-freedom simulation and kinetic parameter such as table 1 institute Show
Table 1. robot for space parameter
The vision time delay error compensation step of robot for space is:
Step one, determine according to the Vision information processing method used and the hardware environment applied visual system time Prolong, set up the mathematical relationship between actual distance information and vision measurement information.
Step 2, utilize formula (26) calculate current time robot for space tip speed estimate.
Step 3, according to formula (30) design vision time delay error controller, via controller export, obtain correct rear space Robot end's Velocity Estimation.
Step 4, utilize formula (32) calculate compensate after vision measurement data.
Designed robot for space vision time delay error compensation actual effect is as shown in Figure 4, Figure 5.By the compensation backsight of Fig. 4 Estimation data and the raw measurement data and true relative position versus figure that feel is measured, it can be seen that due to depositing of vision time delay , on tri-directions of X, Y and Z, there is bigger error to the measured value of target with actual value in visual system, mends through error Repay, the estimation data closely actual value of vision measurement.By the compensation after-vision measurement data estimation difference of Fig. 5 with do not compensate The comparison diagram of anterior optic measurement data time delay error, it can be seen that compensate anterior optic and measure the maximum error with actual value, at X Direction is reduced to 7mm by 18mm, is reduced to 8mm in Z-direction by 38mm, it can be seen that this delay compensation method of the present invention, can So that Slight measurement errors that Free-floating underchassis space robot due to time delay cause is obviously reduced, be conducive to better ensuring that space Robot completes the requirement of accurate operation task the most accurately, and whole process calculates simple, it is not necessary to set up space The accurate mathematical model of robot, is not required to carry out the dynamics calculation of complexity, can meet engineering actual demand.

Claims (4)

1. a robot for space vision time delay error compensating method based on velocity estimation, it is characterised in that: described method by Following steps complete:
Step one, determine the vision time delay of visual system according to the vision processing algorithm used and the hardware applied, set up Mathematical relationship between vision measurement data with time delay and the true relative pose of physics, process is:
Determine what whole vision measurement link caused according to the Vision information processing method used and the hardware platform applied Time delay is m cycle, if time a length of t of system per cycles, then the vision time delay T of Space Robot SystemdFor:
Td=m × ts (1)
Vision measurement information D in definition space robot k momentv(k) and actual distance information Dr(k) be:
Dv(k)=[xv(k) yv(k) zv(k) αv(k) βv(k) γv(k)]T (2)
Dr(k)=[xr(k) yr(k) zr(k) αr(k) βr(k) γr(k)]T (3)
Wherein k is any time, and x (k), y (k), z (k) are relative position information, and α (k), β (k), γ (k) are for describing phase para-position The Eulerian angles of appearance, robot for space vision measurement information Dv(k) and actual distance information DrK the relation between () is:
Dr(k)=Dv(k+m) (4)
K represents that any time, m represent time delay;
If current time is N, vision measurement information is Dv(N), then may know that Dr(N-m) the actual distance information before all can be by The vision measurement data of last time directly obtain, and define true velocity information Vr(k) be:
V r ( k ) = x · r ( k ) y · r ( k ) z · r ( k ) α · r ( k ) β · r ( k ) γ · r ( k ) T - - - ( 5 )
By DrAnd D (k+1)rK () is calculated
V r ( k ) = D r ( k + 1 ) - D r ( k ) t s - - - ( 6 )
True velocity information V of robot for spacerAnd V (N-m-1)r(N-m-1) the true velocity information before is by calculating Arrive: real range information Dk(N) it is
D r ( N ) = D v ( N ) + t s × Σ i = 1 m V r ( N - i ) ; - - - ( 7 )
Step 2, basis instruct with the vision measurement data of time delay and the joint of mechanical arm, estimate current spatial robot Tip speed;
Step 3, design error controller, reduce the error of robot for space tip speed valuation, obtains correcting rear space machine Robot end Velocity Estimation;
Step 4, according to correction rear space robot end's Velocity Estimation, currently vision measurement data with time delay are carried out Compensate, obtain the vision measurement data through overcompensation.
A kind of robot for space vision time delay error compensating method based on velocity estimation the most according to claim 1, its Being characterised by: in step 2, described instructs according to the vision measurement data of time delay and the joint of mechanical arm, estimates to work as The tip speed of front space robot, its process is:
Utilize the vision measurement data with time delay, obtain true velocity sequenceValuation sequenceThen may be used Negotiation speed valuation sequenceTo current vision measurement information Dv(N) compensating, formula (7) is rewritten as:
D ~ r ( N ) = D v ( N ) + t s × Σ i = 1 m V ~ r ( N - i ) - - - ( 8 )
Thus obtain the estimated value of current actual distance information
By combining with the joint instruction of mechanical arm with the robot for space vision measurement data of time delay, construct current time The valuation of robot for space tip speed;
Robot for space owing to being under zero gravity state meets the conservation of momentum and the conservation of angular momentum, sets up space machine according to this Mathematical relationship between device person joint's kinestate and end-effector generalized velocity, meets equation:
v e ω e = J g ( Ψ 0 , Θ , m i , I i ) Θ · = J g _ v J g _ ω Θ · - - - ( 9 )
Wherein parameter matrix JgIt is exactly the broad sense Jacobian matrix of robot for space, is inertia matrix Ii, mass parameter miAnd machine Person joint angle Θ, carrier spacecraft attitude Ψ0Function,It is the joint angle speed of robot;Wherein
v e = x · r y · r z · r T - - - ( 10 )
ω e = 0 - sinα r cosα r cosβ r 0 cosα r sinα r cosβ r 1 0 - sinβ r α · r β · r γ · r - - - ( 11 )
veAnd ωeIt is linear velocity and the angular velocity of robot for space end effector respectively,WithFor robot for space end The differential of attitude Eulerian angles;
Generalized Jacobian is launched, it is considered to the end generalized velocity of robot for space:
v e ω e = Σ i = 1 n J g i θ i - - - ( 12 )
Wherein JgiIt is the i-th row of broad sense Jacobian matrix, θiIt it is i-th joint of robot angle;
Because each kinesiology of robot for space, the actual value of kinetic parameter are included in the measurement data of band time delay, during definition M It is engraved in the neighborhood of current time N, by Jgi(M) expand into:
J g i ( M ) = J g i ( N ) + Σ i = 1 n Σ j = 1 ∞ 1 j ! ∂ j J g i ( N ) ∂ θ i j [ θ i ( M ) - θ i ( N ) ] j + E J i ′ ( M ) - - - ( 13 )
Wherein E'Ji(M) for the error caused because of carrier position variation;Only retain first order therein, be further deformed into
J g i ( M ) = J g i ( N ) + Σ i = 1 n ∂ J g i ( N ) ∂ θ i [ θ i ( M ) - θ i ( N ) ] + E J i ( M ) E J i ( M ) = Σ i = 1 n Σ j = 2 ∞ 1 j / ∂ j J g i ( N ) ∂ θ i j [ θ i ( M ) - θ i ( N ) ] j + E J i ′ ( M ) - - - ( 14 )
Now the robot for space tip speed in M moment is represented by:
v e ( M ) ω e ( M ) = Σ i = 1 n J g i ( M ) θ · i ( M ) = J g ( N ) θ · ( M ) + ΔJ N ( M ) θ · ( M ) + E J ( M ) θ · ( M ) - - - ( 15 )
Wherein
ΔJ N ( M ) = ∂ J g 1 ( N ) ∂ θ 1 [ θ 1 ( M ) - θ 1 ( N ) ] ... ∂ J g n ( N ) ∂ θ n [ θ n ( M ) - θ n ( N ) ] - - - ( 16 )
The robot for space joint angle speed in M moment is expressed as:
θ · ( M ) = θ · ( M ) · θ · ( N ) | θ · ( N ) | 2 2 θ · ( N ) + E θ ( M ) - - - ( 17 )
Wherein
E θ ( M ) = θ · ( M ) - θ · ( M ) · θ · ( N ) | θ · ( N ) | 2 2 θ · ( N ) - - - ( 18 )
When moving due to robot for space, joint angle velocity variations is slow, so formula (17) Section 1 is M moment space machine The major part of device person joint's angular velocity, Eθ(M) it is the remainder relevant to acceleration;
Bring formula (17) into formula (15), obtain the robot for space tip speed in M moment and be expressed as:
v e ( M ) ω e ( M ) = Σ i = 1 n J g i ( M ) θ · i ( M ) = θ · ( M ) · θ · ( N ) | θ · ( N ) | 2 2 J g ( N ) θ · ( N ) + ΔJ N ( M ) θ · ( M ) · θ · ( N ) | θ · ( N ) | 2 2 θ · ( N ) + E v ( M ) = θ · ( M ) · θ · ( N ) | θ · ( N ) | 2 2 v e ( M ) ω e ( M ) + ΔJ N ( M ) θ · ( M ) · θ · ( N ) | θ · ( N ) | 2 2 θ · ( N ) + E v ( M ) - - - ( 19 )
Wherein
E V ( M ) = J g ( N ) E θ ( M ) + Δ J ( M ) E θ ( M ) + E J ( M ) θ · ( M ) - - - ( 20 )
Current time is n-hour, and vision measurement time delay is P sampling period, according to formula (19), obtains the tip speed in N-P moment For:
v e ( N - P ) ω e ( N - P ) = θ · ( N - P ) · θ · ( N ) | θ · ( N ) | 2 2 v e ( M ) ω e ( M ) + ΔJ N ( N - P ) θ · ( N - P ) · θ · ( N ) | θ · ( N ) | 2 2 θ · ( N ) + E V ( N - P ) - - - ( 21 )
According to formula (19), obtaining the tip speed in N-P-1 moment is:
v e ( N - P - 1 ) ω e ( N - P - 1 ) = θ · ( N - P - 1 ) · θ · ( N ) | θ · ( N ) | 2 2 v e ( N ) ω e ( N ) + ΔJ N ( N - P - 1 ) θ · ( N - P - 1 ) · θ · ( N ) | θ · ( N ) | 2 2 θ · ( N ) + E V ( N - P - 1 ) - - - ( 22 )
And the N-P-1 moment is for the target rough estimate matrix Δ J of n-hourNIt is represented by:
ΔJ N ( N - P - 1 ) = ∂ J g 1 ( N ) ∂ θ 1 [ θ 1 ( N - P - 1 ) - θ 1 ( N ) ] ... ∂ J g n ( N ) ∂ θ n [ θ n ( N - P - 1 ) - θ n ( N ) ] = [ θ 1 ( N - P - 1 ) - θ 1 ( N ) ] · [ θ 1 ( N - P ) - θ 1 ( N ) ] | [ θ 1 ( N - P ) - θ 1 ( N ) ] | 2 2 ΔJ N ( N - P ) + T Δ J - - - ( 23 )
Wherein TΔJFor remainder, according to formula (19), formula (21) and formula (22), can obtain the tip speed of current n-hour is:
v e ( N ) ω e ( N ) = - β α 1 ( 1 - β ) v e ( N - P ) ω e ( N - P ) + 1 α 2 ( 1 - β ) v e ( N - P - 1 ) ω e ( N - P - 1 ) + Δ E - - - ( 24 )
Wherein
β = [ θ 1 ( N - P - 1 ) - θ 1 ( N ) ] · [ θ 1 ( N - P ) - θ 1 ( N ) ] | [ θ 1 ( N - P ) - θ 1 ( N ) ] | 2 2 α 1 = θ · ( N - P ) · θ · ( N ) | θ · ( N ) | 2 2 α 2 = θ · ( N - P - 1 ) · θ · ( N ) | θ · ( N ) | 2 2 - - - ( 25 )
α1、α2For second-order linearity estimation coefficient, β is state difference item;Ignoring high-order error delta E, now tip speed can approximate meter It is:
v ~ e ( N ) ω ~ e ( N ) = - β α 1 ( 1 - β ) v e ( N - P ) ω e ( N - P ) + 1 α 2 ( 1 - β ) v e ( N - P - 1 ) ω e ( N - P - 1 ) - - - ( 26 )
Estimation difference is:
Δ E = - β α 1 ( 1 - β ) E V ( N - P ) + 1 α 2 ( 1 - β ) [ T Δ J θ · ( N ) + E V ( N - P - 1 ) ] - - - ( 27 )
According to formula (10) and formula (11), estimating speedFor:
V ~ r ( k ) = x · r ( k ) y · r ( k ) z · r ( k ) α · r ( k ) β · r ( k ) γ · r ( k ) T = v e 0 - sinα r cosα r cosβ r 0 cosα r sinα r cosβ r 1 0 - sinβ r - 1 ω e - - - ( 28 ) .
A kind of robot for space vision time delay error compensating method based on velocity estimation the most according to claim 2, its Being characterised by: in step 3, described error controller is designed as:
The process of the Velocity Estimation sequence described in step 2 is regarded as a differentiation element with error:
P ( s ) = e ~ st s = e st s [ 1 + H Δ ( s ) ] - - - ( 29 )
WhereinFor the error transfer function of differentiation element, the transmission function of vision time delay error control system output is:
Y ( s ) = e - st s e ~ st s + G ( s ) e - st s 1 + G ( s ) e - st s V r ( s ) = V r ( s ) + H Δ ( s ) 1 + G ( s ) e - st s V r ( s ) - - - ( 30 )
Wherein VrS () is the actual value of robot end's speed, Y (s) is robot end's Velocity Estimation after correction, and G (s) is for by mistake Difference controller;
The estimation difference transmission function of system is:
E ( s ) = - H Δ ( s ) e - st s - G ( s ) e - 2 st s 1 + G ( s ) e - st s V r ( s ) - - - ( 31 )
By POLE PLACEMENT USING, design makes error controller G (s) stable for E (s), and estimation difference E (t) the most gradually restrains, school After just, robot end's Velocity Estimation gradually levels off to speed actual value.
A kind of robot for space vision time delay error compensating method based on velocity estimation the most according to claim 3, its It is characterised by: in step 4, described according to correction rear space robot end's Velocity Estimation, to currently regarding with time delay Feel that measurement data compensates, through the vision measurement data of overcompensationIt is calculated as follows:
D ~ r ( N ) = D v ( N ) + t s × Σ i = 1 m Y ( N - i ) - - - ( 32 )
Wherein Y (N-i) is correction rear space robot end's Velocity Estimation that step 4 draws, so far completes vision time delay by mistake The compensation of difference.
CN201410138351.9A 2014-04-08 2014-04-08 Robot for space vision time delay error compensating method based on velocity estimation Expired - Fee Related CN103878770B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410138351.9A CN103878770B (en) 2014-04-08 2014-04-08 Robot for space vision time delay error compensating method based on velocity estimation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410138351.9A CN103878770B (en) 2014-04-08 2014-04-08 Robot for space vision time delay error compensating method based on velocity estimation

Publications (2)

Publication Number Publication Date
CN103878770A CN103878770A (en) 2014-06-25
CN103878770B true CN103878770B (en) 2016-08-31

Family

ID=50948051

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410138351.9A Expired - Fee Related CN103878770B (en) 2014-04-08 2014-04-08 Robot for space vision time delay error compensating method based on velocity estimation

Country Status (1)

Country Link
CN (1) CN103878770B (en)

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104408299B (en) * 2014-11-17 2019-02-15 广东产品质量监督检验研究院 Robot location's error compensating method based on distance identification redundancy kinematics parameters
US10635758B2 (en) * 2016-07-15 2020-04-28 Fastbrick Ip Pty Ltd Brick/block laying machine incorporated in a vehicle
CN108267960A (en) * 2018-02-01 2018-07-10 南阳师范学院 A kind of motion control method of remote operating wheel robot
CN108423427A (en) * 2018-03-05 2018-08-21 菲尼克斯(南京)智能制造技术工程有限公司 Vacuum sucking device and method
CN108714914B (en) * 2018-03-19 2021-09-07 山东超越数控电子股份有限公司 Mechanical arm vision system
CN109606753B (en) * 2018-11-11 2022-03-29 上海宇航系统工程研究所 Control method for cooperatively capturing target by space double-arm robot
CN110109343B (en) * 2019-04-30 2021-09-10 东南大学 Visual measurement time-lag compensation method for photoelectric tracking system
CN110900581B (en) * 2019-12-27 2023-12-22 福州大学 Four-degree-of-freedom mechanical arm vision servo control method and device based on RealSense camera
CN112847323B (en) * 2021-01-06 2022-08-09 中国铁建重工集团股份有限公司 Robot model parameter error compensation method, device, electronic device and medium
CN113427487B (en) * 2021-07-09 2022-03-25 华南理工大学 DH parameter calibration method and system based on electromagnetic wave ranging
CN113799127B (en) * 2021-09-15 2023-05-23 华南理工大学 Six-degree-of-freedom mechanical arm nonstandard positioning pose positioning method under optical binocular positioning system

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4985668A (en) * 1989-09-19 1991-01-15 Kabushiki Kaisha Kobe Seiko Sho Robot controller
CN103302668A (en) * 2013-05-22 2013-09-18 东南大学 Kinect-based space teleoperation robot control system and method thereof

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0727408B2 (en) * 1984-01-19 1995-03-29 株式会社日立製作所 Robot handling device with fixed 3D vision

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4985668A (en) * 1989-09-19 1991-01-15 Kabushiki Kaisha Kobe Seiko Sho Robot controller
CN103302668A (en) * 2013-05-22 2013-09-18 东南大学 Kinect-based space teleoperation robot control system and method thereof

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
史国振.视觉伺服空间机器人运动硬件仿真研究.《系统仿真学报》.2008,第20卷(第13期),第3566-3570页. *
陈俊杰.空问机器人遥操作克服时延影响的研究进展.《测控技术》.2007,第26卷(第2期),第1-4页. *

Also Published As

Publication number Publication date
CN103878770A (en) 2014-06-25

Similar Documents

Publication Publication Date Title
CN103878770B (en) Robot for space vision time delay error compensating method based on velocity estimation
Jiancheng et al. Study on innovation adaptive EKF for in-flight alignment of airborne POS
CN108036785A (en) A kind of aircraft position and orientation estimation method based on direct method and inertial navigation fusion
CN101726295B (en) Unscented Kalman filter-based method for tracking inertial pose according to acceleration compensation
CN106772524B (en) A kind of agricultural robot integrated navigation information fusion method based on order filtering
CN103776446B (en) A kind of pedestrian's independent navigation computation based on double MEMS-IMU
CN105628027B (en) A kind of accurate real-time location method of indoor environment based on MEMS inertia device
CN106289246A (en) A kind of rods arm measure method based on position and orientation measurement system
CN106197428A (en) A kind of SLAM method utilizing metrical information Optimum distribution formula EKF estimation procedure
CN104964685A (en) Judgment method for moving state of mobile phone
CN108490433A (en) Deviation Combined estimator and compensation method and system when sky based on Sequential filter
CN106979780A (en) A kind of unmanned vehicle real-time attitude measuring method
CN104613965B (en) A kind of step-by-step movement pedestrian navigation method based on bidirectional filtering smoothing technique
Wang On adaptive inverse dynamics for free-floating space manipulators
CN107909614A (en) Crusing robot localization method under a kind of GPS failures environment
CN104015191B (en) Based on the motion compensation process under the space manipulator tool coordinates of base satellite angular speed
CN111060099A (en) Real-time positioning method for unmanned automobile
CN104457741A (en) Human arm movement tracing method based on ant colony algorithm error correction
Gao et al. Invariant filtering for legged humanoid locomotion on a dynamic rigid surface
CN110470297A (en) A kind of attitude motion of space non-cooperative target and inertial parameter estimation method
CN103968844A (en) Large ellipse maneuverable spacecraft autonomous navigation method based on low-orbit platform tracking measurement
CN108592907A (en) A kind of quasi real time step-by-step movement pedestrian navigation method based on bidirectional filtering smoothing technique
Sun et al. An improved monocular visual-inertial navigation system
CN108168545A (en) A kind of Compensation for Coning Error algorithm of quasi-Newton method Optimization Compensation coefficient
CN104807465A (en) Method and device for realizing simultaneous localization and mapping of robots

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20160831

CF01 Termination of patent right due to non-payment of annual fee