CN110060277A - A kind of vision SLAM method of multiple features fusion - Google Patents
A kind of vision SLAM method of multiple features fusion Download PDFInfo
- Publication number
- CN110060277A CN110060277A CN201910357796.9A CN201910357796A CN110060277A CN 110060277 A CN110060277 A CN 110060277A CN 201910357796 A CN201910357796 A CN 201910357796A CN 110060277 A CN110060277 A CN 110060277A
- Authority
- CN
- China
- Prior art keywords
- feature
- point
- line
- characteristic
- vision
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 79
- 230000004927 fusion Effects 0.000 title claims abstract description 15
- 238000005457 optimization Methods 0.000 claims abstract description 26
- 238000010276 construction Methods 0.000 claims abstract description 6
- 230000003044 adaptive effect Effects 0.000 claims abstract description 4
- 230000000007 visual effect Effects 0.000 claims abstract 3
- 239000000284 extract Substances 0.000 claims abstract 2
- 230000033001 locomotion Effects 0.000 claims description 30
- 238000013507 mapping Methods 0.000 claims description 18
- 239000011159 matrix material Substances 0.000 claims description 15
- 239000000203 mixture Substances 0.000 claims description 4
- 229910002056 binary alloy Inorganic materials 0.000 claims description 3
- 238000001514 detection method Methods 0.000 claims description 3
- 230000007704 transition Effects 0.000 claims description 3
- 238000004804 winding Methods 0.000 claims description 3
- 238000009472 formulation Methods 0.000 claims 1
- 238000007781 pre-processing Methods 0.000 claims 1
- 238000002203 pretreatment Methods 0.000 claims 1
- 230000007423 decrease Effects 0.000 abstract description 2
- 230000001174 ascending effect Effects 0.000 description 6
- 238000010586 diagram Methods 0.000 description 4
- 238000000605 extraction Methods 0.000 description 3
- 230000004807 localization Effects 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 230000015572 biosynthetic process Effects 0.000 description 2
- 238000012937 correction Methods 0.000 description 2
- 238000001914 filtration Methods 0.000 description 2
- 238000003780 insertion Methods 0.000 description 2
- 230000037431 insertion Effects 0.000 description 2
- 238000003786 synthesis reaction Methods 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000004392 development of vision Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000000452 restraining effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/80—Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Data Mining & Analysis (AREA)
- Multimedia (AREA)
- Software Systems (AREA)
- Geometry (AREA)
- Computer Graphics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Image Analysis (AREA)
Abstract
The vision SLAM method of multiple features fusion is related to Robot visual location and builds figure field.The multiple features fusion vision SLAM method based on depth camera that the invention discloses a kind of, by sufficiently using the dotted line feature extracted from image and according to dotted line feature construction plane characteristic, to solve the problems, such as the vision positioning under pure point feature failure conditions.Point feature is extracted using a kind of adaptive threshold method, to obtain more uniform point feature;It extracts line feature and deletes short and small line segment, merges divided line segment, to improve the accuracy rate of line characteristic matching;Dotted line feature is for the estimation of interframe pose and the building of local map;Region feature is calculated using minimum parameter method, to reduce calculation amount;By constructing back projection's error function of fusion feature, by dotted line region feature close-coupled, and constructs global map and carry out global pose optimization.The present invention is that a kind of precision is high, real-time is good, strong robustness vision SLAM method, solves the problems, such as the vision SLAM accuracy decline even thrashing under low texture environment based on method of characteristic point.
Description
Technical field:
The invention belongs to robot simultaneous localization and mapping technical fields, and in particular to a kind of multiple features fusion
SLAM (simultaneous localization and mapping) method.
Background technique:
With the development of vision SLAM technology, optimization and figure Optimization Framework based on interframe have become vision SLAM and ask
Estimation and light-stream adjustment are introduced into vision SLAM by the mainstream frame of topic, figure Optimization Framework, estimation be by
The position of robot and ambient enviroment feature are solved as Global Optimal Problem, carry out feature by the feature extracted on image
Tracking, establishes error function, constructs linear optimization by linear hypothesis or carries out nonlinear optimization directly to solve so that error
Robot pose when function obtains minimum value optimizes road sign simultaneously.Previous motion structure restores (Structure from
Motion, SFM) in feature extraction and matching and subsequent optimization link the too many time is consumed, therefore can only carry out offline
Pose optimization and three-dimensional reconstruction, can not complete to real-time online itself positioning and map structuring.As light beam is flat these years
The sparsity of poor method is found and the upgrading of computer hardware, and the time of links consumption substantially reduces, therefore base
It can be realized positioning in real time in the vision SLAM of figure Optimization Framework and build figure.
Vision SLAM is broadly divided into two class of the direct method based on gray scale and the indirect method based on feature.Wherein based on gray scale
Direct method shows well in some cases, does not need to carry out feature extracting and matching, need to only carry out the tracking of shade of gray, because
This can save a large amount of calculating time, and this method can enhance the continuity of vision SLAM in the less region of feature.
But simultaneously as direct method only carries out matching and the motion tracking of interframe, therefore this method light using the gradient of pixel
It is more strong according to influences such as intensity, noise jammings.Although the indirect method based on feature can occupy part computing resource, due to
The raising of level of hardware, the feature that the computing resource occupied is substantially reduced, and extracted are engineer, are had relatively stable
Characteristic, and be able to carry out signature tracking, composition, closed loop detection, robot simultaneous localization and mapping can be completed
Overall process, therefore the vision SLAM method based on feature has been increasingly becoming the mainstream of SLAM technical research.
Summary of the invention:
The present invention is special in order to solve the point that the existing vision SLAM algorithm based on point feature calculates under low texture environment
Sign can not carry out stable tracking, and be declined according to the restraining force that point feature generates, and can not construct accurate environmental map
The problems such as, the present invention provides a kind of vision SLAM methods of multiple features fusion.
SLAM problem is modeled first are as follows:
It is the six-vector on a SE (3), indicates the movement at robot moment from t to t+1,Indicate fortune
The covariance matrix of dynamic estimation is approached by the Hessian matrix of loss function in last iteration Lai inverse.
The technical scheme adopted by the invention is that: the vision SLAM method of the point, line, surface Fusion Features based on image,
It is characterized in that, comprising the following steps:
Step 1, the calibration for carrying out depth camera determines the internal reference of camera;
Step 2, the video stream data obtained for camera on mobile robot platform carries out gaussian filtering to reduce noise
Influence to subsequent step;
Step 3, feature extraction, the point feature and line feature of each frame of On-line testing, tool are carried out for the image after correction
Body includes: to describe son for each frame image zooming-out FAST corner feature with rBRIEF and be described, and line feature is calculated using LSD
Method is detected, and is described son using LBD and be described, and the geological information provided using line segment filters out direction and length not
The same and biggish line segment of end-point distances;
Step 4, according to point, the line feature construction region feature extracted, it is special that candidate face is gone out by following three kinds of method constructs
Sign: 1) three coplanar characteristic points;2) characteristic point and a characteristic curve;3) two coplanar characteristic curves;
Step 5, the frame matching that image is carried out using the point between present frame and previous frame, line feature, obtains dotted line feature
Matching pair, establish the geometrical correspondence of interframe;
Step 6, feature back projection is carried out according to the obtained interframe corresponding relationship of step 5, uses Gauss-Newton method minimum
Change back projection's error of feature, in order to handle exceptional value, executes two steps using pseudo- Hans Huber's cost function and minimize, acquire
The incremental motion of two continuous interframe;
Step 7, the selection for carrying out key frame, by the uncertainty of covariance matrix by way of following formula is converted to entropy:
H (ζ)=3 (1+log (2 π))+0.5log (| Ωζ|),
The covariance matrix of Ω expression incremental motion.Ratio calculated:
If α < 0.9, and present frame is more than 300 with the characteristic that is overlapped of previous keyframe, then the image is added to pass
In key frame list, otherwise return step 2, continue with the image data of input;
Step 8, the thought based on figure establishes local map, by all with the related key frame of present frame and according to this
The feature composition that a little key frames observe, defining ψ is the variable under se (3), incremental motion including the continuous interframe of every two, every
The space coordinate point of a point feature, the space coordinate point of line feature endpoint and the minimum of plane indicate vector;Then it minimizes
Data are observed at a distance from road sign projection, construct accurate local map;
Step 9, judge whether the track of robot motion forms closed loop, if so, the local map to both ends is repaired
Just, still using the thought of figure optimization, the robot pose that key frame is represented is closed as node, the constraint of both ends local map
System carries out the amendment of track as side;Otherwise stop signal is judged whether there is, if being not received by stop signal, returns to step
Rapid 2 carry out vision SALM process, otherwise enter step 10;
Step 10, stop motion.
Further, the implementation method of step 1:
The gridiron pattern image data for obtaining the fixed size under multiple different perspectivess using camera in advance, then by opening just
Friendly camera calibration method carries out the calculating of camera internal reference to the gridiron pattern image data got, obtains the result of camera calibration.
Further, the implementation of step 3:
Step 3.1, the FAST corner feature of each frame image is extracted using M16 template, M16 template is as shown in Fig. 2, if
There is the pixel of continuous 12 points to be all larger than Ip+ T is less than Ip- T is then characterized a little;An adaptive threshold T is calculated first:
WhereinIndicate the average value of all pixels gray value in current circle;Judge 1,5,9,13 4 point of gray value with
The size relation of T, when at least there are three point meet above-mentioned condition when continue to calculate remaining 12 point, according to gray scale centroid method determine
The direction of characteristic point: the position C of mass center, the square of neighborhood S are calculated in neighborhood S is defined as:
Mass center are as follows:
The direction of characteristic point are as follows:
2 (m of θ=atan01,m10),
The corner feature extracted is described by rBRIEF description with rotational invariance;
Step 3.2, line segment feature that may be present is detected using LSD algorithm for image, and merging may be same
The line segment of straight line, by calculate any two line feature angle and respective line segment midpoint to another line segment distance with,
If being respectively smaller than threshold value TlangAnd Tldis, then it is assumed that be same line segment, merge, later using LBD algorithm to optimization after
Line feature is described, and obtains description of line feature.
Further, the implementation method of step 4:
Step 4.1, space coordinate solution is carried out for the dotted line feature that step 3 is extracted, particularly, line feature seeks it
The equation that the space line belonging to it is calculated after space coordinate, is divided into two set for feature later, point feature is classified as SpCollection
It closes, line feature is classified as SlSet;
Step 4.2, candidate region feature: 1) three coplanar characteristic points is gone out by following three kinds of method constructs;2) spy
Sign point and a characteristic curve;3) two coplanar characteristic curves;And calculate the general equation of candidate plane: p1x+p2y+p3Z=p4;
Step 4.3, the coefficient of plane equation is normalized:
||[p1,p2,p3,p4]T| |=1,
Plane: p=[p is indicated with three amounts with this1,p2,p3]T;
Further, the implementation method of step 5:
Step 5.1, it is matched according to the point, line, surface feature obtained in step 3, for point feature, according to descriptor
Distance takes preceding 60 percent to be used as optimal matching points according to ascending sort;For line feature, also according to description away from
From according to ascending sort, preceding 50 percent is taken to be used as best match pair;For region feature, calculate:
According to ascending sort, 3-5 is used as best match pair before taking;
Step 5.2, using RANSAC matrix to obtained matching to carry out error characteristic rejecting;
Further, the implementation method of step 6:
Step 6.1, three-dimensional reconstruction is carried out according to the feature matched, three-dimensional coordinate is obtained, further according to the several of adjacent interframe
What relationship calculates the re-projection error of feature to re-projection is carried out, and is missed using the projection that gauss-newton method minimizes point and straight line
Difference;
Step 6.2, it using pseudo- Hans Huber's loss function and executes primary two step and minimizes process, finally obtain two companies
The incremental motion of continuous key interframe.
Further, the implementation method of step 7:
Step 7.1, the uncertainty in covariance matrix is converted to the form of the scalar of entropy:
H (ζ)=3 (1+log (2 π))+0.5log (| Ωζ|),
Wherein Ω indicates the covariance matrix of incremental motion, for giving input frame image, checks previous keyframe i and works as
The entropy of estimation between preceding frame i+u and its ratio with first key frame i+1 calculate:
If the value of α is lower than some threshold value, 0.9, and the feature in present frame and previous keyframe are defined as in this method
Coincidence number is more than 300, then using i+u frame as new key frame insertion system;
Step 7.2, due to ζt,t+1The only estimation of adjacent interframe, thus it is above-mentioned by the synthesis of single order Gauss transmission function
A series of estimations, with obtain two discontinuous interframe covariance.
Further, the implementation method of step 8:
Step 8.1, the variation estimation of the relative pose for present frame and before between all key frames refines, into
Row data correlation, the incremental motion for using step 6 to obtain is as the initial input of Gauss-Newton optimization method;
Step 8.2, defining ψ is the variable under se (3), incremental motion, each point feature including the continuous interframe of every two
Space coordinate point, the space coordinate point of line feature endpoint and the minimum of plane indicate vector, minimize re-projection error:
Wherein Kl、Pl、Ll、MlLocal map set, point feature set, line characteristic set, region feature set are respectively indicated,
The covariance matrix of Ω expression incremental motion.Point feature projection error e thereinijIt is that j-th of mapping point in the i-th frame observes
Two-dimensional distance, according to the following formula calculate:
eij=xij-π(ζiw,Xwj),
π refers to the mapping from se (3) to (3) SE, X in formulawjIt is the three dimensional space coordinate of point feature;The throwing of its middle line feature
Shadow error is different from the projection error of point feature, the endpoint of three-dimensional line segment is projected to the plane of delineation, and it is put down with image
The distance between corresponding Infinite Straight Line is used as error function, e on faceikIt is calculated by following formula:
P, Q are respectively the extreme coordinates of line feature, l in formulaikThe linear equation in plane is represented, is provided by step 3;Face is special
The projection error e of signimIt is calculated according to the logarithmic mapping of quaternary number:
eim=log { q [T (ζiw)Tp(ζim)]-1q(ζiw),
P (ζ in formulaim) indicate plane, it is provided by step 3, q (*) is a quaternary number, T (ζiw) it is a transition matrix
Homogeneous form.Second generation update is carried out using Levenberg-Marquardt optimization method, acquires optimal incremental motion.
Further, the implementation method of step 9:
Step 9.1, the description of BRIEF binary system is carried out to the line feature extracted, description of line feature is obtained, with a spy
The vision entry vector that description of sign constitutes present frame is matched according to the vision bag of words established offline in advance;
Step 9.2, it if successful match, needs to optimize merging to two local maps where the two key frames,
And update with the associated local map of the two local maps, be based on figure Optimization Framework, using interframe incremental motion as figure
The construction of node, side is similar to step 8, according to:
It is updated to be iterated, log therein and exp function respectively indicate logarithmic mapping and Lee of the Lie group to Lie algebra
Index mapping of the algebra to Lie group;
Step 9.3, if matching is unsuccessful, stop signal is judged whether there is, if being not received by stop signal, returns to step
Rapid 2 continue to run, and otherwise enter step 10.
The invention has the following beneficial effects:
1. carrying out close-coupled using multiple features, can effectively solve to be based on tag system precision sharply under low texture environment
The problems such as decline, system down;
2. can be very good to realize that the locating effect of mobile platform and building have structural information according to doors structure environment
Ambient enviroment feature, can be very good to obtain on a variety of open experimental data sets high-precision as a result, it is possible in real time, efficiently
The pose of mobile platform and the environment of surrounding are patterned using matched dotted line region feature, and carry out at winding detection
Reason makes full use of dotted line feature to reduce accumulated error.
Detailed description of the invention:
Fig. 1 is a kind of flow diagram of the vision SLAM method of multiple features fusion described in specific embodiment;
Fig. 2 is FAST feature M16 template schematic diagram used in this paper SLAM method;
Fig. 3 is this paper SLAM method probability graph model used based on figure optimization thought;
Fig. 4 is plane characteristic schematic diagram used in this paper SLAM method;
Fig. 5 is this paper SLAM method dotted line feature re-projection schematic diagram.
Specific embodiment:
Below in conjunction with attached drawing and specific embodiment, the present invention is described further.
The technical scheme adopted by the invention is that: vision SLAM method based on multi-feature fusion, specific implementation process are shown in
Fig. 1 is mainly comprised the steps that
Step 1, the calibration for carrying out depth camera determines the internal reference of camera;
Step 2, the video stream data obtained for camera on mobile robot platform carries out gaussian filtering to reduce noise
Influence to subsequent step;
Step 3, feature extraction, the point feature and line feature of each frame of On-line testing, tool are carried out for the image after correction
Body includes: to describe son for each frame image zooming-out FAST corner feature with rBRIEF and be described, and line feature is calculated using LSD
Method is detected, and is described son using LBD and be described, and the geological information provided using line segment filters out direction and length not
The same and biggish line segment of end-point distances;
Step 4, according to point, the line feature construction region feature extracted, it is special that candidate face is gone out by following three kinds of method constructs
Sign: 1) three coplanar characteristic points;2) characteristic point and a characteristic curve;3) two coplanar characteristic curves;
Step 5, the frame matching that image is carried out using the point between present frame and previous frame, line feature, obtains dotted line feature
Matching pair, establish the geometrical correspondence of interframe;
Step 6, feature back projection is carried out according to the obtained interframe corresponding relationship of step 5, uses Gauss-Newton method minimum
Change back projection's error of feature, in order to handle exceptional value, executes two steps using pseudo- Hans Huber's cost function and minimize, acquire
The incremental motion of two continuous interframe;
Step 7, the selection for carrying out key frame, the uncertainty of covariance matrix is passed through formula, and --- --- -- is converted to entropy
Form:
H (ζ)=3 (1+log (2 π))+0.5log (| Ωζ|),
Ratio calculated:
If α < 0.9, which is added in Key Frames List, otherwise return step 2, the figure of input is continued with
As data;
Step 8, the thought based on figure establishes local map, by all with the related key frame of present frame and according to this
The feature composition that a little key frames observe, defining ψ is the variable under se (3), incremental motion including the continuous interframe of every two, every
The space coordinate point of a point feature, the space coordinate point of line feature endpoint and the minimum of plane indicate vector;Then it minimizes
Data are observed at a distance from road sign projection, construct accurate local map;
Step 9, judge whether the track of robot motion forms closed loop, if so, the local map to both ends is repaired
Just, still using the thought of figure optimization, the robot pose that key frame is represented is closed as node, the constraint of both ends local map
System carries out the amendment of track as side;Otherwise, it is determined whether there is stop signal, if being not received by stop signal, return
Step 2 continues vision SALM process, otherwise enters step 10;
Step 10, stop motion.
Further, the implementation method of step 1:
The gridiron pattern image data for obtaining the fixed size under multiple different perspectivess using camera in advance, then by opening just
Friendly camera calibration method carries out the calculating of camera internal reference to the gridiron pattern image data got, obtains the result of camera calibration.
Further, the implementation of step 3:
Step 3.1, the FAST corner feature of each frame image is extracted using M16 template, M16 template is as shown in Fig. 2, if
There is the pixel of continuous 12 points to be all larger than Ip+ T is less than Ip- T is then characterized a little;An adaptive threshold T is calculated first:
WhereinIndicate the average value of all pixels gray value in current circle;Judge 1,5,9,13 4 point of gray value with
The size relation of T, when at least there are three point meet above-mentioned condition when continue to calculate remaining 12 point, according to gray scale centroid method determine
The direction of characteristic point: the position C of mass center, the square of neighborhood S are calculated in neighborhood S is defined as:
Mass center are as follows:
The direction of characteristic point are as follows:
2 (m of θ=atan01,m10),
The corner feature extracted is described by rBRIEF description with rotational invariance;
Step 3.2, line segment feature that may be present is detected using LSD algorithm for image, and merging may be same
The line segment of straight line, by calculate any two line feature angle and respective line segment midpoint to another line segment distance with,
If being respectively smaller than threshold value TlangAnd Tldis, then it is assumed that be same line segment, merge, later using LBD algorithm to optimization after
Line feature is described, and obtains description of line feature.
Further, the implementation method of step 4:
Step 4.1, space coordinate solution is carried out for the dotted line feature that step 3 is extracted, particularly, line feature seeks it
The equation that the space line belonging to it is calculated after space coordinate, is divided into two set for feature later, point feature is classified as SpCollection
It closes, line feature is classified as SlSet;
Step 4.2, candidate region feature: 1) three coplanar characteristic points is gone out by following three kinds of method constructs;2) spy
Sign point and a characteristic curve;3) two coplanar characteristic curves;And calculate the general equation of candidate plane: p1x+p2y+p3Z=p4;
Step 4.3, the coefficient of plane equation is normalized:
||[p1,p2,p3,p4]T| |=1,
Plane: [p is indicated with three amounts with this1,p2,p3]T;
Further, the implementation method of step 5:
Step 5.1, it is matched according to the point, line, surface feature obtained in step 3, for point feature, according to descriptor
Distance takes preceding 60 percent to be used as optimal matching points according to ascending sort;For line feature, also according to description away from
From according to ascending sort, preceding 50 percent is taken to be used as best match pair;For line feature, calculate:
According to ascending sort, 3-5 is used as best match pair before taking;
Step 5.2, using RANSAC matrix to obtained matching to carry out error characteristic rejecting;
Further, the implementation method of step 6:
Step 6.1, three-dimensional reconstruction is carried out according to the feature matched, three-dimensional coordinate is obtained, further according to the several of adjacent interframe
What relationship calculates the re-projection error of feature to re-projection is carried out, and is missed using the projection that gauss-newton method minimizes point and straight line
Difference;
Step 6.2, it using pseudo- Hans Huber's loss function and executes primary two step and minimizes process, finally obtain two companies
The incremental motion of continuous key interframe.
Further, the implementation method of step 7:
Step 7.1, the uncertainty in covariance matrix is converted to the form of the scalar of entropy:
H (ζ)=3 (1+log (2 π))+0.5log (| Ωζ|),
For given input frame image, check the estimation between previous keyframe i and current frame i+u entropy and
The ratio of itself and first key frame i+1 calculates:
If the value of α is lower than some threshold value, 0.9, and the feature in present frame and previous keyframe are defined as in this method
Coincidence number is more than 300, then using i+u frame as new key frame insertion system;
Step 7.2, due to ζt,t+1The only estimation of adjacent interframe, thus it is above-mentioned by the synthesis of single order Gauss transmission function
A series of estimations, with obtain two discontinuous interframe covariance.
Further, the implementation method of step 8:
Step 8.1, the variation estimation of the relative pose for present frame and before between all key frames refines, into
Row data correlation, the incremental motion for using step 6 to obtain is as the initial input of Gauss-Newton optimization method;
Step 8.2, defining ψ is the variable under se (3), incremental motion, each point feature including the continuous interframe of every two
Space coordinate point, the space coordinate point of line feature endpoint and the minimum of plane indicate vector, minimize re-projection error:
Wherein Kl、Pl、Ll、MlRespectively indicate local map set, point feature set, line characteristic set, region feature set.
Point feature projection error e thereinijIt is the two-dimensional distance that j-th of mapping point in the i-th frame observes, calculates according to the following formula:
eij=xij-π(ζiw,Xwj),
π refers to the mapping from se (3) to (3) SE, X in formulawjIt is the three dimensional space coordinate of point feature;The throwing of its middle line feature
Shadow error is different from the projection error of point feature, the endpoint of three-dimensional line segment is projected to the plane of delineation, and it is put down with image
The distance between corresponding Infinite Straight Line is used as error function, e on faceikIt is calculated by following formula:
P, Q are respectively the extreme coordinates of line feature, l in formulaikThe linear equation in plane is represented, is provided by step 3;Face is special
The projection error e of signimIt is calculated according to the logarithmic mapping of quaternary number:
eim=log { q [T (ζiw)Tp(ζim)]-1q(ζiw),
P (ζ in formulaim) indicate plane, it is provided by step 3, q (*) is a quaternary number, T (ζiw) it is a transition matrix
Homogeneous form.Second generation update is carried out using Levenberg-Marquardt optimization method, acquires optimal incremental motion.
Further, the implementation method of step 9:
Step 9.1, the description of BRIEF binary system is carried out to the line feature extracted, description of line feature is obtained, with a spy
The vision entry vector that description of sign constitutes present frame is matched according to the vision bag of words established offline in advance;
Step 9.2, it if successful match, needs to optimize merging to two local maps where the two key frames,
And update with the associated local map of the two local maps, be based on figure Optimization Framework, using interframe incremental motion as figure
The construction of node, side is similar to step 8, according to:
It is updated to be iterated, log therein and exp function respectively indicate logarithmic mapping and Lee of the Lie group to Lie algebra
Index mapping of the algebra to Lie group;
Step 9.3, if matching is unsuccessful, stop signal is judged whether there is, if being not received by stop signal, returns to step
Rapid 2 continue to run, and otherwise enter step 10.
Claims (5)
1. a kind of vision SLAM method of multiple features fusion, which is characterized in that the reality of visual odometry front end and figure optimization rear end
Existing method, key step are as follows:
Image preprocessing is carried out to the image of acquisition first in visual odometry front end, to reduce in initial pictures noise for spy
Levy the influence of detection;FAST characteristic point is extracted in image after the pre-treatment using a kind of method of adaptive threshold, is selected
Wherein Harris responds highest 500 characteristic points of score and calculates rBRIEF descriptor;The line feature for extracting image, uses LSD
Algorithm extracts the line segment feature in image, for the line segment feature extracted, needs to filter out length less than 20 pixels, simultaneously
It lays down a regulation, it would be possible to be merged for two line segment features of same line segment feature;It is candidate by the method construct of formulation
Region feature indicates region feature using minimum planes representation;Characteristic matching is carried out to the image of adjacent two frame, obtains camera
The variation of interframe pose, and judge whether present frame is key frame, it is then saved if key frame, and construct local map;
The local map of building is optimized in figure optimization rear end, is played a game position appearance using Levenberg-Marquadt algorithm
It optimizes, solves optimal pose;Judge whether present incoming frame can constitute winding, to the part at both ends if it can constitute winding
Map is modified, and the robot pose that key frame is represented comes as node, the constraint relationship of both ends local map as side
Carry out the amendment of track.
2. a kind of vision SLAM method of multiple features fusion as described in claim 1, which is characterized in that use a kind of new side
Method detects FAST feature:
Using the pixel around M16 template selection characteristic point, radius is 3 pixels, selects 10 gray values most in current circle
Big and 10 the smallest pixels of gray value calculate threshold value T:
Wherein IiIndicate the average value of all pixels gray value in current circle;Compare first in 16 points again number be 1,5,9,
13 pixel and the relationship of T, the gray scale there are three or more judge the pass of these points intermediate left point and T if being greater than T
System, if more than T, then the point is characterized a little.
3. a kind of vision SLAM method of multiple features fusion as described in claim 1, which is characterized in that use minimum planes
It indicates to indicate the region feature extracted:
Using point feature obtained in claim 2, line feature is extracted using LSD algorithm, dotted line feature is classified, point is special
Sign is classified as SpSet, line feature are classified as SlSet, particularly, line feature seek calculating after its space coordinate the space belonging to it
The equation of straight line goes out candidate region feature: 1) three coplanar characteristic points by following three kinds of method constructs;2) characteristic point with
One characteristic curve;3) two coplanar characteristic curves;And calculate the general equation of candidate plane: p1x+p2y+p3Z=p4, and will put down
The coefficient of face equation normalizes: | | [p1,p2,p3,p4]T| |=1, plane: p=[p is indicated with three amounts with this1,p2,
p3]T。
4. a kind of vision SLAM method of multiple features fusion as described in claim 1, which is characterized in that sufficiently use dotted line
The re-projection error of region feature completes the optimization of local map:
Present frame is refined with the variation estimation of the relative pose between all key frames before, carries out data correlation,
The incremental motion for using step 6 to obtain is as the initial input of Gauss-Newton optimization method;Defining ψ is the variable under se (3),
The space coordinate point of incremental motion, each point feature including the continuous interframe of every two, line feature endpoint space coordinate point with
And the minimum of plane indicates vector, minimizes re-projection error:
Wherein Kl、Pl、Ll、MlLocal map set, point feature set, line characteristic set, region feature set are respectively indicated, Ω is indicated
The covariance matrix of incremental motion, point feature projection error eijIt is the two-dimensional distance that j-th of mapping point in the i-th frame observes,
It calculates according to the following formula:
eij=xij-π(ζiw,Xwj),
π refers to the mapping from se (3) to (3) SE, ζ in formulaiwIt is the six-vector in se (3) space, XwjIt is the three of point feature
Dimension space coordinate, xijFor image coordinate;The projection error of its middle line feature and the projection error of point feature are different, by three-dimensional line segment
Endpoint project to the plane of delineation, and regard it as error letter the distance between with Infinite Straight Line corresponding on the plane of delineation
Number, eikIt is calculated by following formula:
P, Q are respectively the extreme coordinates of line feature, l in formulaikThe linear equation in plane is represented, is provided by step 3;Region feature
Projection error eimIt is calculated according to the logarithmic mapping of quaternary number:
eim=log { q [T (ζiw)Tp(ζim)]-1q(ζiw),
P (ζ in formulaim) indicate plane, it is provided by step 3, q (*) is a quaternary number, T (ζiw) it is the homogeneous of a transition matrix
Form.It is iterated update using Levenberg-Marquardt optimization method, acquires optimal incremental motion.
5. a kind of vision SLAM method of multiple features fusion as described in claim 1, which is characterized in that offline building in advance
Based on the bag of words of dotted line feature, successful match structural map Optimized model:
The description of BRIEF binary system is carried out to the line feature extracted, obtains description of line feature, the sub- structure of description with point feature
It is matched at the vision entry vector of present frame according to the vision bag of words established offline in advance;If successful match,
Two local maps where the two key frames need to be optimized with merging, and updated associated with the two local maps
Local map is based on figure Optimization Framework, and using interframe incremental motion as the node of figure, the construction on side is similar to step 8, according to:
It is updated to be iterated, log therein and exp function respectively indicate Lie group to the logarithmic mapping and Lie algebra of Lie algebra
To the index mapping of Lie group.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910357796.9A CN110060277A (en) | 2019-04-30 | 2019-04-30 | A kind of vision SLAM method of multiple features fusion |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910357796.9A CN110060277A (en) | 2019-04-30 | 2019-04-30 | A kind of vision SLAM method of multiple features fusion |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110060277A true CN110060277A (en) | 2019-07-26 |
Family
ID=67321742
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910357796.9A Pending CN110060277A (en) | 2019-04-30 | 2019-04-30 | A kind of vision SLAM method of multiple features fusion |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110060277A (en) |
Cited By (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110599569A (en) * | 2019-09-16 | 2019-12-20 | 上海市刑事科学技术研究院 | Method for generating two-dimensional plane graph inside building, storage device and terminal |
CN110631586A (en) * | 2019-09-26 | 2019-12-31 | 珠海市一微半导体有限公司 | Map construction method based on visual SLAM, navigation system and device |
CN110807799A (en) * | 2019-09-29 | 2020-02-18 | 哈尔滨工程大学 | Line feature visual odometer method combining depth map inference |
CN110926405A (en) * | 2019-12-04 | 2020-03-27 | 中科新松有限公司 | ARV attitude measurement method based on monocular vision vanishing point detection |
CN111275764A (en) * | 2020-02-12 | 2020-06-12 | 南开大学 | Depth camera visual mileage measurement method based on line segment shadow |
CN111310772A (en) * | 2020-03-16 | 2020-06-19 | 上海交通大学 | Point line feature selection method and system for binocular vision SLAM |
CN111369594A (en) * | 2020-03-31 | 2020-07-03 | 北京旋极信息技术股份有限公司 | Method, device, computer storage medium and terminal for realizing target tracking |
CN111461141A (en) * | 2020-03-30 | 2020-07-28 | 歌尔科技有限公司 | Equipment pose calculation method device and equipment |
CN111693047A (en) * | 2020-05-08 | 2020-09-22 | 中国航空工业集团公司西安航空计算技术研究所 | Visual navigation method for micro unmanned aerial vehicle in high-dynamic scene |
CN111899334A (en) * | 2020-07-28 | 2020-11-06 | 北京科技大学 | Visual synchronous positioning and map building method and device based on point-line characteristics |
CN112114966A (en) * | 2020-09-15 | 2020-12-22 | 杭州未名信科科技有限公司 | Light beam adjustment calculation method of visual SLAM |
CN112258391A (en) * | 2020-10-12 | 2021-01-22 | 武汉中海庭数据技术有限公司 | Fragmented map splicing method based on road traffic marking |
CN112305558A (en) * | 2020-10-22 | 2021-02-02 | 中国人民解放军战略支援部队信息工程大学 | Mobile robot track determination method and device by using laser point cloud data |
CN112381890A (en) * | 2020-11-27 | 2021-02-19 | 上海工程技术大学 | RGB-D vision SLAM method based on dotted line characteristics |
CN112419497A (en) * | 2020-11-13 | 2021-02-26 | 天津大学 | Monocular vision-based SLAM method combining feature method and direct method |
CN112631271A (en) * | 2020-10-09 | 2021-04-09 | 南京凌华微电子科技有限公司 | Map generation method based on robot perception fusion |
CN112802196A (en) * | 2021-02-01 | 2021-05-14 | 北京理工大学 | Binocular inertia simultaneous positioning and map construction method based on dotted line feature fusion |
CN112880687A (en) * | 2021-01-21 | 2021-06-01 | 深圳市普渡科技有限公司 | Indoor positioning method, device, equipment and computer readable storage medium |
CN112945233A (en) * | 2021-01-15 | 2021-06-11 | 北京理工大学 | Global drift-free autonomous robot simultaneous positioning and map building method |
CN112950709A (en) * | 2021-02-21 | 2021-06-11 | 深圳市优必选科技股份有限公司 | Pose prediction method, pose prediction device and robot |
WO2021135645A1 (en) * | 2019-12-31 | 2021-07-08 | 歌尔股份有限公司 | Map updating method and device |
CN113608236A (en) * | 2021-08-03 | 2021-11-05 | 哈尔滨智兀科技有限公司 | Mine robot positioning and image building method based on laser radar and binocular camera |
CN113689499A (en) * | 2021-08-18 | 2021-11-23 | 哈尔滨工业大学 | Visual rapid positioning method, device and system based on point-surface feature fusion |
CN113763481A (en) * | 2021-08-16 | 2021-12-07 | 北京易航远智科技有限公司 | Multi-camera visual three-dimensional map construction and self-calibration method in mobile scene |
CN114119805A (en) * | 2021-10-28 | 2022-03-01 | 北京理工大学 | Semantic map building SLAM method for point-line-surface fusion |
CN114202035A (en) * | 2021-12-16 | 2022-03-18 | 成都理工大学 | Multi-feature fusion large-scale network community detection algorithm |
CN115578620A (en) * | 2022-10-28 | 2023-01-06 | 北京理工大学 | Point-line-surface multi-dimensional feature-visible light fusion slam method |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150103183A1 (en) * | 2013-10-10 | 2015-04-16 | Nvidia Corporation | Method and apparatus for device orientation tracking using a visual gyroscope |
CN104851094A (en) * | 2015-05-14 | 2015-08-19 | 西安电子科技大学 | Improved method of RGB-D-based SLAM algorithm |
US20180173991A1 (en) * | 2016-02-24 | 2018-06-21 | Soinn Holdings Llc | Feature value extraction method and feature value extraction apparatus |
CN109523589A (en) * | 2018-11-13 | 2019-03-26 | 浙江工业大学 | A kind of design method of more robust visual odometry |
CN109544636A (en) * | 2018-10-10 | 2019-03-29 | 广州大学 | A kind of quick monocular vision odometer navigation locating method of fusion feature point method and direct method |
-
2019
- 2019-04-30 CN CN201910357796.9A patent/CN110060277A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150103183A1 (en) * | 2013-10-10 | 2015-04-16 | Nvidia Corporation | Method and apparatus for device orientation tracking using a visual gyroscope |
CN104851094A (en) * | 2015-05-14 | 2015-08-19 | 西安电子科技大学 | Improved method of RGB-D-based SLAM algorithm |
US20180173991A1 (en) * | 2016-02-24 | 2018-06-21 | Soinn Holdings Llc | Feature value extraction method and feature value extraction apparatus |
CN109544636A (en) * | 2018-10-10 | 2019-03-29 | 广州大学 | A kind of quick monocular vision odometer navigation locating method of fusion feature point method and direct method |
CN109523589A (en) * | 2018-11-13 | 2019-03-26 | 浙江工业大学 | A kind of design method of more robust visual odometry |
Non-Patent Citations (5)
Title |
---|
DYLAN THOMAS等: "A monocular SLAM method for satellite proximity operations", 《2016 AMERICAN CONTROL CONFERENCE (ACC)》 * |
仇翔等: "基于RGB-D相机的视觉里程计实现", 《浙江工业大学学报》 * |
彭天博等: "增强室内视觉里程计实用性的方法", 《模式识别与人工智能》 * |
林志诚等: "移动机器人视觉SLAM过程中图像匹配及相机位姿求解的研究", 《机械设计与制造工程》 * |
梁强: "基于深度相机的室内定位系统研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 * |
Cited By (45)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110599569A (en) * | 2019-09-16 | 2019-12-20 | 上海市刑事科学技术研究院 | Method for generating two-dimensional plane graph inside building, storage device and terminal |
CN110631586A (en) * | 2019-09-26 | 2019-12-31 | 珠海市一微半导体有限公司 | Map construction method based on visual SLAM, navigation system and device |
CN110807799A (en) * | 2019-09-29 | 2020-02-18 | 哈尔滨工程大学 | Line feature visual odometer method combining depth map inference |
CN110807799B (en) * | 2019-09-29 | 2023-05-30 | 哈尔滨工程大学 | Line feature visual odometer method combined with depth map inference |
CN110926405A (en) * | 2019-12-04 | 2020-03-27 | 中科新松有限公司 | ARV attitude measurement method based on monocular vision vanishing point detection |
CN110926405B (en) * | 2019-12-04 | 2022-02-22 | 中科新松有限公司 | ARV attitude measurement method based on monocular vision vanishing point detection |
WO2021135645A1 (en) * | 2019-12-31 | 2021-07-08 | 歌尔股份有限公司 | Map updating method and device |
US12031837B2 (en) | 2019-12-31 | 2024-07-09 | Goertek Inc. | Method and device for updating map |
CN111275764B (en) * | 2020-02-12 | 2023-05-16 | 南开大学 | Depth camera visual mileage measurement method based on line segment shadows |
CN111275764A (en) * | 2020-02-12 | 2020-06-12 | 南开大学 | Depth camera visual mileage measurement method based on line segment shadow |
CN111310772B (en) * | 2020-03-16 | 2023-04-21 | 上海交通大学 | Point line characteristic selection method and system for binocular vision SLAM |
CN111310772A (en) * | 2020-03-16 | 2020-06-19 | 上海交通大学 | Point line feature selection method and system for binocular vision SLAM |
CN111461141A (en) * | 2020-03-30 | 2020-07-28 | 歌尔科技有限公司 | Equipment pose calculation method device and equipment |
CN111461141B (en) * | 2020-03-30 | 2023-08-29 | 歌尔科技有限公司 | Equipment pose calculating method and device |
CN111369594A (en) * | 2020-03-31 | 2020-07-03 | 北京旋极信息技术股份有限公司 | Method, device, computer storage medium and terminal for realizing target tracking |
CN111693047A (en) * | 2020-05-08 | 2020-09-22 | 中国航空工业集团公司西安航空计算技术研究所 | Visual navigation method for micro unmanned aerial vehicle in high-dynamic scene |
CN111693047B (en) * | 2020-05-08 | 2022-07-05 | 中国航空工业集团公司西安航空计算技术研究所 | Visual navigation method for micro unmanned aerial vehicle in high-dynamic scene |
CN111899334A (en) * | 2020-07-28 | 2020-11-06 | 北京科技大学 | Visual synchronous positioning and map building method and device based on point-line characteristics |
CN111899334B (en) * | 2020-07-28 | 2023-04-18 | 北京科技大学 | Visual synchronous positioning and map building method and device based on point-line characteristics |
CN112114966A (en) * | 2020-09-15 | 2020-12-22 | 杭州未名信科科技有限公司 | Light beam adjustment calculation method of visual SLAM |
CN112114966B (en) * | 2020-09-15 | 2024-08-13 | 杭州未名信科科技有限公司 | Light beam adjustment calculation method of vision SLAM |
CN112631271A (en) * | 2020-10-09 | 2021-04-09 | 南京凌华微电子科技有限公司 | Map generation method based on robot perception fusion |
CN112258391A (en) * | 2020-10-12 | 2021-01-22 | 武汉中海庭数据技术有限公司 | Fragmented map splicing method based on road traffic marking |
CN112305558A (en) * | 2020-10-22 | 2021-02-02 | 中国人民解放军战略支援部队信息工程大学 | Mobile robot track determination method and device by using laser point cloud data |
CN112305558B (en) * | 2020-10-22 | 2023-08-01 | 中国人民解放军战略支援部队信息工程大学 | Mobile robot track determination method and device using laser point cloud data |
CN112419497A (en) * | 2020-11-13 | 2021-02-26 | 天津大学 | Monocular vision-based SLAM method combining feature method and direct method |
CN112381890A (en) * | 2020-11-27 | 2021-02-19 | 上海工程技术大学 | RGB-D vision SLAM method based on dotted line characteristics |
CN112381890B (en) * | 2020-11-27 | 2022-08-02 | 上海工程技术大学 | RGB-D vision SLAM method based on dotted line characteristics |
CN112945233A (en) * | 2021-01-15 | 2021-06-11 | 北京理工大学 | Global drift-free autonomous robot simultaneous positioning and map building method |
CN112945233B (en) * | 2021-01-15 | 2024-02-20 | 北京理工大学 | Global drift-free autonomous robot simultaneous positioning and map construction method |
CN112880687A (en) * | 2021-01-21 | 2021-06-01 | 深圳市普渡科技有限公司 | Indoor positioning method, device, equipment and computer readable storage medium |
CN112880687B (en) * | 2021-01-21 | 2024-05-17 | 深圳市普渡科技有限公司 | Indoor positioning method, device, equipment and computer readable storage medium |
CN112802196B (en) * | 2021-02-01 | 2022-10-21 | 北京理工大学 | Binocular inertia simultaneous positioning and map construction method based on dotted line feature fusion |
CN112802196A (en) * | 2021-02-01 | 2021-05-14 | 北京理工大学 | Binocular inertia simultaneous positioning and map construction method based on dotted line feature fusion |
CN112950709B (en) * | 2021-02-21 | 2023-10-24 | 深圳市优必选科技股份有限公司 | Pose prediction method, pose prediction device and robot |
CN112950709A (en) * | 2021-02-21 | 2021-06-11 | 深圳市优必选科技股份有限公司 | Pose prediction method, pose prediction device and robot |
CN113608236A (en) * | 2021-08-03 | 2021-11-05 | 哈尔滨智兀科技有限公司 | Mine robot positioning and image building method based on laser radar and binocular camera |
CN113763481A (en) * | 2021-08-16 | 2021-12-07 | 北京易航远智科技有限公司 | Multi-camera visual three-dimensional map construction and self-calibration method in mobile scene |
CN113763481B (en) * | 2021-08-16 | 2024-04-05 | 北京易航远智科技有限公司 | Multi-camera visual three-dimensional map construction and self-calibration method in mobile scene |
CN113689499A (en) * | 2021-08-18 | 2021-11-23 | 哈尔滨工业大学 | Visual rapid positioning method, device and system based on point-surface feature fusion |
CN114119805B (en) * | 2021-10-28 | 2024-06-04 | 北京理工大学 | Semantic mapping SLAM method for point-line-plane fusion |
CN114119805A (en) * | 2021-10-28 | 2022-03-01 | 北京理工大学 | Semantic map building SLAM method for point-line-surface fusion |
CN114202035A (en) * | 2021-12-16 | 2022-03-18 | 成都理工大学 | Multi-feature fusion large-scale network community detection algorithm |
CN115578620B (en) * | 2022-10-28 | 2023-07-18 | 北京理工大学 | Point-line-plane multidimensional feature-visible light fusion slam method |
CN115578620A (en) * | 2022-10-28 | 2023-01-06 | 北京理工大学 | Point-line-surface multi-dimensional feature-visible light fusion slam method |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110060277A (en) | A kind of vision SLAM method of multiple features fusion | |
CN106055091B (en) | A kind of hand gestures estimation method based on depth information and correcting mode | |
CN109166149A (en) | A kind of positioning and three-dimensional wire-frame method for reconstructing and system of fusion binocular camera and IMU | |
CN111563442A (en) | Slam method and system for fusing point cloud and camera image data based on laser radar | |
KR101257207B1 (en) | Method, apparatus and computer-readable recording medium for head tracking | |
CN109460267A (en) | Mobile robot offline map saves and real-time method for relocating | |
CN107563323A (en) | A kind of video human face characteristic point positioning method | |
CN106296742A (en) | A kind of online method for tracking target of combination Feature Points Matching | |
CN112446882A (en) | Robust visual SLAM method based on deep learning in dynamic scene | |
CN105760898A (en) | Vision mapping method based on mixed group regression method | |
CN116449384A (en) | Radar inertial tight coupling positioning mapping method based on solid-state laser radar | |
CN114677323A (en) | Semantic vision SLAM positioning method based on target detection in indoor dynamic scene | |
CN111709997B (en) | SLAM implementation method and system based on point and plane characteristics | |
CN110070578A (en) | A kind of winding detection method | |
CN112465858A (en) | Semantic vision SLAM method based on probability grid filtering | |
CN111027586A (en) | Target tracking method based on novel response map fusion | |
Zhang | Multiple moving objects detection and tracking based on optical flow in polar-log images | |
CN115018999A (en) | Multi-robot-cooperation dense point cloud map construction method and device | |
Xu et al. | 3D joints estimation of the human body in single-frame point cloud | |
CN118071873A (en) | Dense Gaussian map reconstruction method and system in dynamic environment | |
CN112432653B (en) | Monocular vision inertial odometer method based on dotted line characteristics | |
CN118010024A (en) | Point-line-surface decoupling-based 6-DoF pose estimation optimization method and system | |
CN115855018A (en) | Improved synchronous positioning and mapping method based on point-line comprehensive characteristics | |
Lai et al. | A survey of deep learning application in dynamic visual SLAM | |
CN106558065A (en) | The real-time vision tracking to target is realized based on color of image and texture analysiss |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20190726 |
|
WD01 | Invention patent application deemed withdrawn after publication |