CN108986037B - Monocular vision odometer positioning method and positioning system based on semi-direct method - Google Patents
Monocular vision odometer positioning method and positioning system based on semi-direct method Download PDFInfo
- Publication number
- CN108986037B CN108986037B CN201810512342.XA CN201810512342A CN108986037B CN 108986037 B CN108986037 B CN 108986037B CN 201810512342 A CN201810512342 A CN 201810512342A CN 108986037 B CN108986037 B CN 108986037B
- Authority
- CN
- China
- Prior art keywords
- points
- point
- frame
- image
- camera
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 70
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 9
- 230000008569 process Effects 0.000 claims abstract description 8
- 230000008859 change Effects 0.000 claims abstract description 4
- 238000007781 pre-processing Methods 0.000 claims abstract description 4
- 239000011159 matrix material Substances 0.000 claims description 54
- 230000036544 posture Effects 0.000 claims description 21
- 230000000007 visual effect Effects 0.000 claims description 16
- 238000012937 correction Methods 0.000 claims description 8
- 238000004364 calculation method Methods 0.000 claims description 6
- 238000005070 sampling Methods 0.000 claims description 6
- 230000000694 effects Effects 0.000 claims description 5
- 238000001914 filtration Methods 0.000 claims description 5
- 238000005457 optimization Methods 0.000 claims description 5
- 238000010276 construction Methods 0.000 claims description 4
- 238000009795 derivation Methods 0.000 claims description 4
- 238000006073 displacement reaction Methods 0.000 claims description 4
- 230000033001 locomotion Effects 0.000 claims description 4
- 238000001514 detection method Methods 0.000 claims description 3
- 238000009825 accumulation Methods 0.000 claims description 2
- 125000004432 carbon atom Chemical group C* 0.000 claims description 2
- 238000000546 chi-square test Methods 0.000 claims description 2
- 230000007423 decrease Effects 0.000 claims description 2
- 230000003287 optical effect Effects 0.000 claims description 2
- 238000012545 processing Methods 0.000 claims description 2
- 238000012546 transfer Methods 0.000 claims description 2
- 238000005286 illumination Methods 0.000 abstract description 2
- 230000035945 sensitivity Effects 0.000 abstract description 2
- 238000010586 diagram Methods 0.000 description 3
- 238000005259 measurement Methods 0.000 description 3
- 239000000463 material Substances 0.000 description 2
- 238000005096 rolling process Methods 0.000 description 2
- 230000004075 alteration Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000011084 recovery Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/80—Geometric correction
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C11/00—Photogrammetry or videogrammetry, e.g. stereogrammetry; Photographic surveying
- G01C11/04—Interpretation of pictures
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C22/00—Measuring distance traversed on the ground by vehicles, persons, animals or other moving solid bodies, e.g. using odometers, using pedometers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/80—Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/46—Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30244—Camera pose
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
- Length Measuring Devices By Optical Means (AREA)
Abstract
The invention discloses a monocular vision odometer positioning method and a positioning system based on a semi-direct method, wherein the method comprises the following steps: entering a target scene by using a monocular camera and recording scene image data; preprocessing the obtained scene image, and detecting the characteristic points; initializing a system, tracking the obtained characteristic points, taking the characteristic point pair set obtained by tracking as a data set of a robustness algorithm, extracting line segment characteristics in a scene image, and performing attitude estimation of a camera by using a semi-direct method; selecting a key frame in a camera shooting image; and (5) constructing a map. The invention realizes the initialization process with robustness and high quality. According to the method, when each three-dimensional feature is added into the depth filter for the first time, the key frame is selected to update the depth information of the three-dimensional feature for one time, so that the uncertainty of the new three-dimensional feature depth information is reduced, and the convergence of the depth is accelerated; and the reference frame with the three-dimensional characteristics is updated, the time interval between the reference frame and the current frame is shortened, and the sensitivity to illumination change is reduced.
Description
Technical Field
The invention belongs to the technical field of visual mileage in three-dimensional reconstruction, and particularly relates to a monocular visual odometer positioning method and a monocular visual odometer positioning system based on a semi-direct method.
Background
With the continuous evolution of computer vision, image processing and other technologies in recent years, vision sensors are applied in more and more scenes, and the mileometers based on the vision sensors are also receiving more and more attention of researchers. Vision sensors are becoming one of the main sensors for odometry applications due to their low cost, rich information, etc. The visual odometer is more efficient and requires less hardware than SLAM-based (simultaneous localization and mapping). The traditional wheel type odometer mainly calculates the mileage through the rolling of the wheel, but the wheel may slip and idle on a rugged road surface, which brings a large error to the odometer measurement, and the odometer based on the visual sensor is not affected by the slip and idle.
The current mainstream visual odometer is classified by hardware, and can be roughly classified into a monocular system using only one camera and a binocular system using a stereo camera. Monocular means using only one camera as a unique sensor, while binocular means using one stereo camera to simultaneously capture two images at each moment. Compared with a binocular vision mileometer, the monocular vision mileometer has the advantages of low cost, no need of considering the influence of a camera baseline on the system precision and the like.
In 2014, Christian Forster proposed a visual odometer combining the features of the direct method and the feature point method, named SVO. The SVO firstly extracts FAST feature points from the image, but does not calculate descriptors of the feature points, but estimates the initial pose of a camera by using a direct method, so that a large amount of feature detection and matching time is saved, and meanwhile, the FAST feature points are only calculated, so that the speed of the FAST feature points is rapidly improved, and the method is very suitable for platforms with limited computing resources, such as unmanned aerial vehicles, smart phones and the like. However, SVO is not robust enough because it pursues extremely fast.
Disclosure of Invention
The invention aims to at least solve the technical problems in the prior art, and particularly innovatively provides a monocular vision odometer positioning method and a monocular vision odometer positioning system based on a semi-direct method.
In order to achieve the above object of the present invention, according to a first invention of the present invention, there is provided a monocular vision odometer positioning method based on a semi-direct method, comprising the steps of:
s1, entering a target scene by using an unmanned aerial vehicle or a vehicle-mounted monocular camera and recording scene image data;
s2, preprocessing the scene image obtained in the step S1, correcting distortion and detecting characteristic points of the corrected image;
s3, initializing the system, tracking the characteristic points obtained in the step S2, taking the tracked characteristic point pair set as a data set of a robustness algorithm, estimating a basic matrix F of two images by using an eight-point method, calculating a homography matrix H by using four point pairs, and selecting one of the basic matrix F and the homography matrix H for subsequent calculation by calculating the inner point pairs of the two matrixes;
s4, tracking and positioning the camera to form a motion track of the camera, extracting line segment characteristics in a scene image, and estimating the posture of the camera by using a semi-direct method, wherein the sparse images are aligned, and a photometric error is established through characteristic points on a current frame image and a previous frame image to obtain an initial camera posture; then, aligning the characteristics, projecting all map points in the map to the current frame by using the camera postures obtained in the last step, establishing photometric errors by using the gray difference of the projection points and the map points on the reference frame, and optimizing the positions of the projection points on the image; finally, a re-projection error about the three-dimensional point and the camera attitude is established by utilizing the optimized position and the optimized projection point, and the camera attitude and the three-dimensional point are optimized to obtain the final camera attitude of the current frame;
s5, selecting key frames in the camera shooting images;
s6, constructing a map, carrying out depth filtering on end points of line segment features in a key frame, newly constructing a seed point from the newly acquired key frame, carrying out one-time pre-updating on the seed point by using a plurality of key frames closest to the newly acquired key frame, adding the seed point into a seed point queue, if the seed point queue is converged, newly constructing a key frame, projecting all three-dimensional points in the map onto the new key frame, if the three-dimensional points are visible on the new key frame, changing a reference frame of the three-dimensional points into the new key frame, and if the seed point queue is not converged, continuously acquiring the key frame and updating the seed point queue.
In order to achieve the above object, according to a second aspect of the present invention, the present invention provides a monocular vision odometer positioning system based on a semi-direct method, which includes a monocular vision camera and a processor, wherein the monocular vision camera is carried by an unmanned aerial vehicle or a vehicle, the monocular vision camera acquires images and transmits the images to the processor, and the processor processes the images by using the method of the present invention, constructs a map, positions and displays a movement track of the camera on the map.
The invention adopts the visual mileage recorder to measure the mileage, thereby reducing the error of mileage measurement. The conventional wheel-type odometer mainly calculates the mileage through the rolling of the wheel, but the wheel may slip or idle on a rugged road surface, which brings a large error to the odometer measurement, and the odometer based on the visual sensor is not affected by the slip or idle.
Compared with a binocular vision mileometer and even a multi-ocular vision mileometer, the monocular vision mileometer has the advantages of low cost, no need of considering the influence of a camera baseline on the system precision and the like.
The invention realizes the initialization process with robustness and high quality. The new key frame selection strategy enhances and improves the key frame selection strategy based on scene depth, and eliminates the limitation that the key frame selection strategy can only be used for camera downward viewing.
According to the method, when each three-dimensional feature is added into the depth filter for the first time, the key frame is selected to update the depth information of the three-dimensional feature for one time, so that the uncertainty of the new three-dimensional feature depth information is reduced, and the convergence of the depth is accelerated; and the reference frame with the three-dimensional characteristics is updated, the time interval between the reference frame and the current frame is shortened, and the sensitivity to illumination change is reduced.
Additional aspects and advantages of the invention will be set forth in part in the description which follows and, in part, will be obvious from the description, or may be learned by practice of the invention.
Drawings
The above and/or additional aspects and advantages of the present invention will become apparent and readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings of which:
FIG. 1 is a flow chart of a monocular visual odometer positioning method based on the semi-direct method in a preferred embodiment of the present invention;
FIG. 2 is a diagram showing the effect of distortion correction in a preferred embodiment of the present invention, wherein FIG. 2(a) is an image before correction, and FIG. 2(b) is an image before correction;
FIG. 3 is a schematic diagram of an initialization process in a preferred embodiment of the present invention;
fig. 4 is a diagram illustrating robust estimation in a preferred embodiment of the present invention.
Detailed Description
Reference will now be made in detail to embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like or similar reference numerals refer to the same or similar elements or elements having the same or similar function throughout. The embodiments described below with reference to the accompanying drawings are illustrative only for the purpose of explaining the present invention, and are not to be construed as limiting the present invention.
In the description of the present invention, unless otherwise specified and limited, it is to be noted that the terms "mounted," "connected," and "connected" are to be interpreted broadly, and may be, for example, a mechanical connection or an electrical connection, a communication between two elements, a direct connection, or an indirect connection via an intermediate medium, and specific meanings of the terms may be understood by those skilled in the art according to specific situations.
The invention provides a monocular vision odometer positioning method based on a semi-direct method, which comprises the following steps as shown in figure 1:
s1, entering a target scene by using an unmanned aerial vehicle or a vehicle-mounted monocular camera and recording scene image data;
s2, preprocessing the scene image obtained in the step S1, correcting distortion and detecting characteristic points of the corrected image;
s3, initializing the system, tracking the characteristic points obtained in the step S2, using the tracked characteristic point pair set as a data set of a robustness algorithm, estimating a basic matrix F of two images by using an eight-point method and calculating a homography matrix H by using four point pairs, in another preferred embodiment of the invention, calculating by using a method such as nonlinear optimization, and selecting one of the basic matrix F and the homography matrix H by calculating inner point pairs of the two matrices to perform subsequent calculation;
s4, tracking the camera, extracting line segment characteristics in a scene image, and estimating the camera posture by using a semi-direct method, wherein the sparse images are aligned, and a photometric error is established through characteristic points on a current frame image and a previous frame image to obtain a rough camera posture; then, aligning the characteristics, projecting all map points in the map to the current frame by using the camera postures obtained in the last step, establishing photometric errors by using the gray difference of the projection points and the map points on the reference frame, and optimizing the positions of the projection points on the image; finally, a re-projection error about the three-dimensional point and the camera attitude is established by utilizing the optimized position and the optimized projection point, and the camera attitude and the three-dimensional point are optimized to obtain the final camera attitude of the current frame;
s5, selecting key frames in the camera shooting images;
s6, constructing a map, carrying out depth filtering on end points of line segment features in a key frame, newly constructing a seed point from the newly acquired key frame, carrying out one-time pre-updating on the seed point by using a plurality of key frames closest to the newly acquired key frame, adding the seed point into a seed point queue, if the seed point queue is converged, newly constructing a key frame, projecting all three-dimensional points in the map onto the new key frame, if the three-dimensional points are visible on the new key frame, changing a reference frame of the three-dimensional points into the new key frame, and if the seed point queue is not converged, continuously acquiring the key frame and updating the seed point queue.
In the embodiment, the imaging distortion of the camera is composed of two parts of radial distortion and tangential distortion, and the distortion coefficient is a 5-dimensional vector (k)1,k2,k3,p1,p2) To indicate. k is a radical of3Has a small influence, and in the present invention, it does not participate in the operation. The distortion coefficients of the camera can be solved during camera calibration,the distortion coefficients are related only to the hardware of the camera, as is the camera reference. In the distortion correction process, the distorted image is known and the undistorted image is required, and it is common practice to calculate the coordinates (u, v) of the point which should be in the distorted image after distortion for a given undistorted image and then calculate the coordinates (u, v) of the distorted image of the image at this pointd,vd) Thus, the pixel value of the undistorted image coordinate (u, v) is obtained, and the complete distortion corrected image is obtained after traversing the coordinates of all the distortion corrected images.
The distortion correction method comprises the following steps:
solving its physical coordinates (x ', y') on the image plane from the given pixel coordinates:
the physical coordinates after distortion are:
r2=x'2+y'2
converting the distorted physical coordinates to pixel coordinates on the image:
wherein (k)1,k2,k3,p1,p2) Is the distortion coefficient of the camera, (u, v) is the coordinates of the undistorted image, (ud,vd) Is the coordinates of the corresponding distorted image (u, v), (c)x,cy) Is the position of the camera optical center; (f)x,fy) Are the focal lengths of the camera in the x-direction and the y-direction.
In fig. 2, the left side shows an image before distortion correction, and the right side shows an effect image after distortion correction. The red line in the figure is a straight line in the real world, however, because of the distortion, the left figure has been seriously bent, and after the distortion is corrected, the position of the red line in the right figure is corrected to be a straight line.
In the present embodiment, the method for detecting the feature point includes:
establishing a circle by taking a pixel point in the image as the center of the circle and j as the radius, and judging whether the pixel point is a characteristic point or not according to the difference between the pixel point on the circle and the center of the circle, wherein j is a positive integer greater than 1; if the difference between the gray value of the central point p and the gray value of any continuous k pixels on the circle is larger than or smaller than the threshold, the point p is a feature point, and the value of k is j × j. In the present invention, j is preferably 3. According to the method, a large number of useless feature points are generated at the edge in the image, so that after the initial detection is completed, non-maximum suppression is performed on all detected feature points, and the best feature point in the area is selected.
In the present embodiment, as shown in fig. 3, the initialization specifically includes:
s31, determining the number of feature points in the picture, if the number of feature points exceeds a threshold (e.g. 100), determining whether there is a reference frame (i.e. the first image that satisfies the threshold exceeding the feature points), if there is a reference frame, executing step S32, if there is no reference frame, saving the picture as a reference frame, executing step S32, and when the number of feature points does not exceed the threshold, waiting for the next scene picture to be input;
s32, tracking the characteristic points, and acquiring gray level images I for two continuous time points t, t +1t,It+1Let I assumetA window w to be tracked, the positions of the pixels in the two images within the window w satisfy:
It(x,y)=It+1(x+δx,y+δy),
finding the matching position of the window to be tracked by minimizing the sum of squares of the gray differences of the window to be tracked between the images:
wherein e (w) is the matching position of the window to be tracked,
taylor expansion is carried out on the formula to a high-order linear term, minimum values can be obtained by respectively making partial derivatives zero, and the movement (delta x, delta y) of the window to be tracked between the images is obtained;
s33, judging the tracking effect, if the requirement of the threshold value is satisfied (the number of the feature points successfully tracked is more than 50), executing the step S34, otherwise, judging whether the interval between the frame and the reference frame exceeds 50 frames, if so, deleting the reference frame, inputting the picture again, and if not, inputting the picture again;
s34, as shown in fig. 4, randomly selecting the minimum number of data (4 pairs) required for calculating the target model from the matched feature point pairs in the two images, generating a model by using the selected minimum data, where the model is a basis matrix F estimated by using an eight-point method for the two images or a homography matrix H calculated by using four point pairs, adapting the data in the data set by using the model, and if the data is in accordance with the current model, marking the data as the inner point of the current iteration, otherwise, marking the data as the outer point, evaluating the quality of the current model by the proportion of the inner points, recording the model with the highest proportion of the inner points (the proportion of the inner points is the highest), if the iteration end condition is satisfied, ending, otherwise, returning to the beginning of the step, where the iteration end condition includes two parts: the first part is to reach the maximum number of iterations; the second part is to set an iteration number k, which is updated with each iteration, and the update formula is as follows:
wherein p is a confidence coefficient, w is a proportion of inner points under the current optimal model, m is the minimum data quantity reduced by the fitting model, and iteration is stopped when the iteration number is greater than k or any one of the maximum iteration numbers;
estimating a basic matrix F of the two images by using an eight-point method and calculating a homography matrix H by using four point pairs, wherein when the basic matrix is estimated, the minimum subset is 8 randomly selected point pairs, and when the homography matrix is estimated, the minimum subset is four randomly selected point pairs;
for the simultaneously estimated basic matrix FAnd evaluating the homography matrix H, and for the estimated model M, the score under the characteristic point pair set P can be expressed as sM:
Wherein (p)i,pi') indicates a pair of feature points for which tracking was successful, d2(p, p', M) represents the symmetric transfer error of two points under model M, ρM(d2(p, p', M)) represents a pair d2(p, p', M) performing Chi-square test to determine whether the pair of points is an interior point pair, the number of the last interior point pair is the final score of the model, and calculating the score s of the basic matrix in this wayFAnd the score s of the homography matrixH:
If R >0.4, then the homography matrix H is selected, otherwise the basis matrix F is selected.
In this embodiment, the step S4 specifically includes:
s41, extracting line segment characteristics of the scene image by using an LSD algorithm (a specific algorithm reference paper LSD: A line segment detector), uniformly sampling the extracted line segment characteristics, and converting the extracted line segment characteristics into point characteristics;
s42, extracting feature points of a newly arrived image frame to be used as feature points, aligning sparse images, establishing luminosity errors through the feature points on a current frame image and a previous frame image, and solving a rough camera posture; aligning features, projecting all map points in a map onto a current frame by using the camera postures obtained in the previous step, establishing luminosity errors by using the gray level difference of the projection points and the map points on a reference frame, and optimizing the positions of the projection points on the image; finally, a re-projection error about the three-dimensional point and the camera attitude is established by utilizing the optimized position and the optimized projection point, and the camera attitude and the three-dimensional point are optimized to obtain the final camera attitude of the current frame;
wherein sparse image alignment refers to:
assume that the current time is k and the current image frame is IkThe previous time is k-1 and the previous image frame is Ik-1At this time, the relative attitude of two adjacent frames is solved, and a point piAs an image Ik-1The point of any one of the above characteristic points, back-projected to the three-dimensional space under the camera coordinate system thereof, is piA 1 is to piProjection to IkThe projection point obtained above is pi', then the photometric error is:
e=Ik-1(pi)-Ik(pi')
for the line segment characteristics, carrying out uniform sampling on the line segment, and back-projecting the line segment characteristics on the previous frame image to a three-dimensional space and then projecting the line segment characteristics to the current frame;
and the luminosity error of the line segment characteristic is the sum of the luminosity errors of the sampling points, if the camera internal reference matrix is K and the camera posture of the current frame relative to the previous frame is ξ:
pi'=Kexp(ξ^)Pi,
solving for an increment of camera pose per iteration such that the value of the entire cost function decreases until the condition to stop the iteration is satisfied, and pre-multiplying ξ by a small perturbation δ ξ:
expanding the above equation into taylor to the first order and simplifying it can be:
in the above formula, the first and second carbon atoms are,is a point piThe gradient at the point of' is,for the derivative of the projection equation with respect to the camera pose, the chain rule:
wherein, Pi'=exp(ξ^)PiIf the three-dimensional point is a three-dimensional point in the camera coordinate system and the coordinates thereof are (X ', Y ', Z '), then:
thus:
then, in the sparse image alignment stage, the jacobian matrix using the photometric error resume cost function is:
by means of continuous iteration, the total error is reduced, the relative postures of the cameras at the previous moment and the current moment can be obtained finally, and the camera postures of the current cameras under the world coordinate system are obtained through accumulation;
the feature alignment includes: projecting all three-dimensional points visible in the current image in the map onto the image by using the camera postures obtained by sparse image alignment to obtain a series of projection points, and for the three-dimensional points P in the mapi(X, Y, Z), the key frame in which it was first observed is called its reference key frame, and the image of its reference key frame is recorded as IrefIts position in the reference frame is piThe three-dimensional point being visible on the image at time k and whichProjection point uiAnd adopting the photometric error:
e=Iref(pi)-Ik(ui),
the parameter to be optimized for feature alignment is the position u of the projection pointi,
The line segment is represented by two end points in a three-dimensional space, the end points of the three-dimensional line segment are respectively projected into the current image in the characteristic alignment stage, and the alignment of the line segment characteristics can be completed by aligning the positions of the end points,
and finally, in the attitude and three-dimensional structure optimization stage, establishing a reprojection error by using the distance between the optimized position of the projection point and the non-optimized position in the feature alignment stage, optimizing the camera attitude obtained in the sparse alignment stage and the three-dimensional point position in the map, and assuming any three-dimensional point p in the mapiProjected point u ofiAssume its optimized position as ui', then:
e=u′i-ui,
for the line segment characteristics, the reprojection error is composed of two end points, and is represented by a perturbation model of lie algebra:
wherein, Pi' representing a three-dimensional Point PiBy using three-dimensional points under a camera coordinate system, the above formula actually represents derivation of a projection equation to the camera attitude, when the three-dimensional points are optimized, how e changes along with the change of the three-dimensional points needs to be known, and the e is analyzed about the three-dimensional point PiDerivative of' by the chain derivative rule:
assuming the current camera pose is R, t, then from the projection equation:
Pi'=RPi+t,
Pi' Pair PiThe derivation is in fact a rotation matrix R,the Jacobian matrix of the reprojection error relative to the three-dimensional points is:
combining the Jacobian matrix of the reprojection error with respect to pose and the Jacobian matrix with respect to the three-dimensional points is an iterative Jacobian matrix in the pose and structure optimization process.
In this embodiment, the step S5 specifically includes:
one of the following three conditions is satisfied to begin selecting a key frame:
① when the map construction thread is idle, the method is mainly divided into two aspects, namely that the key frame queue to be processed is empty, and meanwhile, in the map construction thread, the number of three-dimensional points to be updated is small, and at the moment, the addition of key frames is considered to improve the density of the map;
② the visibility of the feature point on the key frame nearest to the current frame on the current frame is low, the key frame far from the current frame and the current frame should be overlapped in a larger scene, when this happens, it indicates that the camera is entering a new area, and the key frame needs to be added immediately to ensure the correct tracking of the camera;
③ the displacement in relative pose of the current image and the previous image exceeds a certain proportion of the average depth of the scene;
an image frame also needs to satisfy the following condition to be selected as a key frame:
① is within a certain distance from the nearest key frame, (0.5-2, this is set manually, is dependent on the data used and therefore has no fixed standard)
② in the feature alignment stage, the number of three-dimensional points of the feature points which can be re-projected on the current image is not lower than a certain value (40%) compared with the last frame;
the key frame selection algorithm is as follows:
inputting a current image frame and a previous image frame;
if the keyframe queue is empty, and if the non-converged three-dimensional point <100, then the pseudo code needKfs is true, start selecting the keyframe, projecting all three-dimensional points on the keyframe onto the current image frame, if all three-dimensional points on the current frame visible three-dimensional points/keyframe <0.5, then needKfs is true,
if the displacement/scene mean depth of the current frame is >0.1, needks ═ true,
if needks is true, the distance of the current frame from the key frame is calculated,
if the distance between the current frame and the key frame is less than the maximum distance threshold and the distance between the current frame and the key frame is greater than the minimum distance threshold, then all points in the map are projected onto the current image frame, the visibility number is m, all points in the map are projected onto the previous image frame, the visibility number is n,
if m/n > is 0.4, the current frame is a key frame.
In this embodiment, the step S6 specifically includes:
performing depth filtering on the end points of the line segment characteristics, initializing the characteristic points on the key frame into seed points and putting the seed points into a seed queue when the depth filter obtains a key frame, and updating all the seed points of the seed queue by using the image frame when the depth filter obtains a common image frame, namely updating the depth value and the uncertainty value of the depth of the seed point, wherein the seed points are uncertain in position on a new image but on a straight line (polar line);
searching epipolar lines of the feature points u of the seed points in the reference frame on the current image to obtain corresponding points u' (epipolar line matching), and taking pixels in the feature point field as templates to perform template matching during searching;
calculating depth using triangulation after obtaining the matching points; updating the original depth estimation and uncertainty by using a Bayesian probability model, adding the updated depth information into a map if the updated depth information is converged, deleting the seed point if the updated depth information is dispersed, and waiting for the next update if the updated depth information is dispersed;
and (2) deep pre-updating, namely newly building a seed point from a new key frame, pre-updating the seed point once by using a plurality of key frames closest to the new key frame, and then adding the seed point into a seed point queue, wherein the specific deep pre-updating step comprises the following steps:
acquiring a new key frame;
acquiring the average depth of a new key frame;
find the n key frames closest to the new key frame,
for feature points ft in the new keyframe:
newly building a seed point by using the characteristic point ft, and initializing the depth to be the average depth;
for a key-frame kf of the n key-frames:
carrying out epipolar line search with a new key frame;
the triangulation calculates the depth and uncertainty of the characteristic point ft, and updates the seed points based on a Bayesian probability model; the specific calculation method can adopt the method described in the paper, "REMODE: Probasilictic, monoclonal dense recovery time";
adding seed points to a seed point queue
Outputting a new seed point queue;
every time a key frame is newly built, all three-dimensional points in the map are projected onto the new key frame, if the three-dimensional points are visible on the new key frame, the reference frame of the three-dimensional points is changed into the new key frame, all the three-dimensional points in the image are projected into the latest key frame, the gray level of the projection points in the latest image and the gray level of the corresponding three-dimensional points in the reference frame are used for establishing luminosity errors, the positions of the projection points between the latest image frames are optimized, and the map is constructed.
It should be noted that the calculation method of the basis matrix F, the homography matrix H and the rotation matrix R in the present invention adopts the existing calculation method, and the acquisition method of the camera internal reference matrix K adopts the acquisition method commonly used in the art.
The invention also provides a monocular vision mileage system based on the semi-direct method, which comprises a monocular vision camera and a processor, wherein the monocular vision camera is carried by an unmanned aerial vehicle or a vehicle, the monocular vision camera acquires images and transmits the images to the processor, and the processor processes the images by using the method and constructs a map.
In the description herein, references to the description of the term "one embodiment," "some embodiments," "an example," "a specific example," or "some examples," etc., mean that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the invention. In this specification, the schematic representations of the terms used above do not necessarily refer to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples.
While embodiments of the invention have been shown and described, it will be understood by those of ordinary skill in the art that: various changes, modifications, substitutions and alterations can be made to the embodiments without departing from the principles and spirit of the invention, the scope of which is defined by the claims and their equivalents.
Claims (6)
1. A monocular vision odometer positioning method based on a semi-direct method is characterized by comprising the following steps:
s1, entering a target scene by using an unmanned aerial vehicle or a vehicle-mounted monocular camera and recording scene image data;
s2, preprocessing the scene image obtained in the step S1, correcting distortion and detecting characteristic points of the corrected image;
s3, initializing the system, tracking the characteristic points obtained in the step S2, taking the tracked characteristic point pair set as a data set of a robustness algorithm, estimating a basic matrix F of two images by using an eight-point method, calculating a homography matrix H by using four point pairs, and selecting one of the basic matrix F and the homography matrix H for subsequent calculation by calculating the inner point pairs of the two matrixes;
s4, tracking and positioning the camera, extracting line segment characteristics in a scene image, and estimating the camera posture by using a semi-direct method, wherein the sparse images are aligned, and a photometric error is established through characteristic points on a current frame image and a previous frame image to obtain a rough camera posture; then, aligning the characteristics, projecting all map points in the map to the current frame by using the camera postures obtained in the last step, establishing photometric errors by using the gray difference of the projection points and the map points on the reference frame, and optimizing the positions of the projection points on the image; finally, a re-projection error about the three-dimensional point and the camera attitude is established by utilizing the optimized position and the optimized projection point, and the camera attitude and the three-dimensional point are optimized to obtain the final camera attitude of the current frame;
s5, selecting a key frame from the camera image, specifically including:
one of the following three conditions is satisfied to begin selecting a key frame:
① when the map construction thread is idle, the method is mainly divided into two aspects, namely that the key frame queue to be processed is empty, and meanwhile, in the map construction thread, the number of three-dimensional points to be updated is small, and at the moment, the addition of key frames is considered to improve the density of the map;
② the visibility of the feature point on the key frame nearest to the current frame on the current frame is low, the key frame far from the current frame and the current frame should be overlapped in a larger scene, when this happens, it indicates that the camera is entering a new area, and the key frame needs to be added immediately to ensure the correct tracking of the camera;
③ the displacement in relative pose of the current image and the previous image exceeds a certain proportion of the average depth of the scene;
an image frame also needs to satisfy the following condition to be selected as a key frame:
① are within a certain distance from the nearest key frame;
② in the feature alignment stage, the number of three-dimensional points of the feature points which can be re-projected on the current image is not lower than a certain value compared with the last frame;
the key frame selection algorithm is as follows:
inputting a current image frame and a previous image frame;
if the keyframe queue is empty, and if the three-dimensional point that does not converge <100, then the pseudocode needKfs is true, start selecting the keyframe, find the closest keyframe to the current frame, project all three-dimensional points on the keyframe onto the current image frame, if the three-dimensional points visible for the current frame/all three-dimensional points on the keyframe <0.5, then needKfs is true,
if the displacement/scene mean depth of the current frame is >0.1, needks ═ true,
if needks is true, the distance of the current frame from the key frame is calculated,
if the distance between the current frame and the key frame is less than the maximum distance threshold and the distance between the current frame and the key frame is greater than the minimum distance threshold, then all points in the map are projected onto the current image frame, the visibility number is m, all points in the map are projected onto the previous image frame, the visibility number is n,
if m/n > is 0.4, the current frame is a key frame;
s6, constructing a map, carrying out depth filtering on end points of line segment features in a key frame, newly constructing a seed point from the newly acquired key frame, carrying out one-time pre-updating on the seed point by using a plurality of key frames closest to the newly acquired key frame, adding the seed point into a seed point queue, if the seed point queue is converged, newly constructing a key frame, projecting all three-dimensional points in the map onto the new key frame, if the three-dimensional points are visible on the new key frame, changing a reference frame of the three-dimensional points into the new key frame, and if the seed point queue is not converged, continuously acquiring the key frame and updating the seed point queue;
the method specifically comprises the following steps:
performing depth filtering on the end points of the line segment characteristics, initializing the characteristic points on the key frame into seed points and putting the seed points into a seed queue when the depth filter obtains a key frame, updating all the seed points of the seed queue by using the image frame when the depth filter obtains a common image frame, wherein the positions of the seed points on a new image are uncertain and are on the same straight line;
searching according to the epipolar line of the feature point u of the seed point in the reference frame on the current image to obtain a corresponding point u', and taking the pixel in the feature point field as a template to perform template matching during searching;
calculating depth using triangulation after obtaining the matching points; updating the original depth estimation and uncertainty by using a Bayesian probability model, adding the updated depth information into a map if the updated depth information is converged, deleting the seed point if the updated depth information is dispersed, and waiting for the next update if the updated depth information is dispersed;
and (2) deep pre-updating, namely newly building a seed point from a new key frame, pre-updating the seed point once by using a plurality of key frames closest to the new key frame, and then adding the seed point into a seed point queue, wherein the specific deep pre-updating step comprises the following steps:
acquiring a new key frame;
acquiring the average depth of a new key frame;
find the n key frames closest to the new key frame,
for feature points ft in the new keyframe:
newly building a seed point by using the characteristic point ft, and initializing the depth to be the average depth;
for a key-frame kf of the n key-frames:
carrying out epipolar line search with a new key frame;
triangulation calculates the depth and uncertainty of the feature point ft;
updating the seed points based on a Bayesian probability model;
the seed points are added to a seed point queue,
outputting a new seed point queue;
every time a key frame is newly built, all three-dimensional points in the map are projected onto the new key frame, if the three-dimensional points are visible on the new key frame, the reference frame of the three-dimensional points is changed into the new key frame, all the three-dimensional points in the image are projected into the latest key frame, the gray level of the projection points in the latest image and the gray level of the corresponding three-dimensional points in the reference frame are used for establishing luminosity errors, the positions of the projection points between the latest image frames are optimized, and the map is constructed.
2. The monocular visual odometer positioning method based on the semi-direct method according to claim 1, wherein the distortion correction method is:
solving its physical coordinates (x ', y') on the image plane from the given pixel coordinates:
the physical coordinates after distortion are:
r2=x'2+y'2
converting the distorted physical coordinates to pixel coordinates on the image:
wherein (k)1,k2,k3,p1,p2) Is the distortion coefficient of the camera, (u, v) is the coordinates of the undistorted image, (ud,vd) Is the coordinates of the corresponding distorted image (u, v), (c)x,cy) Is the position of the camera optical center; (f)x,fy) Are the focal lengths of the camera in the x-direction and the y-direction.
3. The monocular visual odometer positioning method based on the semi-direct method according to claim 1, wherein the method for feature point detection is:
establishing a circle by taking a pixel point in the image as the center of the circle and j as the radius, and judging whether the pixel point is a characteristic point or not according to the difference between the pixel point on the circle and the center of the circle, wherein j is a positive integer greater than 1; if the difference between the gray value of the central point p and the gray value of any continuous k pixels on the circle is larger than or smaller than the threshold, the point p is a feature point, and the value of k is j × j.
4. The monocular visual odometer positioning method based on the semi-direct method according to claim 1, wherein the step S3 specifically comprises:
s31, judging the number of the feature points in the picture, if the number of the feature points exceeds the threshold value, judging whether a reference frame exists, if so, executing a step S32, if not, saving the picture as the reference frame, executing a step S32, and when the number of the feature points does not exceed the threshold value, waiting for inputting the next scene picture;
s32, tracking the characteristic points, and acquiring gray level images I for two continuous time points t, t +1t,It+1Let I assumetA window w to be tracked, the positions of the pixels in the two images within the window w satisfy:
It(x,y)=It+1(x+δx,y+δy),
finding the matching position of the window to be tracked by minimizing the sum of squares of the gray differences of the window to be tracked between the images:
wherein e (w) is the matching position of the window to be tracked,
taylor expansion is carried out on the formula to a high-order linear term, minimum values can be obtained by respectively making partial derivatives zero, and the movement (delta x, delta y) of the window to be tracked between the images is obtained;
s33, judging the tracking effect, if the tracking effect meets the requirement, executing the step S34, otherwise judging whether the interval between the frame and the reference frame exceeds 50 frames, if so, deleting the reference frame, inputting the picture again, and if not, inputting the picture again;
s34, randomly selecting the minimum data number needed by calculating a target model from the matched feature point pairs in the two images, generating a model by using the selected minimum data, wherein the model is a basic matrix F estimated by using an eight-point method for the two images or a homography matrix H calculated by using four point pairs, adapting the data in a data set by using the model, marking the data which conform to the current model as an inner point of the iteration, otherwise marking the data as an outer point, evaluating the quality of the current model by using the proportion of the inner points, recording the model with the highest proportion of the inner points, finishing if the iteration finishing condition is met, otherwise returning to the beginning of the step, and the iteration finishing condition comprises two parts: the first part is to reach the maximum number of iterations; the second part is to set an iteration number k, which is updated with each iteration, and the update formula is as follows:
wherein p is a confidence coefficient, w is a proportion of inner points under the current optimal model, m is the minimum data quantity reduced by the fitting model, and iteration is stopped when the iteration number is greater than k or any one of the maximum iteration numbers;
estimating a basic matrix F of the two images by using an eight-point method and calculating a homography matrix H by using four point pairs, wherein when the basic matrix is estimated, the minimum subset is 8 randomly selected point pairs, and when the homography matrix is estimated, the minimum subset is four randomly selected point pairs;
evaluating the simultaneously estimated basis matrix F and homography matrix H, and expressing the score of the estimated model M under the characteristic point pair set P as sM:
Wherein (p)i,p′i) A pair of feature points representing successful tracking, d2(p, p', M) represents the symmetric transfer error of two points under model M, ρM(d2(p, p', M)) represents a pair d2(p, p', M) performing Chi-square test to determine whether the pair of points is an interior point pair, the number of the last interior point pair is the final score of the model, and calculating the score s of the basic matrix in this wayFAnd the score s of the homography matrixH:
If R >0.4, then the homography matrix H is selected, otherwise the basis matrix F is selected.
5. The monocular visual odometer positioning method based on the semi-direct method according to claim 1, wherein the step S4 specifically comprises:
s41, extracting line segment characteristics of the scene image by using an LSD algorithm, uniformly sampling the extracted line segment characteristics, and converting the line segment characteristics into point characteristics;
s42, extracting feature points of a newly arrived image frame to be used as feature points, aligning sparse images, establishing luminosity errors through the feature points on a current frame image and a previous frame image, and solving a rough camera posture; aligning features, projecting all map points in a map onto a current frame by using the camera postures obtained in the previous step, establishing luminosity errors by using the gray level difference of the projection points and the map points on a reference frame, and optimizing the positions of the projection points on the image; finally, a re-projection error about the three-dimensional point and the camera attitude is established by utilizing the optimized position and the optimized projection point, and the camera attitude and the three-dimensional point are optimized to obtain the final camera attitude of the current frame;
wherein sparse image alignment refers to:
assume that the current time is k and the current image frame is IkThe previous time is k-1 and the previous image frame is Ik-1At this time, the relative attitude of two adjacent frames is solved, and a point piAs an image Ik-1The point of any one of the above characteristic points, back-projected to the three-dimensional space under the camera coordinate system thereof, is piA 1 is to piProjection to IkThe projection point obtained above is pi', then the photometric error is:
e=Ik-1(pi)-Ik(pi')
for the line segment characteristics, carrying out uniform sampling on the line segment, and back-projecting the line segment characteristics on the previous frame image to a three-dimensional space and then projecting the line segment characteristics to the current frame;
and the luminosity error of the line segment characteristic is the sum of the luminosity errors of the sampling points, if the camera internal reference matrix is K and the camera posture of the current frame relative to the previous frame is ξ:
p′i=Kexp(ξ^)Pi,
solving for an increment of camera pose per iteration such that the value of the entire cost function decreases until the condition to stop the iteration is satisfied, and pre-multiplying ξ by a small perturbation δ ξ:
expanding the above equation into taylor to the first order and simplifying it can be:
in the above formula, the first and second carbon atoms are,is a point piThe gradient at the point of' is,for the derivative of the projection equation with respect to the camera pose, the chain rule:
wherein, Pi'=exp(ξ^)PiIf the three-dimensional point is a three-dimensional point in the camera coordinate system and the coordinates thereof are (X ', Y ', Z '), then:
thus:
then, in the sparse image alignment stage, the jacobian matrix using the photometric error resume cost function is:
by means of continuous iteration, the total error is reduced, the relative postures of the cameras at the previous moment and the current moment can be obtained finally, and the camera postures of the current cameras under the world coordinate system are obtained through accumulation;
the feature alignment includes: projecting all three-dimensional points visible in the current image in the map onto the image by using the camera postures obtained by sparse image alignment to obtain a series of projection points, and for the three-dimensional points P in the mapi(X, Y, Z), the key frame in which it was first observed is called its reference key frame, and the image of its reference key frame is recorded as IrefIts position in the reference frame is piThe three-dimensional point is visible on the image at time k and its projection point is uiAnd adopting the photometric error:
e=Iref(pi)-Ik(ui),
the parameter to be optimized for feature alignment is the position u of the projection pointi,
The line segment is represented by two end points in a three-dimensional space, the end points of the three-dimensional line segment are respectively projected into the current image in the characteristic alignment stage, and the alignment of the line segment characteristics can be completed by aligning the positions of the end points,
and finally, in the attitude and three-dimensional structure optimization stage, establishing a reprojection error by using the distance between the optimized position of the projection point and the non-optimized position in the feature alignment stage, and solving the reprojection error in the sparse alignment stageIs optimized with the three-dimensional point position in the map, assuming any three-dimensional point p in the mapiProjected point u ofiAssume its optimized position to be u'iAnd then:
e=u′i-ui,
for the line segment characteristics, the reprojection error is composed of two end points, and is represented by a perturbation model of lie algebra:
wherein, Pi' representing a three-dimensional Point PiBy using three-dimensional points under a camera coordinate system, the above formula actually represents derivation of a projection equation to the camera attitude, when the three-dimensional points are optimized, how e changes along with the change of the three-dimensional points needs to be known, and the e is analyzed about the three-dimensional point PiDerivative of' by the chain derivative rule:
assuming the current camera pose is R, t, then from the projection equation:
Pi'=RPi+t,
Pi' Pair PiThe derivation is actually a rotation matrix R, and the jacobian matrix of the reprojection error with respect to the three-dimensional points is:
combining the Jacobian matrix of the reprojection error with respect to pose and the Jacobian matrix with respect to the three-dimensional points is an iterative Jacobian matrix in the pose and structure optimization process.
6. A monocular visual odometer positioning system based on a semi-direct method, comprising a monocular visual camera carried by an unmanned plane or a vehicle, and a processor for acquiring images and transmitting the images to the processor, wherein the processor performs processing by using the method of any one of claims 1 to 5 and constructs a map.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810512342.XA CN108986037B (en) | 2018-05-25 | 2018-05-25 | Monocular vision odometer positioning method and positioning system based on semi-direct method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810512342.XA CN108986037B (en) | 2018-05-25 | 2018-05-25 | Monocular vision odometer positioning method and positioning system based on semi-direct method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN108986037A CN108986037A (en) | 2018-12-11 |
CN108986037B true CN108986037B (en) | 2020-06-16 |
Family
ID=64542055
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810512342.XA Active CN108986037B (en) | 2018-05-25 | 2018-05-25 | Monocular vision odometer positioning method and positioning system based on semi-direct method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108986037B (en) |
Families Citing this family (41)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111322993B (en) * | 2018-12-13 | 2022-03-04 | 杭州海康机器人技术有限公司 | Visual positioning method and device |
CN109443320A (en) * | 2019-01-10 | 2019-03-08 | 轻客小觅智能科技(北京)有限公司 | Binocular vision speedometer and measurement method based on direct method and line feature |
CN109816726B (en) * | 2019-01-29 | 2021-10-01 | 京东方科技集团股份有限公司 | Visual odometer map updating method and system based on depth filter |
CN109752008B (en) * | 2019-03-05 | 2021-04-13 | 长安大学 | Intelligent vehicle multi-mode cooperative positioning system and method and intelligent vehicle |
CN109727269B (en) * | 2019-03-29 | 2019-07-09 | 中国人民解放军国防科技大学 | Monocular vision and road map based matching positioning method |
CN110189390B (en) * | 2019-04-09 | 2023-02-14 | 南京航空航天大学 | Monocular vision SLAM method and system |
CN111862146B (en) * | 2019-04-30 | 2023-08-29 | 北京魔门塔科技有限公司 | Target object positioning method and device |
CN110335316B (en) * | 2019-06-28 | 2023-04-18 | Oppo广东移动通信有限公司 | Depth information-based pose determination method, device, medium and electronic equipment |
CN110363821B (en) * | 2019-07-12 | 2021-09-28 | 顺丰科技有限公司 | Monocular camera installation deviation angle acquisition method and device, camera and storage medium |
CN110473258B (en) * | 2019-07-24 | 2022-05-13 | 西北工业大学 | Monocular SLAM system initialization algorithm based on point-line unified framework |
CN110514212A (en) * | 2019-07-26 | 2019-11-29 | 电子科技大学 | A kind of intelligent vehicle map terrestrial reference localization method merging monocular vision and difference GNSS |
CN110807809B (en) * | 2019-10-25 | 2021-04-09 | 中山大学 | Light-weight monocular vision positioning method based on point-line characteristics and depth filter |
CN112967311B (en) * | 2019-12-12 | 2024-06-07 | 浙江商汤科技开发有限公司 | Three-dimensional line graph construction method and device, electronic equipment and storage medium |
CN111145255B (en) * | 2019-12-27 | 2022-08-09 | 浙江省北大信息技术高等研究院 | Pose calculation method and system combining deep learning and geometric optimization |
CN113409368B (en) * | 2020-03-16 | 2023-11-03 | 北京京东乾石科技有限公司 | Mapping method and device, computer readable storage medium and electronic equipment |
CN111462179B (en) * | 2020-03-26 | 2023-06-27 | 北京百度网讯科技有限公司 | Three-dimensional object tracking method and device and electronic equipment |
CN111553300B (en) * | 2020-05-08 | 2022-03-11 | 北京工商大学 | Multi-time-domain resolution lip language behavior detection method for three-dimensional point cloud video |
CN111583331B (en) * | 2020-05-12 | 2023-09-01 | 北京轩宇空间科技有限公司 | Method and device for simultaneous localization and mapping |
CN111721318B (en) * | 2020-05-26 | 2022-03-25 | 南京航空航天大学 | Template matching visual odometer based on self-adaptive search area |
CN111829522B (en) * | 2020-07-02 | 2022-07-12 | 浙江大华技术股份有限公司 | Instant positioning and map construction method, computer equipment and device |
CN111950599B (en) * | 2020-07-20 | 2022-07-01 | 重庆邮电大学 | Dense visual odometer method for fusing edge information in dynamic environment |
CN112115980A (en) * | 2020-08-25 | 2020-12-22 | 西北工业大学 | Binocular vision odometer design method based on optical flow tracking and point line feature matching |
CN112580683B (en) * | 2020-11-17 | 2024-01-12 | 中山大学 | Multi-sensor data time alignment system and method based on cross correlation |
CN112665575B (en) * | 2020-11-27 | 2023-12-29 | 重庆大学 | SLAM loop detection method based on mobile robot |
CN112633122B (en) * | 2020-12-17 | 2024-01-23 | 厦门大学 | Front-end mileage calculation method and system of monocular VIO system |
CN112767538B (en) * | 2021-01-11 | 2024-06-07 | 浙江商汤科技开发有限公司 | Three-dimensional reconstruction and related interaction and measurement methods, related devices and equipment |
KR20220112575A (en) * | 2021-02-04 | 2022-08-11 | 삼성전자주식회사 | Method for performing simultaneous localization and mapping and device using same |
CN112862803B (en) * | 2021-02-26 | 2023-09-26 | 中国人民解放军93114部队 | Infrared imaging SLAM method and device based on edge and feature point fusion |
CN113108771B (en) * | 2021-03-05 | 2022-08-16 | 华南理工大学 | Movement pose estimation method based on closed-loop direct sparse visual odometer |
CN113112542A (en) * | 2021-03-25 | 2021-07-13 | 北京达佳互联信息技术有限公司 | Visual positioning method and device, electronic equipment and storage medium |
CN113206949B (en) * | 2021-04-01 | 2023-04-28 | 广州大学 | Semi-direct monocular vision SLAM method based on entropy weighted image gradient |
CN113362377B (en) * | 2021-06-29 | 2022-06-03 | 东南大学 | VO weighted optimization method based on monocular camera |
CN113674340A (en) * | 2021-07-05 | 2021-11-19 | 北京物资学院 | Binocular vision navigation method and device based on landmark points |
WO2023280274A1 (en) * | 2021-07-07 | 2023-01-12 | The Hong Kong University Of Science And Technology | Geometric structure aided visual localization method and system |
CN113592947B (en) * | 2021-07-30 | 2024-03-12 | 北京理工大学 | Method for realizing visual odometer by semi-direct method |
CN113689485B (en) * | 2021-08-25 | 2022-06-07 | 北京三快在线科技有限公司 | Method and device for determining depth information of unmanned aerial vehicle, unmanned aerial vehicle and storage medium |
CN113781567B (en) * | 2021-10-08 | 2024-05-31 | 西北工业大学 | Aerial image target geographic positioning method based on three-dimensional map generation |
CN113870353A (en) * | 2021-10-11 | 2021-12-31 | 重庆邮电大学 | Monocular vision-based indoor positioning method for unmanned aerial vehicle |
CN114708392B (en) * | 2022-03-22 | 2024-05-14 | 重庆大学 | Octree map construction method based on closed-loop track |
CN116625380B (en) * | 2023-07-26 | 2023-09-29 | 广东工业大学 | Path planning method and system based on machine learning and SLAM |
CN116681733B (en) * | 2023-08-03 | 2023-11-07 | 南京航空航天大学 | Near-distance real-time pose tracking method for space non-cooperative target |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103680291A (en) * | 2012-09-09 | 2014-03-26 | 复旦大学 | Method for realizing simultaneous locating and mapping based on ceiling vision |
CN105809687A (en) * | 2016-03-08 | 2016-07-27 | 清华大学 | Monocular vision ranging method based on edge point information in image |
CN105869136A (en) * | 2015-01-22 | 2016-08-17 | 北京雷动云合智能技术有限公司 | Collaborative visual SLAM method based on multiple cameras |
CN107341814A (en) * | 2017-06-14 | 2017-11-10 | 宁波大学 | The four rotor wing unmanned aerial vehicle monocular vision ranging methods based on sparse direct method |
CN107610175A (en) * | 2017-08-04 | 2018-01-19 | 华南理工大学 | The monocular vision SLAM algorithms optimized based on semi-direct method and sliding window |
CN107687850A (en) * | 2017-07-26 | 2018-02-13 | 哈尔滨工业大学深圳研究生院 | A kind of unmanned vehicle position and orientation estimation method of view-based access control model and Inertial Measurement Unit |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140212001A1 (en) * | 2013-01-27 | 2014-07-31 | Quantum Signal Llc | Visual odometry |
-
2018
- 2018-05-25 CN CN201810512342.XA patent/CN108986037B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103680291A (en) * | 2012-09-09 | 2014-03-26 | 复旦大学 | Method for realizing simultaneous locating and mapping based on ceiling vision |
CN105869136A (en) * | 2015-01-22 | 2016-08-17 | 北京雷动云合智能技术有限公司 | Collaborative visual SLAM method based on multiple cameras |
CN105809687A (en) * | 2016-03-08 | 2016-07-27 | 清华大学 | Monocular vision ranging method based on edge point information in image |
CN107341814A (en) * | 2017-06-14 | 2017-11-10 | 宁波大学 | The four rotor wing unmanned aerial vehicle monocular vision ranging methods based on sparse direct method |
CN107687850A (en) * | 2017-07-26 | 2018-02-13 | 哈尔滨工业大学深圳研究生院 | A kind of unmanned vehicle position and orientation estimation method of view-based access control model and Inertial Measurement Unit |
CN107610175A (en) * | 2017-08-04 | 2018-01-19 | 华南理工大学 | The monocular vision SLAM algorithms optimized based on semi-direct method and sliding window |
Non-Patent Citations (3)
Title |
---|
"SVO:Semidirect visual odometry for monocular and multicamera systems";Christian Forester等;《IEEE Transactions on Robotics》;20161231;第33卷(第2期);第249-265页 * |
"基于视觉的同时定位与地图构建的研究进展";陈常等;《计算机应用研究》;20170818(第3期);第641-647页 * |
"点线特征融合的单目视觉里程计";袁梦等;《激光与光电子学进展》;20170912(第2期);第1-8页 * |
Also Published As
Publication number | Publication date |
---|---|
CN108986037A (en) | 2018-12-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108986037B (en) | Monocular vision odometer positioning method and positioning system based on semi-direct method | |
CN110807809B (en) | Light-weight monocular vision positioning method based on point-line characteristics and depth filter | |
CN110223348B (en) | Robot scene self-adaptive pose estimation method based on RGB-D camera | |
CN104732518B (en) | A kind of PTAM improved methods based on intelligent robot terrain surface specifications | |
CN108682027A (en) | VSLAM realization method and systems based on point, line Fusion Features | |
JP5832341B2 (en) | Movie processing apparatus, movie processing method, and movie processing program | |
CN108010081B (en) | RGB-D visual odometer method based on Census transformation and local graph optimization | |
CN108615246B (en) | Method for improving robustness of visual odometer system and reducing calculation consumption of algorithm | |
CN108648194B (en) | Three-dimensional target identification segmentation and pose measurement method and device based on CAD model | |
CN108519102B (en) | Binocular vision mileage calculation method based on secondary projection | |
CN113108771A (en) | Movement pose estimation method based on closed-loop direct sparse visual odometer | |
CN110827321B (en) | Multi-camera collaborative active target tracking method based on three-dimensional information | |
JP7173471B2 (en) | 3D position estimation device and program | |
CN116449384A (en) | Radar inertial tight coupling positioning mapping method based on solid-state laser radar | |
CN111829522B (en) | Instant positioning and map construction method, computer equipment and device | |
CN116222543B (en) | Multi-sensor fusion map construction method and system for robot environment perception | |
CN112652020B (en) | Visual SLAM method based on AdaLAM algorithm | |
CN116468786B (en) | Semantic SLAM method based on point-line combination and oriented to dynamic environment | |
CN111882602A (en) | Visual odometer implementation method based on ORB feature points and GMS matching filter | |
CN111105467A (en) | Image calibration method and device and electronic equipment | |
CN115147344A (en) | Three-dimensional detection and tracking method for parts in augmented reality assisted automobile maintenance | |
KR101766823B1 (en) | Robust visual odometry system and method to irregular illumination changes | |
CN112767481B (en) | High-precision positioning and mapping method based on visual edge features | |
CN104156933A (en) | Image registering method based on optical flow field | |
CN111260725B (en) | Dynamic environment-oriented wheel speed meter-assisted visual odometer method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |