CN108053445A - The RGB-D camera motion methods of estimation of Fusion Features - Google Patents

The RGB-D camera motion methods of estimation of Fusion Features Download PDF

Info

Publication number
CN108053445A
CN108053445A CN201711297500.6A CN201711297500A CN108053445A CN 108053445 A CN108053445 A CN 108053445A CN 201711297500 A CN201711297500 A CN 201711297500A CN 108053445 A CN108053445 A CN 108053445A
Authority
CN
China
Prior art keywords
mrow
mtr
mtd
msub
dimensional
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201711297500.6A
Other languages
Chinese (zh)
Inventor
陈白帆
刘春发
宋德臻
王斌
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Central South University
Original Assignee
Central South University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Central South University filed Critical Central South University
Priority to CN201711297500.6A priority Critical patent/CN108053445A/en
Publication of CN108053445A publication Critical patent/CN108053445A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a kind of RGB D camera motion methods of estimation of Fusion Features.The present invention extracts two-dimensional points and two-dimentional linear feature first in RGB image, and two dimensional character back projection is obtained three-dimensional feature by the depth information in D images.Then, by RGB measurement errors and the error of depth survey error structure three-dimensional point, the uncertainty of straight line is measured by calculating the tripleplane of two-dimentional line-sampling point with the mahalanobis distance of estimated 3 d-line.Finally, the three-dimensional point and 3 d-line characteristic matching pair of adjacent two frame are merged, using unascertained information, the movement of RGB D cameras is calculated by Maximum-likelihood estimation.The present invention has merged the linear feature insensitive to illumination variation, and the error model of reasonable construction system improves robustness and the accuracy of camera motion estimation.

Description

Feature-fused RGB-D camera motion estimation method
Technical Field
The invention belongs to the technical field of machine vision, and particularly relates to a feature fusion RGB-D camera motion estimation method.
Background
In recent years, with the rapid development of image processing technology and the emergence of various vision sensors, vision-based mobile robots have received increasing attention. Compare laser radar, millimeter wave radar etc. and vision sensor can acquire abundanter environmental information, can also reduce cost simultaneously. Visual Odometer (VO) is a device that estimates the course of motion of a camera or a body connected to it (e.g., an automobile, a human or a mobile robot, etc.) by means of a Visual sensor only. The method is a sub-problem of visual simultaneous localization and map creation (VSLAM), and is a core problem of realizing autonomous navigation of a mobile robot. Commonly used visual sensors include monocular cameras, binocular cameras, panoramic cameras and RGB-D cameras. RGB-D cameras, such as Microsoft Kinect, Primer Sense, Hua Shuo Xtion PRO Live, Australian, etc., have attracted considerable interest not only because of their light weight and low cost, but more importantly because of the simultaneous color and depth information that such cameras provide. Depth information may solve scale problems in monocular cameras. Physical measurement of an RGB-D camera saves a large amount of depth information computation process compared to a binocular camera.
At present, the mainstream camera motion estimation methods are roughly classified into feature point methods (such as SIFT, SURF, ORB, etc.) and direct methods. The method is sensitive to illumination change, the characteristic point method is easily affected by uneven distribution of characteristics, the robustness of the algorithm is poor, and the application requirement cannot be met.
Disclosure of Invention
The invention aims to solve the problems that the existing camera motion estimation method is easily influenced by illumination and large noise is caused by uneven feature distribution. A robust RGB-D camera motion estimation method based on point and line feature fusion is provided.
In order to achieve the technical purpose, the invention adopts the technical scheme that the RGB-D camera motion estimation method with feature fusion comprises the following steps:
step one, extracting two-dimensional point and straight line features: on an RGB picture, extracting two-dimensional feature points and two-dimensional feature straight lines by using a feature point detection algorithm and a straight line segmentation detection algorithm respectively;
step two, performing feature back projection to the three dimensions and performing uncertainty analysis: combining depth information, utilizing a pinhole camera model to perform back projection on the extracted two-dimensional feature points and two-dimensional feature straight lines to three dimensions, and performing uncertainty analysis on the three-dimensional points and the three-dimensional straight lines under the assumption of Gaussian noise distribution;
step three, feature matching: for the point characteristics, calculating a characteristic point detection algorithm descriptor and matching the point characteristics of two continuous frames; calculating and matching the mean standard deviation descriptors of the linear features; then removing mismatching by using a random sampling consistency algorithm;
step four, estimating the motion of the camera: carrying out maximum likelihood estimation on the camera motion by using the uncertainty information; and solving the objective function of the problem through a Levenberg algorithm to obtain the pose of the camera.
The RGB-D camera motion estimation method with feature fusion comprises the following steps:
for one RGB image, two-dimensional feature points are obtained through a feature point detection algorithm, meanwhile, two-dimensional feature straight lines are obtained through a straight line segmentation detection algorithm, and a feature set { p is obtainedi,lj1,2, …, j 1,2, … }, where the two-dimensional point p isi=[ui,vi]TTwo-dimensional straight line[ui,vi]TRepresents a point piPixel coordinates of ajAnd bjIs a straight line ljTwo end points of (a).
The second step of the feature-fused RGB-D camera motion estimation method comprises the following steps:
step 1, three-dimensional point characteristic and uncertainty analysis:
and (3) performing back projection on the two-dimensional point P in the image to obtain a three-dimensional point P through a pinhole camera model:
where (u, v) represents the pixel coordinate corresponding to the two-dimensional point p, d is the depth value corresponding to the two-dimensional point p, [ cu,cv]TIs the center of the aperture of the camera, fcIs the focal length;
the noise at two-dimensional point p is the mean 0 and the covarianceWith the noise of the depth value d of the two-dimensional point p being a quadratic function of the measured value, i.e. deltad=c1d2+c2d+c3Constant coefficient c1、c2And c3The uncertainty of the three-dimensional point is obtained through experiments as follows:
wherein the covariance matrix of the noiseJacobi matrix
Wherein, I2Is a 2-dimensional unit matrix, δ is the noise variance;
step 2, three-dimensional straight line characteristic and uncertainty analysis:
sampling enough points of the two-dimensional straight line, discarding points with abnormal depth values, and calculating three-dimensional coordinates of the remaining points according to a formula (1); then adopting a random sampling consistency algorithm to resist and remove the local outer points which appear after the two-dimensional straight line is back projected to the three-dimensional due to the influence of depth noise, and obtaining a fitted three-dimensional straight line equation; with L ═ AT,BT]TTo represent a corresponding three-dimensional straight line, wherein AT,BTAre two three-dimensional points on a three-dimensional straight line;
measuring the uncertainty of the three-dimensional straight line by calculating the Mahalanobis distance between the three-dimensional point corresponding to the two-dimensional sampling points and the estimated three-dimensional straight line L;
three-dimensional point P ═ x, y, z]TTo a three-dimensional line L ═ AT,BT]TThe mahalanobis distance of (a) is defined as follows:
wherein Q epsilon L is an arbitrary point on the straight line L; with Q ═ a + λ (B-a), the optimal estimate of Q is then Q*,
Definition of
δ(P-L)=P-Q*(4)
The mahalanobis distance from the three-dimensional point P to the three-dimensional line L is
The three-dimensional straight line is processed by a group of three-dimensional points { P after being processed by a random sampling consistency algorithmi,i=1,…nLIs formed of, wherein P1Andrespectively representing two end points of a straight line, and making the error function of a three-dimensional straight line be
Maximum likelihood estimation L of three-dimensional straight line L*Equivalent to minimizing the following:
whereinPAnd delta (P-L) are respectively calculated by a formula (2) and a formula (4); solving the nonlinear minimization problem of equation (7) using the Levenberg algorithm with an optimized uncertainty of the three-dimensional line of
Wherein,
the feature-fused RGB-D camera motion estimation method comprises the following steps:
for the point features, SURF descriptors of the point features are calculated, and the point features of two adjacent frames of images are matched through the similarity between the descriptors; for straight line features, a mean normalized descriptor is calculated and then matched.
The feature-fused RGB-D camera motion estimation method comprises the following four steps:
t represents the motion conversion of two adjacent frames, and comprises T (X), RX + T, wherein R is a rotation matrix, T is a displacement vector, and a six-dimensional vector ξ is formed by taking the non-0 items of the rotation matrix and the displacement vector;
to be provided withAs a set of matched three-dimensional point sets of two adjacent frames F and F',then the three-dimensional straight line feature matching pairs of two adjacent frames are obtained; if three-dimensional point matching pairsThe corresponding real physical point is X under the local coordinate of the previous frameiThree-dimensional points, matching pairs in three-dimensional linesCorresponding to a true physical straight line of YjThen, the variable to be optimized in the camera odometer system is
The error function defining the system includes errors for point features and errors for straight line features, i.e.
Wherein the error of the point feature is
Error of straight line feature is
Error of a single straight line is η (L)i-Yi)=[δ(Ai,Yi)T,δ(Bi,Yi)T]And is and
setting the target to find the variable Δ that minimizes the system error of the camera odometer system, i.e.
Wherein, sigmaf=diag(ΣPL) Solving the above formula (10) by using a Levenberg algorithm to obtain an initial pose of the camera motion; and constructing a pose graph by using initial pose information of camera motion, and optimizing the motion track of the camera by using the pose graph at the back end of the SLAM through a graph optimization method.
In the RGB-D camera motion estimation method using feature fusion, in the step 2, the point with an abnormal depth value is discarded as a point with a null depth value.
The invention has the technical effects that:
(1) the linear characteristic is insensitive to illumination change, and the method combines the point characteristic and the line characteristic, thereby greatly improving the robustness of motion estimation. Under the indoor environment, the linear features are particularly rich, the linear features in the environment are very easy to extract, and the complexity of the algorithm is reduced.
(2) The invention carries out uncertainty analysis on the three-dimensional points and the three-dimensional straight lines, carries out maximum likelihood estimation on the camera motion by using uncertain information, obtains the camera pose through optimization, and provides an effective and reasonable initial value for back-end optimization.
Drawings
FIG. 1 is a block diagram of a feature fused RGB-D camera motion estimation system.
FIG. 2 is an exemplary diagram of a two-dimensional straight-line backprojection into three-dimensional space.
Fig. 3 is a pose diagram of the camera.
Detailed Description
Embodiments of the present invention will be further described with reference to the accompanying drawings.
The invention provides an RGB-D camera motion estimation method based on point and line feature fusion, a whole system block diagram is shown as an attached figure 1, and the method comprises the following steps:
s1, two-dimensional feature extraction: the system respectively extracts two-dimensional point features and two-dimensional straight line features by using the input RGB image.
In the invention, a group of two-dimensional point features are obtained by SURF (speeded up robust feature) feature point detection algorithm, and two-dimensional line features are obtained by LSD (line Segment detector) line segmentation detection algorithm. For an RGB map, its feature set { p }i,lj1,2, …, j 1,2, … }, where the two-dimensional point p isi=[ui,vi]TTwo-dimensional straight line[ui,vi]TRepresents a point piPixel coordinates of ajAnd bjIs a straight line ljTwo end points of (a).
S2, three-dimensional feature acquisition and uncertainty analysis: and back projecting the two-dimensional features to three dimensions, respectively carrying out uncertain analysis on the three-dimensional points and the three-dimensional straight lines, and optimizing the estimation of the three-dimensional features.
(1) Three-dimensional point feature and uncertainty analysis
Two-dimensional point features are back-projected to three dimensions through a pinhole camera model. Assuming that a two-dimensional point P in an image corresponds to a depth value d, its corresponding three-dimensional point P is as follows:
here [ c ]u,cv]TIs the center of the aperture of the camera, fcIs the focal length.
By formula (1), the two-dimensional point feature can be back projected to a three-dimensional space to obtain a three-dimensional point feature (as shown in fig. 2). The RGB-D camera has larger measurement errors, and the noise of the acquired three-dimensional point cloud mainly comes from the RGB measurement errors and the depth value measurement errors.
Suppose that the noise at two-dimensional point p is mean 0 and covarianceA gaussian distribution of (a). Wherein, I2Is a 2-dimensional identity matrix and δ is the noise variance. The noise of the depth value d of the three-dimensional point P is quadratic in relation to the measured value, i.e. deltad=c1d2+c2d+c3Constant coefficient c1、c2And c3Obtained through experiments. Assuming that the measurement error of the two-dimensional point p and the error of the depth value d are independent of each other, the uncertainty of the three-dimensional point is:
wherein the covariance matrix of the noiseJacobi matrix
(2) Three-dimensional straight line feature and uncertainty analysis
The projection of a straight line in three-dimensional space to two dimensions still retains the properties of a straight line. Therefore, the present invention first detects straight lines in a two-dimensional RGB map and then back-projects to a three-dimensional space. The idea of back projection is to sample a two-dimensional straight line with enough points, calculate the three-dimensional coordinates of these points according to equation (1), and discard those points with invalid depth values, as shown in fig. 2. Due to the influence of depth noise, local points (such as small circles in fig. 2) appear after a two-dimensional straight line is back-projected to three dimensions, and a random sampling consensus algorithm (RANSAC) is adopted to remove the local points and obtain a fitted three-dimensional straight line equation. We use L ═ AT,BT]TTo represent a corresponding three-dimensional straight line, wherein AT,BTAre two three-dimensional points on a three-dimensional straight line.
And measuring the uncertainty of the straight line by calculating the mahalanobis distance between the three-dimensional point corresponding to the two-dimensional sampling points and the estimated three-dimensional straight line L.
Three-dimensional point P ═ x, y, z]TTo a three-dimensional line L ═ AT,BT]TThe mahalanobis distance of (a) is defined as follows:
wherein Q ∈ L is an arbitrary point on the straight line L. Assuming Q is a + λ (B-a), then the minimization problem of equation (3) is equivalent to the univariate quadratic function minimization problem. The optimal estimation of Q can be obtained through derivation and calculationQ*I.e. by
Definition of
δ(P-L)=P-Q*(4)
Thus, the mahalanobis distance of the three-dimensional point P to the three-dimensional straight line L can be written as
The three-dimensional straight line is processed by a group of three-dimensional points { P after being processed by a random sampling consistency algorithmi,i=1,…nLIs formed of, wherein P1Andrespectively representing the two end points of the straight line. Let the error function of the three-dimensional straight line be
Maximum likelihood estimation L of three-dimensional straight line L*Equivalent to minimizing the following:
whereinPAnd delta (P-L) are respectively calculated by a formula (2) and a formula (4); the non-linear minimization problem of equation (7) is solved using the Levenberg algorithm. Uncertainty of the optimized three-dimensional straight line is
Wherein,
s3, feature matching of two adjacent frames: including point feature matching and straight line feature matching.
For point features, the invention calculates SURF descriptors thereof, and matches the point features of two adjacent frames of images through the similarity between the descriptors. For straight line features, MSLD descriptors are computed and then matched.
S4, motion estimation: the invention estimates the motion pose of the camera by using the point characteristics and the straight line characteristics.
The motion transformation of two adjacent frames is represented by T (X) ═ RX + T, where R is the rotation matrix and T is the displacement vector, and the non-0 terms of the rotation matrix and the displacement vector are taken to form a six-dimensional vector ξ.
Suppose thatAs a set of matched three-dimensional point sets of two adjacent frames F and F',then it is a matching pair of three-dimensional straight line features of two adjacent frames. If three-dimensional point matching pairsThe corresponding real physical point is X under the local coordinate of the previous frameiThree-dimensional points, matching pairs in three-dimensional linesCorresponding to a true physical straight line of YjThen, the variable to be optimized in the camera odometer system is
The error function defining the system includes errors for point features and errors for straight line features, i.e.
Wherein the error of the point feature is
Error of straight line feature is
Error of a single straight line is η (L)i-Yi)=[δ(Ai,Yi)T,δ(Bi,Yi)T]And is and
setting the target to find the variable Δ that minimizes the system error of the camera odometer system, i.e.
Wherein, sigmaf=diag(ΣPL). The invention solves the above formula (10) by using LM algorithm to obtain the pose of the camera motion. The pose diagram of the camera is shown in fig. 3. The estimated information of the initial pose of the camera can construct a pose graph, and the track of the camera is optimized through a graph optimization method.

Claims (6)

1. A feature-fused RGB-D camera motion estimation method is characterized by comprising the following steps:
step one, extracting two-dimensional point and straight line features: on an RGB picture, extracting two-dimensional feature points and two-dimensional feature straight lines by using a feature point detection algorithm and a straight line segmentation detection algorithm respectively;
step two, performing feature back projection to the three dimensions and performing uncertainty analysis: combining depth information, utilizing a pinhole camera model to perform back projection on the extracted two-dimensional feature points and two-dimensional feature straight lines to three dimensions, and performing uncertainty analysis on the three-dimensional points and the three-dimensional straight lines under the assumption of Gaussian noise distribution;
step three, feature matching: for the point characteristics, calculating a characteristic point detection algorithm descriptor and matching the point characteristics of two continuous frames; calculating and matching the mean standard deviation descriptors of the linear features; then removing mismatching by using a random sampling consistency algorithm;
step four, estimating the motion of the camera: carrying out maximum likelihood estimation on the camera motion by using the uncertainty information; and solving the objective function of the problem through a Levenberg algorithm to obtain the pose of the camera.
2. The method as claimed in claim 1, wherein the step one comprises the steps of:
for one RGB image, two-dimensional feature points are obtained through a feature point detection algorithm, meanwhile, two-dimensional feature straight lines are obtained through a straight line segmentation detection algorithm, and a feature set { p is obtainedi,lj1,2, …, j 1,2, … }, where the two-dimensional point p isi=[ui,vi]TTwo-dimensional straight line[ui,vi]TRepresents a point piPixel coordinates of ajAnd bjIs a straight line ljTwo end points of (a).
3. The feature-fused RGB-D camera motion estimation method as claimed in claim 1, wherein the second step comprises the steps of:
step 1, three-dimensional point characteristic and uncertainty analysis:
and (3) performing back projection on the two-dimensional point P in the image to obtain a three-dimensional point P through a pinhole camera model:
<mrow> <mi>P</mi> <mo>=</mo> <mfenced open = "[" close = "]"> <mtable> <mtr> <mtd> <mi>x</mi> </mtd> </mtr> <mtr> <mtd> <mi>y</mi> </mtd> </mtr> <mtr> <mtd> <mi>z</mi> </mtd> </mtr> </mtable> </mfenced> <mo>=</mo> <mfenced open = "[" close = "]"> <mtable> <mtr> <mtd> <mrow> <mo>(</mo> <mi>u</mi> <mo>-</mo> <msub> <mi>c</mi> <mi>u</mi> </msub> <mo>)</mo> <mi>d</mi> <mo>/</mo> <msub> <mi>f</mi> <mi>c</mi> </msub> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mo>(</mo> <mi>v</mi> <mo>-</mo> <msub> <mi>c</mi> <mi>v</mi> </msub> <mo>)</mo> <mi>d</mi> <mo>/</mo> <msub> <mi>f</mi> <mi>c</mi> </msub> </mrow> </mtd> </mtr> <mtr> <mtd> <mi>d</mi> </mtd> </mtr> </mtable> </mfenced> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>)</mo> </mrow> </mrow>
where (u, v) represents the pixel coordinate corresponding to the two-dimensional point p, d is the depth value corresponding to the two-dimensional point p, [ cu,cv]TIs the center of the aperture of the camera, fcIs the focal length;
the noise at two-dimensional point p is the mean 0 and the covarianceWith the noise of the depth value d of the two-dimensional point p being a quadratic function of the measured value, i.e. deltad=c1d2+c2d+c3Constant coefficient c1、c2And c3The uncertainty of the three-dimensional point is obtained through experiments as follows:
<mrow> <msub> <mo>&amp;Sigma;</mo> <mi>P</mi> </msub> <mo>=</mo> <msub> <mi>J</mi> <mi>P</mi> </msub> <mi>c</mi> <mi>o</mi> <mi>v</mi> <mrow> <mo>(</mo> <mfenced open = "[" close = "]"> <mtable> <mtr> <mtd> <mi>p</mi> </mtd> </mtr> <mtr> <mtd> <mi>d</mi> </mtd> </mtr> </mtable> </mfenced> <mo>)</mo> </mrow> <msubsup> <mi>J</mi> <mi>P</mi> <mi>T</mi> </msubsup> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>2</mn> <mo>)</mo> </mrow> </mrow>
wherein the covariance matrix of the noiseJacobi matrix
Wherein, I2Is a 2-dimensional unit matrix, δ is the noise variance;
step 2, three-dimensional straight line characteristic and uncertainty analysis:
sampling enough points of the two-dimensional straight line, discarding points with abnormal depth values, and calculating three-dimensional coordinates of the remaining points according to a formula (1); then adopting a random sampling consistency algorithm to resist and remove the local outer points which appear after the two-dimensional straight line is back projected to the three-dimensional due to the influence of depth noise, and obtaining a fitted three-dimensional straight line equation; with L ═ AT,BT]TTo represent a corresponding three-dimensional straight line, wherein AT,BTAre two three-dimensional points on a three-dimensional straight line;
measuring the uncertainty of the three-dimensional straight line by calculating the Mahalanobis distance between the three-dimensional point corresponding to the two-dimensional sampling points and the estimated three-dimensional straight line L;
three-dimensional point P ═ x, y, z]TTo a three-dimensional line L ═ AT,BT]TThe mahalanobis distance of (a) is defined as follows:
<mrow> <msub> <mi>d</mi> <mrow> <mi>M</mi> <mi>A</mi> <mi>H</mi> </mrow> </msub> <mrow> <mo>(</mo> <mi>P</mi> <mo>,</mo> <mi>L</mi> <mo>)</mo> </mrow> <mo>=</mo> <munder> <mrow> <mi>m</mi> <mi>i</mi> <mi>n</mi> </mrow> <mrow> <mi>Q</mi> <mo>&amp;Element;</mo> <mi>L</mi> </mrow> </munder> <msqrt> <mrow> <msup> <mrow> <mo>(</mo> <mi>P</mi> <mo>-</mo> <mi>Q</mi> <mo>)</mo> </mrow> <mi>T</mi> </msup> <msubsup> <mo>&amp;Sigma;</mo> <mi>P</mi> <mrow> <mo>-</mo> <mn>1</mn> </mrow> </msubsup> <mrow> <mo>(</mo> <mi>P</mi> <mo>-</mo> <mi>Q</mi> <mo>)</mo> </mrow> </mrow> </msqrt> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>3</mn> <mo>)</mo> </mrow> </mrow>
wherein Q epsilon L is an arbitrary point on the straight line L; with Q ═ a + λ (B-a), the optimal estimate of Q is then Q*,
Definition of
δ(P-L)=P-Q*(4)
The mahalanobis distance from the three-dimensional point P to the three-dimensional line L is
<mrow> <msub> <mi>d</mi> <mrow> <mi>M</mi> <mi>A</mi> <mi>H</mi> </mrow> </msub> <mrow> <mo>(</mo> <mi>P</mi> <mo>,</mo> <mi>L</mi> <mo>)</mo> </mrow> <mo>=</mo> <msqrt> <mrow> <mi>&amp;delta;</mi> <msup> <mrow> <mo>(</mo> <mi>P</mi> <mo>-</mo> <mi>L</mi> <mo>)</mo> </mrow> <mi>T</mi> </msup> <msubsup> <mo>&amp;Sigma;</mo> <mi>P</mi> <mrow> <mo>-</mo> <mn>1</mn> </mrow> </msubsup> <mi>&amp;delta;</mi> <mrow> <mo>(</mo> <mi>P</mi> <mo>-</mo> <mi>L</mi> <mo>)</mo> </mrow> </mrow> </msqrt> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>5</mn> <mo>)</mo> </mrow> </mrow>
The three-dimensional straight line is processed by a group of three-dimensional points { P after being processed by a random sampling consistency algorithmi,i=1,…nLIs formed of, wherein P1Andrespectively representing two end points of a straight line, and making the error function of a three-dimensional straight line be
<mrow> <mi>w</mi> <mrow> <mo>(</mo> <mi>L</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfenced open = "[" close = "]"> <mtable> <mtr> <mtd> <mrow> <msub> <mi>P</mi> <mn>1</mn> </msub> <mo>-</mo> <mi>A</mi> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mi>&amp;delta;</mi> <mrow> <mo>(</mo> <msub> <mi>P</mi> <mn>2</mn> </msub> <mo>,</mo> <mi>L</mi> <mo>)</mo> </mrow> </mrow> </mtd> </mtr> <mtr> <mtd> <mo>.</mo> </mtd> </mtr> <mtr> <mtd> <mo>.</mo> </mtd> </mtr> <mtr> <mtd> <mo>.</mo> </mtd> </mtr> <mtr> <mtd> <mrow> <mi>&amp;delta;</mi> <mrow> <mo>(</mo> <msub> <mi>P</mi> <mrow> <msub> <mi>n</mi> <mi>L</mi> </msub> <mo>-</mo> <mn>1</mn> </mrow> </msub> <mo>,</mo> <mi>L</mi> <mo>)</mo> </mrow> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <msub> <mi>P</mi> <msub> <mi>n</mi> <mi>L</mi> </msub> </msub> <mo>-</mo> <mi>A</mi> </mrow> </mtd> </mtr> </mtable> </mfenced> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>6</mn> <mo>)</mo> </mrow> </mrow>
Maximum likelihood estimation L of three-dimensional straight line L*Equivalent to minimizing the following:
<mrow> <msup> <mi>L</mi> <mo>*</mo> </msup> <mo>=</mo> <munder> <mi>min</mi> <mi>L</mi> </munder> <mi>w</mi> <msup> <mrow> <mo>(</mo> <mi>L</mi> <mo>)</mo> </mrow> <mi>T</mi> </msup> <msubsup> <mo>&amp;Sigma;</mo> <mi>w</mi> <mrow> <mo>-</mo> <mn>1</mn> </mrow> </msubsup> <mi>w</mi> <mrow> <mo>(</mo> <mi>L</mi> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>7</mn> <mo>)</mo> </mrow> </mrow>
whereinPAnd delta (P-L) are respectively calculated by a formula (2) and a formula (4); solving the nonlinear minimization problem of equation (7) using the Levenberg algorithm with an optimized uncertainty of the three-dimensional line of
<mrow> <msub> <mo>&amp;Sigma;</mo> <mi>L</mi> </msub> <mo>=</mo> <msup> <mrow> <mo>(</mo> <msubsup> <mi>J</mi> <mi>w</mi> <mi>T</mi> </msubsup> <msubsup> <mo>&amp;Sigma;</mo> <mi>w</mi> <mrow> <mo>-</mo> <mn>1</mn> </mrow> </msubsup> <msub> <mi>J</mi> <mi>w</mi> </msub> <mo>)</mo> </mrow> <mrow> <mo>-</mo> <mn>1</mn> </mrow> </msup> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>8</mn> <mo>)</mo> </mrow> </mrow>
Wherein,
4. the feature-fused RGB-D camera motion estimation method according to claim 1, wherein said step three comprises the steps of:
for the point features, SURF descriptors of the point features are calculated, and the point features of two adjacent frames of images are matched through the similarity between the descriptors; for straight line features, a mean normalized descriptor is calculated and then matched.
5. The feature-fused RGB-D camera motion estimation method according to claim 1, wherein said step four includes the steps of:
t represents the motion conversion of two adjacent frames, and comprises T (X), RX + T, wherein R is a rotation matrix, T is a displacement vector, and a six-dimensional vector ξ is formed by taking the non-0 items of the rotation matrix and the displacement vector;
to be provided withAs a set of matched three-dimensional point sets of two adjacent frames F and F',then the three-dimensional straight line feature matching pairs of two adjacent frames are obtained; if three-dimensional point matching pairsThe corresponding real physical point is X under the local coordinate of the previous frameiThree-dimensional points, matching pairs in three-dimensional linesCorresponding to a true physical straight line of YjThen, the variable to be optimized in the camera odometer system is
The error function defining the system includes errors for point features and errors for straight line features, i.e.
<mrow> <mi>f</mi> <mrow> <mo>(</mo> <mi>&amp;Delta;</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfenced open = "[" close = "]"> <mtable> <mtr> <mtd> <msub> <mi>f</mi> <mi>P</mi> </msub> </mtd> </mtr> <mtr> <mtd> <msub> <mi>f</mi> <mi>L</mi> </msub> </mtd> </mtr> </mtable> </mfenced> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>9</mn> <mo>)</mo> </mrow> </mrow>
Wherein the error of the point feature is
<mrow> <msub> <mi>f</mi> <mi>P</mi> </msub> <mo>=</mo> <mfenced open = "[" close = "]"> <mtable> <mtr> <mtd> <mrow> <msub> <mi>X</mi> <mn>1</mn> </msub> <mo>-</mo> <msub> <mi>P</mi> <mn>1</mn> </msub> </mrow> </mtd> </mtr> <mtr> <mtd> <mo>.</mo> </mtd> </mtr> <mtr> <mtd> <mo>.</mo> </mtd> </mtr> <mtr> <mtd> <mo>.</mo> </mtd> </mtr> <mtr> <mtd> <mrow> <msub> <mi>X</mi> <mi>n</mi> </msub> <mo>-</mo> <msub> <mi>P</mi> <mi>n</mi> </msub> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mi>T</mi> <mrow> <mo>(</mo> <msub> <mi>X</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> <mo>-</mo> <msubsup> <mi>P</mi> <mn>1</mn> <mo>&amp;prime;</mo> </msubsup> </mrow> </mtd> </mtr> <mtr> <mtd> <mo>.</mo> </mtd> </mtr> <mtr> <mtd> <mo>.</mo> </mtd> </mtr> <mtr> <mtd> <mo>.</mo> </mtd> </mtr> <mtr> <mtd> <mrow> <mi>T</mi> <mrow> <mo>(</mo> <msub> <mi>X</mi> <mi>n</mi> </msub> <mo>)</mo> </mrow> <mo>-</mo> <msubsup> <mi>P</mi> <mi>n</mi> <mo>&amp;prime;</mo> </msubsup> </mrow> </mtd> </mtr> </mtable> </mfenced> <mo>,</mo> </mrow>
Error of straight line feature is
<mrow> <msub> <mi>f</mi> <mi>L</mi> </msub> <mo>=</mo> <mfenced open = "[" close = "]"> <mtable> <mtr> <mtd> <mrow> <mi>&amp;eta;</mi> <mrow> <mo>(</mo> <msub> <mi>L</mi> <mn>1</mn> </msub> <mo>-</mo> <msub> <mi>Y</mi> <mn>1</mn> </msub> <mo>)</mo> </mrow> </mrow> </mtd> </mtr> <mtr> <mtd> <mo>.</mo> </mtd> </mtr> <mtr> <mtd> <mo>.</mo> </mtd> </mtr> <mtr> <mtd> <mo>.</mo> </mtd> </mtr> <mtr> <mtd> <mrow> <mi>&amp;eta;</mi> <mrow> <mo>(</mo> <msub> <mi>L</mi> <mi>m</mi> </msub> <mo>-</mo> <msub> <mi>Y</mi> <mi>m</mi> </msub> <mo>)</mo> </mrow> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mi>&amp;eta;</mi> <mrow> <mo>(</mo> <msubsup> <mi>L</mi> <mn>1</mn> <mo>&amp;prime;</mo> </msubsup> <mo>-</mo> <mi>T</mi> <mo>(</mo> <msub> <mi>Y</mi> <mn>1</mn> </msub> <mo>)</mo> <mo>)</mo> </mrow> </mrow> </mtd> </mtr> <mtr> <mtd> <mo>.</mo> </mtd> </mtr> <mtr> <mtd> <mo>.</mo> </mtd> </mtr> <mtr> <mtd> <mo>.</mo> </mtd> </mtr> <mtr> <mtd> <mrow> <mi>&amp;eta;</mi> <mrow> <mo>(</mo> <msubsup> <mi>L</mi> <mi>m</mi> <mo>&amp;prime;</mo> </msubsup> <mo>-</mo> <mi>T</mi> <mo>(</mo> <msub> <mi>Y</mi> <mi>m</mi> </msub> <mo>)</mo> <mo>)</mo> </mrow> </mrow> </mtd> </mtr> </mtable> </mfenced> <mo>,</mo> </mrow>
Error of a single straight line is η (L)i-Yi)=[δ(Ai,Yi)T,δ(Bi,Yi)T]And is and
setting the target to find the variable Δ that minimizes the system error of the camera odometer system, i.e.
<mrow> <munder> <mrow> <mi>m</mi> <mi>i</mi> <mi>n</mi> </mrow> <mi>&amp;Delta;</mi> </munder> <mi>f</mi> <msup> <mrow> <mo>(</mo> <mi>&amp;Delta;</mi> <mo>)</mo> </mrow> <mi>T</mi> </msup> <msubsup> <mo>&amp;Sigma;</mo> <mi>f</mi> <mrow> <mo>-</mo> <mn>1</mn> </mrow> </msubsup> <mi>f</mi> <mrow> <mo>(</mo> <mi>&amp;Delta;</mi> <mo>)</mo> </mrow> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>10</mn> <mo>)</mo> </mrow> </mrow>
Wherein, sigmaf=diag(ΣPL) Solving the above formula (10) by using a Levenberg algorithm to obtain an initial pose of the camera motion; and constructing a pose graph by using initial pose information of camera motion, and optimizing the motion track of the camera by using the pose graph at the back end of the SLAM through a graph optimization method.
6. The method as claimed in claim 3, wherein the step 2 of discarding the points with abnormal depth values is discarding the points with null depth values.
CN201711297500.6A 2017-12-08 2017-12-08 The RGB-D camera motion methods of estimation of Fusion Features Pending CN108053445A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711297500.6A CN108053445A (en) 2017-12-08 2017-12-08 The RGB-D camera motion methods of estimation of Fusion Features

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711297500.6A CN108053445A (en) 2017-12-08 2017-12-08 The RGB-D camera motion methods of estimation of Fusion Features

Publications (1)

Publication Number Publication Date
CN108053445A true CN108053445A (en) 2018-05-18

Family

ID=62123166

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711297500.6A Pending CN108053445A (en) 2017-12-08 2017-12-08 The RGB-D camera motion methods of estimation of Fusion Features

Country Status (1)

Country Link
CN (1) CN108053445A (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109993747A (en) * 2019-03-22 2019-07-09 上海理工大学 Merge the rapid image matching method of dotted line feature
CN110162038A (en) * 2019-05-07 2019-08-23 杭州迦智科技有限公司 Control method for movement, device, storage medium and processor
CN110517301A (en) * 2019-07-22 2019-11-29 杭州电子科技大学 The method of feature is effectively matched under a kind of quick camera motion
CN110807799A (en) * 2019-09-29 2020-02-18 哈尔滨工程大学 Line feature visual odometer method combining depth map inference
CN111047620A (en) * 2019-11-15 2020-04-21 广东工业大学 Unmanned aerial vehicle visual odometer method based on depth point-line characteristics
CN114800504A (en) * 2022-04-26 2022-07-29 平安普惠企业管理有限公司 Robot posture analysis method, device, equipment and storage medium
CN117351140A (en) * 2023-09-15 2024-01-05 中国科学院自动化研究所 Three-dimensional reconstruction method, device and equipment integrating panoramic camera and laser radar

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102692236A (en) * 2012-05-16 2012-09-26 浙江大学 Visual milemeter method based on RGB-D camera
CN104851094A (en) * 2015-05-14 2015-08-19 西安电子科技大学 Improved method of RGB-D-based SLAM algorithm
EP2910187A1 (en) * 2014-02-24 2015-08-26 Université de Strasbourg (Etablissement Public National à Caractère Scientifique, Culturel et Professionnel) Automatic multimodal real-time tracking of moving instruments for image plane alignment inside a MRI scanner
CN105783913A (en) * 2016-03-08 2016-07-20 中山大学 SLAM device integrating multiple vehicle-mounted sensors and control method of device
CN105938619A (en) * 2016-04-11 2016-09-14 中国矿业大学 Visual odometer realization method based on fusion of RGB and depth information
CN106127739A (en) * 2016-06-16 2016-11-16 华东交通大学 A kind of RGB D SLAM method of combination monocular vision
CN106780592A (en) * 2016-06-30 2017-05-31 华南理工大学 Kinect depth reconstruction algorithms based on camera motion and image light and shade

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102692236A (en) * 2012-05-16 2012-09-26 浙江大学 Visual milemeter method based on RGB-D camera
EP2910187A1 (en) * 2014-02-24 2015-08-26 Université de Strasbourg (Etablissement Public National à Caractère Scientifique, Culturel et Professionnel) Automatic multimodal real-time tracking of moving instruments for image plane alignment inside a MRI scanner
CN104851094A (en) * 2015-05-14 2015-08-19 西安电子科技大学 Improved method of RGB-D-based SLAM algorithm
CN105783913A (en) * 2016-03-08 2016-07-20 中山大学 SLAM device integrating multiple vehicle-mounted sensors and control method of device
CN105938619A (en) * 2016-04-11 2016-09-14 中国矿业大学 Visual odometer realization method based on fusion of RGB and depth information
CN106127739A (en) * 2016-06-16 2016-11-16 华东交通大学 A kind of RGB D SLAM method of combination monocular vision
CN106780592A (en) * 2016-06-30 2017-05-31 华南理工大学 Kinect depth reconstruction algorithms based on camera motion and image light and shade

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
YAN LU 等: "Robust RGB-D Odometry Using Point and Line Features", 《2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION》 *
王亚龙 等: "基于Kinect的三维视觉里程计的设计", 《计算机应用》 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109993747A (en) * 2019-03-22 2019-07-09 上海理工大学 Merge the rapid image matching method of dotted line feature
CN110162038A (en) * 2019-05-07 2019-08-23 杭州迦智科技有限公司 Control method for movement, device, storage medium and processor
CN110517301A (en) * 2019-07-22 2019-11-29 杭州电子科技大学 The method of feature is effectively matched under a kind of quick camera motion
CN110517301B (en) * 2019-07-22 2022-04-01 杭州电子科技大学 Method for effectively matching features under rapid camera motion
CN110807799A (en) * 2019-09-29 2020-02-18 哈尔滨工程大学 Line feature visual odometer method combining depth map inference
CN110807799B (en) * 2019-09-29 2023-05-30 哈尔滨工程大学 Line feature visual odometer method combined with depth map inference
CN111047620A (en) * 2019-11-15 2020-04-21 广东工业大学 Unmanned aerial vehicle visual odometer method based on depth point-line characteristics
CN114800504A (en) * 2022-04-26 2022-07-29 平安普惠企业管理有限公司 Robot posture analysis method, device, equipment and storage medium
CN117351140A (en) * 2023-09-15 2024-01-05 中国科学院自动化研究所 Three-dimensional reconstruction method, device and equipment integrating panoramic camera and laser radar
CN117351140B (en) * 2023-09-15 2024-04-05 中国科学院自动化研究所 Three-dimensional reconstruction method, device and equipment integrating panoramic camera and laser radar

Similar Documents

Publication Publication Date Title
Fan et al. Road surface 3D reconstruction based on dense subpixel disparity map estimation
CN108053445A (en) The RGB-D camera motion methods of estimation of Fusion Features
WO2022188094A1 (en) Point cloud matching method and apparatus, navigation method and device, positioning method, and laser radar
CN108229416B (en) Robot SLAM method based on semantic segmentation technology
CN107358629B (en) Indoor mapping and positioning method based on target identification
CN113223045B (en) Vision and IMU sensor fusion positioning system based on dynamic object semantic segmentation
CN111862673A (en) Parking lot vehicle self-positioning and map construction method based on top view
Reina et al. 3D traversability awareness for rough terrain mobile robots
CN104281148A (en) Mobile robot autonomous navigation method based on binocular stereoscopic vision
Chen et al. A stereo visual-inertial SLAM approach for indoor mobile robots in unknown environments without occlusions
CN111998862A (en) Dense binocular SLAM method based on BNN
Guan et al. Minimal solvers for relative pose estimation of multi-camera systems using affine correspondences
CN109544632B (en) Semantic SLAM object association method based on hierarchical topic model
Beauvisage et al. Robust multispectral visual-inertial navigation with visual odometry failure recovery
CN112731503A (en) Pose estimation method and system based on front-end tight coupling
Wong et al. Monocular localization within sparse voxel maps
Aggarwal Machine vision based SelfPosition estimation of mobile robots
CN113744301B (en) Motion trail estimation method and device for mobile robot and storage medium
Kim et al. Image-Based ICP algorithm for visual odometry using a RGB-D sensor in a dynamic environment
Song et al. MF-LIO: integrating multi-feature LiDAR inertial odometry with FPFH loop closure in SLAM
Sun Dance training movement depth information recognition based on artificial intelligence
Ren An improved binocular LSD_SLAM method for object localization
Zhi et al. A Dynamic Visual-Inertial-Wheel Odometry With Semantic Constraints And Denoised IMU-Odometer Prior For Autonomous Driving
Shilin et al. Application of a Depth Camera for Constructing Complex Three-Dimensional Models in Multiple Scanning Complexes
Tamura et al. Visual odometry with effective feature sampling for untextured outdoor environment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20180518