CN103983270A - Graphic sonar data processing method - Google Patents
Graphic sonar data processing method Download PDFInfo
- Publication number
- CN103983270A CN103983270A CN201410210477.2A CN201410210477A CN103983270A CN 103983270 A CN103983270 A CN 103983270A CN 201410210477 A CN201410210477 A CN 201410210477A CN 103983270 A CN103983270 A CN 103983270A
- Authority
- CN
- China
- Prior art keywords
- point
- calculate
- max
- sonar
- current
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000003672 processing method Methods 0.000 title abstract description 5
- 238000000034 method Methods 0.000 claims abstract description 55
- 230000004807 localization Effects 0.000 claims abstract description 6
- 238000013507 mapping Methods 0.000 claims abstract 3
- 238000006243 chemical reaction Methods 0.000 claims description 10
- 230000008878 coupling Effects 0.000 claims description 7
- 238000010168 coupling process Methods 0.000 claims description 7
- 238000005859 coupling reaction Methods 0.000 claims description 7
- 239000011159 matrix material Substances 0.000 claims description 6
- 238000000605 extraction Methods 0.000 claims description 3
- 238000004364 calculation method Methods 0.000 claims description 2
- 238000001914 filtration Methods 0.000 claims description 2
- 229910052757 nitrogen Inorganic materials 0.000 claims description 2
- 238000013341 scale-up Methods 0.000 claims description 2
- 238000013480 data collection Methods 0.000 claims 1
- 235000012489 doughnuts Nutrition 0.000 claims 1
- 230000004927 fusion Effects 0.000 claims 1
- 238000012545 processing Methods 0.000 abstract description 10
- 238000010586 diagram Methods 0.000 description 8
- 230000007613 environmental effect Effects 0.000 description 5
- 230000000694 effects Effects 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 229910052717 sulfur Inorganic materials 0.000 description 2
- 241001061260 Emmelichthys struhsakeri Species 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000013506 data mapping Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 229910052760 oxygen Inorganic materials 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 229910052698 phosphorus Inorganic materials 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000002604 ultrasonography Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S15/00—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
- G01S15/02—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems using reflection of acoustic waves
- G01S15/06—Systems determining the position data of a target
- G01S15/08—Systems for measuring distance only
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/52—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
- G01S7/539—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 using analysis of echo signal for target characterisation; Target signature; Target cross-section
Landscapes
- Engineering & Computer Science (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Networks & Wireless Communication (AREA)
- Automation & Control Theory (AREA)
- Acoustics & Sound (AREA)
- Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)
Abstract
The invention discloses a graphic sonar data processing method which can be effectively applied to mobile robot localization, map building and route planning processes. The method comprises the following steps: performing noise reduction treatment on sonar data, extracting angular point information from the data, and expanding an angular point range based on an error ellipse and a largest error circle so as to build a local expansion map; mapping the local expansion map to a binary image space, establishing a local expansion map in the binary image space, and matching the current local expansion map with a historical local expansion map through a rotational invariance matching method, a two-dimensional scanning matching method and a scale invariant feature transform (SIFT) matching method. Compared with a traditional direct sonar data processing method, the novel sonar distance data processing method provided by the invention has the advantages that the information contained in the sonar data can be fully explored due to graphic sonar data processing, the sonar data processing mode is enriched, and the precision and robustness of sonar data processing are improved.
Description
Technical field
The present invention relates to a kind of disposal route of sonar range data, by range data being mapped to the mode of image space, realized the image conversion of sonar data and processed.Belong to Mobile Robotics Navigation field.
Background technology
Along with the progress of airmanship, mobile robot is widely used to assist the mankind to complete the tasks such as circumstances not known detection, and " Jade Hare " lunar rover of China is typical example.The inter-related task of surveying for completing circumstances not known, robot need possess the function of independent navigation.Generally, does independent navigation comprise three subproblems: 1. I am at that? which will 2. I go? which do 3. how I go? the corresponding robot of difference location, map building and path planning.The perception of robot self pose and environmental characteristic positional information is the prerequisite addressing the above problem.
At present, mobile robot mainly uses vision sensor, laser sensor and ultrasonic sensor etc. to obtain the positional information of self and environmental characteristic.Wherein, the abundant information that vision sensor obtains, but require robot to possess data-handling capacity fast, in addition, vision sensor disturbs responsive to light, block etc., limited its range of application.Laser sensor and ultrasonic sensor are range sensor, by the distance between robot measurement and environmental characteristic, provide robot navigation required information.Laser sensor response is fast, and the precision of information of obtaining is high, and still, laser sensor installation accuracy requires high, expensive.Relative, ultrasonic sensor is installed simple, and price is relatively low, can obtain the information that precision is relatively high, and therefore, ultrasonic sensor is still widely used.But because field angle is larger, the information that ultrasonic sensor obtains exists uncertain.Probability theory, fuzzy theory and gray system theory etc. are all used to express, process ultrasound information, and finally realize robot map building, location and path planning.
Taking a broad view of current sonar information by using disposal route, is generally directly to process the range data that sonar obtains, and by calculating the statistical information of range data, sets up the formalized description of environmental characteristic.Typical example is as the rasterizing describing method of environment and characteristics map creation method.Affected by field angle, the range information that sonar obtains is inevitably with error.The statistical information that current sonar range data processing method has utilized packet to contain to greatest extent, and in mobile robot autonomous navigation, obtained successful Application.But when error is larger, the accuracy meeting of statistical information is affected, meanwhile, existing method is also difficult to further from raw data, excavate useful information.For this reason, sonar range data are mapped to image space from metric space herein, the related art method of utilizing image to process realizes the processing of sonar range information.The method can effectively be applied in localization for Mobile Robot, map building and path planning process.
Summary of the invention
The object of the invention is the range data of sonar to map to image space, utilize image processing techniques to realize the processing of sonar range data, by the processing of image conversion, excavate to greatest extent the environmental information that sonar data comprises, the new approaches that one aspect of the present invention provides sonar data to process, on the other hand, by the image conversion matching technique of multiple sonar data, precision and robustness that sonar data is processed have been improved.
The invention provides a kind of image conversion disposal route of sonar data, by sonar range information is mapped to image space, the method for utilizing image to process, realizes the processing of sonar range information, mainly comprises the following steps:
Step 1, filter owing to measuring blind area or exceeding measured value that sonar to measure scope produces (if note sonar data is (x, y, θ, ra), wherein (x, y) represents the coordinate of target, θ is the orientation of the relative robot of target, and ra is that target is to the distance of robot.Undesirable data markers is ra=R, the maximum measuring distance that wherein R is sonar sensor, R=5000mm in the present invention), the data set of recording a demerit after filter is S.
Step 2, filtration singular value.In the present invention, singular value refers to not represent measured value any physical presence feature, relatively sparse.Data-oriented collection S, calculate respectively in S Euclidean distance a little between any two, and according to the size of coordinate and distance, the point in S is classified.Add up each classification and comprise number a little, when the number of classification mid point is while being less than threshold value Num, remove institute in respective class a little, Num be the threshold value of setting in advance.The data set of recording a demerit after filter singular value is S
0, the present invention claims S
0for range data space, referred to as metric space.
Step 3, extraction angle point.Definition length is N, and sliding step is s, and glide direction is
moving window.Edge
direction, from S
0in get successively N point, remember that its transverse and longitudinal coordinate is respectively X
t=[x
1, x
2..., x
n] and Y
t=[y
1, y
2..., y
n], X
t, Y
tcovariance matrix be:
Wherein:
with
be respectively X
tand Y
tthe average of element.Note C
teigenwert be λ
maxand λ
min, its ratio is EVR=λ
min/ λ
max.Along with the slip of window, can calculate S
0all EVR values, obtain current EVR curve.At corner point, EVR can reach extreme value.The present invention calculates the peak value of current EVR curve by the mode of more adjacent EVR value, the corresponding angle point of each peak value.
Step 4, error of calculation ellipse.Calculate the peak value of current EVR curve, this step be take peak value O as example, and the computation process that elaborates error ellipse is as follows: calculate sonar data point (x corresponding to O, y), then get and take (x, y) as the center of circle, r' is n data point in radius, and its transverse and longitudinal coordinate is designated as respectively X=[x
1..., x
n] and Y=[y
1..., y
n].If the covariance matrix of note X, Y is C, the eigenwert of C is λ
1and λ
2(λ
1>=λ
2), corresponding proper vector is v
1and v
2, error ellipse is v
1and v
2in the coordinate system forming with
for the center of circle, λ
1and λ
2be respectively the ellipse of major axis and minor axis, wherein
Step 5, calculating maximum error circle, set up the local map of expanding.With
for the center of circle, λ
1for radius, set up the maximum error circle that error ellipse is corresponding.According to the actual connection of angle point, calculate the public outer tangent line of the maximum error circle that the actual angle point that is connected is corresponding, further calculate the point of contact of tangent line and maximum error circle, connect the actual point of contact that connects the corresponding maximum error circle of angle point and form the local map M that expands, the center of circle of all maximum error circles, point of contact form the local crucial point set im_point that expands map M.
Step 6, map to image space.The transverse and longitudinal coordinate that note im_point comprises is a little respectively P=(p
1..., p
m), Q=(q
1..., q
m), p
maxand q
maxthe maximal value that represents respectively transverse and longitudinal coordinate, p
minand q
minbe respectively the minimum value of transverse and longitudinal coordinate.A bit (p in im_point
i, q
i) by formula (2) and formula (3), map to a bit (h in image space
i, k
i):
Wherein, η is scale-up factor, and η is relevant to the yardstick of sonar data and image space.
Step 7, note (h
i, k
i) be the picture in the center of circle in im_point, r
ifor corresponding radius of a circle, i=1 ..., j, j is center of circle number.With (h
i, k
i) be the center of circle, r
ifor radius, determine border circular areas, making the pixel value in the definite region of the picture at point of contact in this border circular areas and im_point is 0, otherwise is 1.Remember that the region that in this bianry image space, pixel value is 0 is the part expansion map p_M in image space.
Step 8, location matches.Location matches mainly comprises two aspects: rotational invariance coupling and two-dimensional scan coupling.
Steps A, rotational invariance coupling.
Between steps A 1, note p_M horizontal and vertical direction pixel, the maximal value of distance is respectively h
maxand v
max, the present invention defines P
c(int (h
max/ 2), int (v
max/ 2)) be the central point of p_M.With P
cfor true origin, along the horizontal and vertical direction of p_M, set up coordinate system Σ
c;
Steps A 2, make D=even ((h
max/ 2)
2+ (v
max/ 2)
2)
1/2, " even " represents ((h
max/ 2)
2+ (v
max/ 2)
2)
1/2upwards get even number.At coordinate system Σ
cin, with P
cfor the center of circle, as radius, be respectively the concentric circles of r and r+ Δ r, form donut R
r.R
rthe number of pixels comprising is designated as N
r, the pixel count wherein being occupied by p_M is designated as U
r, its ratio is designated as v (r)=U
r/ N
r, be called effective duty cycle corresponding to radius r.According to this, calculate initial point to all effective duty cycles within the scope of D/2, form effective duty cycle vector V=[v (0), v (Δ r) ... v (r) ... v (D/2)]
t.
Steps A 3, the effective duty cycle vector of establishing current and historical p_M are respectively V
l=[v
l(0), v
l(Δ r) ... v
l(r) ... v
l(D/2)]
tand V
d=[v
d(0), v
d(Δ r) ... v
d(r) ... v
d(D/2)]
t, V
land V
dmatching rate be:
Wherein,
Step B, two-dimensional scan coupling.
Step B1, the establishment of coordinate system method of mentioning according to steps A 1, set up respectively the coordinate system of current p_M and historical p_M;
Step B2, remember that the pixel count of picture horizontal and vertical direction in current bianry image space is N, the pixel count of the capable and i row of i that current p_M occupies is N respectively
riand N
ci, the effective duty cycle of i row and column is designated as respectively w
ri=N
ri/ N and w
ci=N
ci/ N.Generally, due to current p_M and historical p_M towards difference, the dutycycle of same sequence number ranks is generally different.For this reason, the present invention realizes the coupling of historical p_M and current p_M according to following rule:
2.1 fixing current p_M, remember that the effective duty cycle of its i row and column is
with
calculate the effective duty cycle of its all ranks, generate the effective duty cycle vector of current p_M
with
2.2 with 2.1, remember that the effective duty cycle vector of historical p_M is respectively
with
2.3 utilize formula (4) to calculate respectively the canonical correlation coefficient of two dimensions
with
.Historical p_M is rotated counterclockwise 1 degree, recalculates the canonical correlation coefficient of current p_M and historical p_M, is designated as
with
.Successively, calculate respectively 360 groups of canonical correlation coefficients in two dimension one-periods, generate two dimensions related coefficient vector separately:
with
If the current p_M of step B3 and historical p_M represent identical local environment, related coefficient vector λ
rand λ
cthe some continuous element of middle correspondence will surpass the threshold value λ setting
th; Otherwise, λ
rand λ
celement be all less than λ
thor only have some discrete elements to surpass threshold value.Get λ
rand λ
cin comprise surpass threshold value λ
ththe average of element, be designated as respectively
with
the matching rate of current p_M and historical p_M
Step 9, images match.Utilize SIFT operator to extract respectively the point of interest of current p_M and historical p_M, and carry out images match, note matching rate is P
p, P in the present invention
pmake a comment or criticism really the point of interest logarithm of coupling and the ratio of all match interest point logarithms.
Step 10, matching rate merge.In the present invention, stipulate that final matching rate is P
f=α p
rpt+ β p
2d+ γ p
p, wherein α, β and γ are respectively the weight of each matching rate, meet alpha+beta+γ=1.The value of α, β and γ determines according to the matching precision of each method in practical application.If P
fdo not meet threshold value requirement, store current p_M; Otherwise, can utilize match information to carry out the tasks such as localization for Mobile Robot, map building and path planning.
The present invention has following beneficial effect
1, by sonar range data-mapping to image space, utilize image processing techniques to process sonar range data, provide a kind of new sonar data to process thinking.
2, by setting up the local map of expanding, improved the robustness that sonar data is processed.
3, proposed three kinds of sonar data image conversion matching process, three kinds of methods complement each other, and have improved matching precision.
4, the image conversion of sonar data is processed and can be effectively applied in localization for Mobile Robot, map building and path planning process, improves precision and the robustness of robot navigation's task.
Accompanying drawing explanation
Fig. 1 is sonar data image conversion Processing Algorithm process flow diagram;
Fig. 2 (a), Fig. 2 (b) are respectively data set S and S
0, Fig. 2 (b) has shown error ellipse and maximum error circle simultaneously;
Fig. 3 (a), Fig. 3 (b) are respectively rotational invariance matching process schematic diagram and rotational invariance matching method matches result schematic diagram;
Fig. 4 (a), Fig. 4 (b) are respectively two-dimensional scan matching process schematic diagram and two-dimensional scan matching method matches result schematic diagram;
Fig. 5 (a), Fig. 5 (b) are respectively current p_M and historical p_M interest point extraction result schematic diagram;
Fig. 6 (a) is the schematic diagram of current and historical p_M SIFT interest points matching effect, and Fig. 6 (b) is that current p_M point of interest is from the schematic diagram of matching effect.
Embodiment
The present embodiment is implemented take invention technical scheme under prerequisite, provided detailed embodiment and process, but practical range of the present invention is not limited to following embodiment.
The present embodiment utilizes 16 sonar sensors image data under corridor-office environment of the Pioneer3-DX of robot equipment, utilize Visual Studio2008, OpenCV-1.0.0 and Matlab R2009a hybrid programming are realized the image conversion of sonar data and are processed, and algorithm flow as shown in Figure 1.The concrete execution step of the present invention is as follows:
(1) utilize robot sonar to gather current local environment data, the sonar data point of separated ra=5000mm, separating resulting, as shown in Fig. 2 (a), filters the singular value in sonar data, sets up metric space S
0, as shown in Fig. 2 (b);
(2) in the present embodiment, get N=10, s=1, X
t=[x
1, x
2..., x
10], Y
t=[y
1, y
2..., y
10].First, by formula (1), calculate X
tand Y
tcovariance matrix C
t, then, utilize the eig () function providing in Matlab, calculate C
tproper vector and individual features value, further calculate EVR=λ
min/ λ
max.
(3) edge
with step-length s moving window, be less than N to data in window.Repeating step (2), calculates the EVR that all windows are corresponding, forms current distance space S
0eVR curve.
(4) more adjacent EVR value, calculates the peak value of current EVR curve.Suppose that O is one of them peak value, the sonar data point that O is corresponding is (x, y).Get and take (x, y) as the center of circle, the data point of the n in the circle that r' is radius, its coordinate is designated as X=[x
1..., x
n] and Y=[y
1..., y
n].Calculate the covariance matrix C of X, Y, the eigenvalue λ of C
1and λ
2and corresponding proper vector v
1and v
2.At v
1and v
2in the coordinate system forming, with
for the center of circle, λ
1and λ
2for major axis and minor axis are determined error ellipse, wherein
with
for the center of circle, λ
1for radius, determine maximum error circle.
(5) according to step (4), calculate the maximum error circle that all peak values of current EVR curve are corresponding.According to the actual connection of angle point, calculate the public outer tangent line of the maximum error circle that the actual angle point that is connected is corresponding, further calculate the point of contact of tangent line and maximum error circle, build the local map M that expands, determine the crucial point set im_point of M (as shown in Fig. 3 (a) and Fig. 4 (a)).
(6) the transverse and longitudinal coordinate that note im_point comprises is a little respectively P=(p
1..., p
m) and Q=(q
1..., q
m), determine p
maxand q
max, and p
minand q
min.By formula (2) and formula (3) by the point (p in im_point
i, q
i) map to the point (h in image space
i, k
i), η=0.02 in the present invention.
(7) determine (h
i, k
i), r
i, i=1 ..., j.With (h
i, k
i) be the center of circle, r
ifor radius is determined border circular areas.Order is 0 by the pixel value in the definite region of the picture at this border circular areas and im_point point of contact, otherwise is 1, determines that map p_M (as shown in Fig. 5 (a) and Fig. 5 (b)) is expanded in the part in image space.
(8) location matches.
Steps A, rotational invariance coupling.
The maximal value h of distance between steps A 1, calculating p_M horizontal and vertical direction pixel
maxand v
max, further calculate the center point P of p_M
c(int (h
max/ 2), int (v
max/ 2)), set up coordinate system Σ
c, as shown in Fig. 3 (a);
Steps A 2, first calculate D=even ((h
max/ 2)
2+ (v
max/ 2)
2)
1/2, then calculate the effective duty cycle vector V of current and historical p_M
l=[v
l(0), v
l(Δ r) ... v
l(r) ... v
l(D/2)]
tand V
d=[v
d(0), v
d(Δ r) ... v
d(r) ... v
d(D/2)]
t, by formula (4), calculate V
land V
dmatching rate p
rpt(as shown in Fig. 3 (b)).
Step B, two-dimensional scan coupling.
Step B1, the establishment of coordinate system method proposing according to steps A 1, set up respectively the coordinate system of current p_M and historical p_M, as shown in Fig. 4 (a);
Step B2, calculate the effective duty cycle vector of current p_M
with
Step B3, calculate the effective duty cycle vector of historical p_M
with
Step B4, utilize formula (4) to calculate the canonical correlation coefficient vector of two dimensions
with
Step B5, calculating P
2d(as shown in Fig. 4 (b)).
(9) utilize SIFT operator to extract respectively the point of interest (as shown in Fig. 5 (a) and Fig. 5 (b)) of current p_M and historical p_M, and carry out images match (as shown in Fig. 6 (a) and Fig. 6 (b)), note matching rate is P
p.
(10) calculate final matching rate P
f=α p
rpt+ β p
2d+ γ p
p.In the present invention, test of many times result shows, the precision of three kinds of matching algorithms is just followed successively by: two-dimensional scan matching algorithm, and rotational invariance matching algorithm and image matching algorithm, so in this experiment, get α=0.33, β=0.37, γ=0.30.If P
fdo not meet threshold value requirement, store current p_M, otherwise, can utilize match information to carry out the tasks such as localization for Mobile Robot, map building and path planning.
Claims (11)
1. an image conversion disposal route for sonar data, is characterized in that specifically comprising the following steps:
(1) filter due to sonar blind area or exceed the measured value that sonar to measure scope produces;
(2) by the method for cluster, filter singular value and the discrete point in sonar data;
(3) calculate the EVR curve of sonar to measure value, by detecting the mode of current EVR peak of curve, obtain the environment angle point information that sonar data comprises;
(4) error ellipse of each Corner Feature that calculating sonar data comprises;
(5) calculate the maximum error circle of each error ellipse, build the current local map M that expands;
(6) set up the mapping relations in the current local map M of expansion and bianry image space;
(7) M is mapped to bianry image space, map p_M is expanded in the part in design of graphics image space;
(8) utilize rotational invariance matching process and two-dimensional scan matching process to realize the local coupling of expanding map;
(9) utilize SIFT operator extraction point of interest, and carry out images match;
(10) be rotated the fusion of unchangeability matching process, two-dimensional scan matching process and image matching method matching result.
2. the method for claim 1, is characterized in that step (1) comprising: note sonar data is (x, y, θ, ra), (x wherein, y) be coordinates of targets, θ is the orientation of the relative robot of target, and ra is that target is to the distance of robot, due to the impact of measuring blind area or exceeding sonar to measure scope, can produce undesirable data, be labeled as ra=R, the maximum measuring distance that wherein R is sonar sensor, the point of removing all ra=R, the data set after filtration is designated as S.
3. the method for claim 1, it is characterized in that step (2) comprising: calculate in sonar data collection S Euclidean distance a little between any two, and according to the size of coordinate and distance, the point in S is classified, add up each classification and comprise number a little, when the number of certain classification mid point is less than threshold value Num, remove institute in respective class a little, left point forms metric space S
0.
4. the method for claim 1, is characterized in that step (3) comprising: definition length is N, and sliding step is s, and glide direction is
moving window, edge
direction, from S
0in get successively N point, its coordinate is designated as respectively X
t=[x
1, x
2..., x
n] and Y
t=[y
1, y
2..., y
n], by:
Calculate X
tand Y
tcovariance matrix C
t, wherein
with
be respectively X
tand Y
tthe average of element, note C
tthe ratio EVR=λ of eigenwert
min/ λ
max, edge
with step-length s moving window, be less than N to data in window, calculate current EVR curve, by the mode of more adjacent EVR value, calculate the peak value of current EVR curve, the corresponding angle point of each peak value.
5. the method for claim 1, is characterized in that the described error of calculation ellipse of step (4) specifically comprises following content:
Suppose that O is one of them peak value, the sonar data point that O is corresponding is (x, y), get and take (x, y) as the center of circle, and the data point of the n in the circle that r' is radius, its coordinate is designated as X=[x
1..., x
n] and Y=[y
1..., y
n], the covariance matrix C of calculating X and Y, the eigenwert of C is λ
1and λ
2(λ
1>=λ
2), corresponding proper vector is v
1and v
2, error ellipse is v
1and v
2in the coordinate system forming with
for the center of circle, λ
1, λ
2be respectively the ellipse of major axis and minor axis, wherein
6. the method for claim 1, is characterized in that the current local map M that expands of the described structure of step (5), specifically comprises following content:
With
for the center of circle, λ
1for radius, build the maximum error circle that error ellipse is corresponding, according to the actual connection of angle point, calculate the public outer tangent line of the maximum error circle that the actual angle point that is connected is corresponding, further calculate the point of contact of tangent line and maximum error circle, the point of contact that connects the corresponding maximum error circle of the actual angle point that is connected forms the local map M that expands, and the center of circle of all maximum error circles, point of contact form the local crucial point set im_point that expands map M.
7. the method for claim 1, is characterized in that the described mapping relations of setting up M and bianry image space of step (6) comprise:
Horizontal ordinate and ordinate that note im_point comprises are a little respectively P=(p
1..., p
m) and Q=(q
1..., q
m), p
maxand q
maxbe respectively the maximal value of transverse and longitudinal coordinate, p
minand q
minbe respectively the minimum value of transverse and longitudinal coordinate, a bit (p in im_point
i, q
i) pass through formula
With
Map to a bit (h in image space
i, k
i), wherein, η is scale-up factor.
8. the method for claim 1, is characterized in that the described part in image space set up of step (7) expands map and comprise following content:
In note im_point, the picture in the center of circle and corresponding radius are respectively (h
i, k
i) and r
i, i=1 ..., j, j is center of circle number, with (h
i, k
i) be the center of circle, r
ifor radius, determine border circular areas, making the pixel value in the definite region of the picture at point of contact in this border circular areas and im_point is 0, otherwise is 1, remembers that region that in this bianry image space, pixel value is 0 is that map p_M is expanded in part in image space.
9. the method for claim 1, is characterized in that step (8) comprising:
(8.1) utilize the current p_M of rotational invariance matching method matches and historical p_M, specifically comprise following content:
(i) calculate the maximal value h of distance between p_M horizontal and vertical direction pixel
maxand v
max, further calculate the center point P of p_M
c(int (h
max/ 2), int (v
max/ 2)), along horizontal and vertical direction, set up coordinate system Σ
c, at coordinate system Σ
cin, with P
cfor the center of circle, as radius, be respectively the concentric circles of r and r+ Δ r, form donut R
r, R
rthe number of pixels comprising is designated as N
r, the pixel count wherein being occupied by p_M is designated as U
r, its ratio is designated as v (r)=U
r/ N
r, be called effective duty cycle corresponding to radius r;
(ii) calculate D=even ((h
max/ 2)
2+ (v
max/ 2)
2)
1/2, " even " represents ((h
max/ 2)
2+ (v
max/ 2)
2)
1/2upwards get even number, then calculate the effective duty cycle vector V of current and historical p_M
land V
d, and pass through
calculate V
land V
dmatching rate p
rpt, wherein, V
l=[v
l(0), v
l(Δ r) ... v
l(r) ... v
l(D/2)]
t, V
d=[v
d(0), v
d(Δ r) ... v
d(r) ... v
d(D/2)]
t,
(8.2) utilize the current p_M of two-dimensional scan matching method matches and historical p_M, specifically comprise following content:
(i), according to the establishment of coordinate system method in (8.1), set up respectively the coordinate system of current p_M and historical p_M;
(ii) pixel count of remembering picture horizontal and vertical direction in current bianry image space is N, and the pixel count of the capable and i row of the i that occupied by p_M is N respectively
riand N
ci, the effective duty cycle of i row and column is designated as respectively w
ri=N
ri/ N and w
ci=N
ci/ N, the effective duty cycle that calculates respectively all ranks of current p_M forms effective duty cycle vector
with
, in like manner, calculate the effective duty cycle vector of historical p_M
with
(iii) calculate respectively canonical correlation coefficient and the final matching degree P2d of current and historical p_M ranks effective duty cycle vector.
10. the method for claim 1, is characterized in that step (9) comprises following content: utilize SIFT operator to extract respectively the point of interest of current p_M and historical p_M, carry out images match, calculate matching rate P
p.
11. the method for claim 1, is characterized in that step (10) comprising:
Calculate final matching rate P
f=α p
rpt+ β p
2d+ γ p
p, alpha+beta+γ=1 wherein, α, the value of β and γ is determined according to the precision of each matching process, if P
fdo not meet threshold value requirement, store current p_M, otherwise, can utilize match information to carry out localization for Mobile Robot, map building and path planning task.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410210477.2A CN103983270B (en) | 2014-05-16 | 2014-05-16 | A kind of image conversion processing method of sonar data |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410210477.2A CN103983270B (en) | 2014-05-16 | 2014-05-16 | A kind of image conversion processing method of sonar data |
Publications (2)
Publication Number | Publication Date |
---|---|
CN103983270A true CN103983270A (en) | 2014-08-13 |
CN103983270B CN103983270B (en) | 2016-09-28 |
Family
ID=51275334
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410210477.2A Expired - Fee Related CN103983270B (en) | 2014-05-16 | 2014-05-16 | A kind of image conversion processing method of sonar data |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN103983270B (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106197421A (en) * | 2016-06-24 | 2016-12-07 | 北京工业大学 | A kind of forward position impact point for moving robot autonomous exploration generates method |
CN108268427A (en) * | 2018-01-10 | 2018-07-10 | 中国地质大学(武汉) | A kind of free kick goal probability analysis method, equipment and storage device |
CN108764080A (en) * | 2018-05-17 | 2018-11-06 | 中国电子科技集团公司第五十四研究所 | A kind of unmanned plane vision barrier-avoiding method based on cloud space binaryzation |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103247040A (en) * | 2013-05-13 | 2013-08-14 | 北京工业大学 | Layered topological structure based map splicing method for multi-robot system |
CN103278170A (en) * | 2013-05-16 | 2013-09-04 | 东南大学 | Mobile robot cascading map building method based on remarkable scenic spot detection |
-
2014
- 2014-05-16 CN CN201410210477.2A patent/CN103983270B/en not_active Expired - Fee Related
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103247040A (en) * | 2013-05-13 | 2013-08-14 | 北京工业大学 | Layered topological structure based map splicing method for multi-robot system |
CN103278170A (en) * | 2013-05-16 | 2013-09-04 | 东南大学 | Mobile robot cascading map building method based on remarkable scenic spot detection |
Non-Patent Citations (4)
Title |
---|
CHOI J ET AL.: "Autonomous topological modeling of a home environment and topological localization using a sonar grid map", 《AUTONOMOUS ROBOTS》 * |
CHOI Y-H ET AL.: "A line feature based SLAM with low grade range sensors using geometric constraints and active exploration for mobile robot", 《AUTONOMOUS ROBOTS》 * |
李润伟: "基于声纳的室内环境栅格地图创建方法的研究", 《中国优秀硕士学位论文全文数据库信息科技辑》 * |
高丽华等: "一种基于声纳信息的移动机器人地图创建方法", 《制造业自动化》 * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106197421A (en) * | 2016-06-24 | 2016-12-07 | 北京工业大学 | A kind of forward position impact point for moving robot autonomous exploration generates method |
CN106197421B (en) * | 2016-06-24 | 2019-03-22 | 北京工业大学 | A kind of forward position target point generation method independently explored for mobile robot |
CN108268427A (en) * | 2018-01-10 | 2018-07-10 | 中国地质大学(武汉) | A kind of free kick goal probability analysis method, equipment and storage device |
CN108764080A (en) * | 2018-05-17 | 2018-11-06 | 中国电子科技集团公司第五十四研究所 | A kind of unmanned plane vision barrier-avoiding method based on cloud space binaryzation |
CN108764080B (en) * | 2018-05-17 | 2021-10-01 | 中国电子科技集团公司第五十四研究所 | Unmanned aerial vehicle visual obstacle avoidance method based on point cloud space binarization |
Also Published As
Publication number | Publication date |
---|---|
CN103983270B (en) | 2016-09-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN103236064B (en) | A kind of some cloud autoegistration method based on normal vector | |
CN103727930B (en) | A kind of laser range finder based on edge matching and camera relative pose scaling method | |
CN104240297A (en) | Rescue robot three-dimensional environment map real-time construction method | |
CN112184736B (en) | Multi-plane extraction method based on European clustering | |
CN103295239A (en) | Laser-point cloud data automatic registration method based on plane base images | |
CN103617328A (en) | Airplane three-dimensional attitude computation method | |
CN112362072A (en) | High-precision point cloud map creation system and method in complex urban area environment | |
CN114998276B (en) | Robot dynamic obstacle real-time detection method based on three-dimensional point cloud | |
Qian et al. | Robust visual-lidar simultaneous localization and mapping system for UAV | |
CN105137412A (en) | Accurate fitting method of line segment features in 2D laser radar distance image | |
CN112949407A (en) | Remote sensing image building vectorization method based on deep learning and point set optimization | |
CN113177593A (en) | Fusion method of radar point cloud and image data in water traffic environment | |
CN103983270A (en) | Graphic sonar data processing method | |
Aissou et al. | Building roof superstructures classification from imbalanced and low density airborne LiDAR point cloud | |
CN102184536B (en) | Method and system for extracting straight line and/or line segment end points from image | |
CN109583626B (en) | Road network topology reconstruction method, medium and system | |
CN103700095A (en) | Log end surface image partitioning algorithm for improving active contour model based on circle constraint | |
CN111275748A (en) | Point cloud registration method based on laser radar in dynamic environment | |
CN113791400B (en) | Stair parameter autonomous detection method based on laser radar | |
CN113554705B (en) | Laser radar robust positioning method under changing scene | |
CN115453570A (en) | Multi-feature fusion mining area dust filtering method | |
CN111932612B (en) | Intelligent vehicle vision positioning method and device based on second-order hidden Markov model | |
CN111239761B (en) | Method for indoor real-time establishment of two-dimensional map | |
Fang et al. | ContinuityLearner: Geometric continuity feature learning for lane segmentation | |
CN113240721A (en) | Navigation vanishing point detection method applied to complex scene of roadway |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20160928 |