CN103983270A - Graphic sonar data processing method - Google Patents

Graphic sonar data processing method Download PDF

Info

Publication number
CN103983270A
CN103983270A CN201410210477.2A CN201410210477A CN103983270A CN 103983270 A CN103983270 A CN 103983270A CN 201410210477 A CN201410210477 A CN 201410210477A CN 103983270 A CN103983270 A CN 103983270A
Authority
CN
China
Prior art keywords
point
calculate
max
sonar
current
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201410210477.2A
Other languages
Chinese (zh)
Other versions
CN103983270B (en
Inventor
陈宗海
王鹏
张启彬
包鹏
徐子伟
孙建
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Science and Technology of China USTC
Original Assignee
University of Science and Technology of China USTC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Science and Technology of China USTC filed Critical University of Science and Technology of China USTC
Priority to CN201410210477.2A priority Critical patent/CN103983270B/en
Publication of CN103983270A publication Critical patent/CN103983270A/en
Application granted granted Critical
Publication of CN103983270B publication Critical patent/CN103983270B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S15/00Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
    • G01S15/02Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems using reflection of acoustic waves
    • G01S15/06Systems determining the position data of a target
    • G01S15/08Systems for measuring distance only
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/52Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
    • G01S7/539Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 using analysis of echo signal for target characterisation; Target signature; Target cross-section

Landscapes

  • Engineering & Computer Science (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Automation & Control Theory (AREA)
  • Acoustics & Sound (AREA)
  • Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)

Abstract

The invention discloses a graphic sonar data processing method which can be effectively applied to mobile robot localization, map building and route planning processes. The method comprises the following steps: performing noise reduction treatment on sonar data, extracting angular point information from the data, and expanding an angular point range based on an error ellipse and a largest error circle so as to build a local expansion map; mapping the local expansion map to a binary image space, establishing a local expansion map in the binary image space, and matching the current local expansion map with a historical local expansion map through a rotational invariance matching method, a two-dimensional scanning matching method and a scale invariant feature transform (SIFT) matching method. Compared with a traditional direct sonar data processing method, the novel sonar distance data processing method provided by the invention has the advantages that the information contained in the sonar data can be fully explored due to graphic sonar data processing, the sonar data processing mode is enriched, and the precision and robustness of sonar data processing are improved.

Description

A kind of image conversion disposal route of sonar data
Technical field
The present invention relates to a kind of disposal route of sonar range data, by range data being mapped to the mode of image space, realized the image conversion of sonar data and processed.Belong to Mobile Robotics Navigation field.
Background technology
Along with the progress of airmanship, mobile robot is widely used to assist the mankind to complete the tasks such as circumstances not known detection, and " Jade Hare " lunar rover of China is typical example.The inter-related task of surveying for completing circumstances not known, robot need possess the function of independent navigation.Generally, does independent navigation comprise three subproblems: 1. I am at that? which will 2. I go? which do 3. how I go? the corresponding robot of difference location, map building and path planning.The perception of robot self pose and environmental characteristic positional information is the prerequisite addressing the above problem.
At present, mobile robot mainly uses vision sensor, laser sensor and ultrasonic sensor etc. to obtain the positional information of self and environmental characteristic.Wherein, the abundant information that vision sensor obtains, but require robot to possess data-handling capacity fast, in addition, vision sensor disturbs responsive to light, block etc., limited its range of application.Laser sensor and ultrasonic sensor are range sensor, by the distance between robot measurement and environmental characteristic, provide robot navigation required information.Laser sensor response is fast, and the precision of information of obtaining is high, and still, laser sensor installation accuracy requires high, expensive.Relative, ultrasonic sensor is installed simple, and price is relatively low, can obtain the information that precision is relatively high, and therefore, ultrasonic sensor is still widely used.But because field angle is larger, the information that ultrasonic sensor obtains exists uncertain.Probability theory, fuzzy theory and gray system theory etc. are all used to express, process ultrasound information, and finally realize robot map building, location and path planning.
Taking a broad view of current sonar information by using disposal route, is generally directly to process the range data that sonar obtains, and by calculating the statistical information of range data, sets up the formalized description of environmental characteristic.Typical example is as the rasterizing describing method of environment and characteristics map creation method.Affected by field angle, the range information that sonar obtains is inevitably with error.The statistical information that current sonar range data processing method has utilized packet to contain to greatest extent, and in mobile robot autonomous navigation, obtained successful Application.But when error is larger, the accuracy meeting of statistical information is affected, meanwhile, existing method is also difficult to further from raw data, excavate useful information.For this reason, sonar range data are mapped to image space from metric space herein, the related art method of utilizing image to process realizes the processing of sonar range information.The method can effectively be applied in localization for Mobile Robot, map building and path planning process.
Summary of the invention
The object of the invention is the range data of sonar to map to image space, utilize image processing techniques to realize the processing of sonar range data, by the processing of image conversion, excavate to greatest extent the environmental information that sonar data comprises, the new approaches that one aspect of the present invention provides sonar data to process, on the other hand, by the image conversion matching technique of multiple sonar data, precision and robustness that sonar data is processed have been improved.
The invention provides a kind of image conversion disposal route of sonar data, by sonar range information is mapped to image space, the method for utilizing image to process, realizes the processing of sonar range information, mainly comprises the following steps:
Step 1, filter owing to measuring blind area or exceeding measured value that sonar to measure scope produces (if note sonar data is (x, y, θ, ra), wherein (x, y) represents the coordinate of target, θ is the orientation of the relative robot of target, and ra is that target is to the distance of robot.Undesirable data markers is ra=R, the maximum measuring distance that wherein R is sonar sensor, R=5000mm in the present invention), the data set of recording a demerit after filter is S.
Step 2, filtration singular value.In the present invention, singular value refers to not represent measured value any physical presence feature, relatively sparse.Data-oriented collection S, calculate respectively in S Euclidean distance a little between any two, and according to the size of coordinate and distance, the point in S is classified.Add up each classification and comprise number a little, when the number of classification mid point is while being less than threshold value Num, remove institute in respective class a little, Num be the threshold value of setting in advance.The data set of recording a demerit after filter singular value is S 0, the present invention claims S 0for range data space, referred to as metric space.
Step 3, extraction angle point.Definition length is N, and sliding step is s, and glide direction is moving window.Edge direction, from S 0in get successively N point, remember that its transverse and longitudinal coordinate is respectively X t=[x 1, x 2..., x n] and Y t=[y 1, y 2..., y n], X t, Y tcovariance matrix be:
C t = cov ( X t , X t ) cov ( X t , Y t ) cov ( Y t , X t ) cov ( Y t , Y t ) , - - - ( 1 )
Wherein: with be respectively X tand Y tthe average of element.Note C teigenwert be λ maxand λ min, its ratio is EVR=λ min/ λ max.Along with the slip of window, can calculate S 0all EVR values, obtain current EVR curve.At corner point, EVR can reach extreme value.The present invention calculates the peak value of current EVR curve by the mode of more adjacent EVR value, the corresponding angle point of each peak value.
Step 4, error of calculation ellipse.Calculate the peak value of current EVR curve, this step be take peak value O as example, and the computation process that elaborates error ellipse is as follows: calculate sonar data point (x corresponding to O, y), then get and take (x, y) as the center of circle, r' is n data point in radius, and its transverse and longitudinal coordinate is designated as respectively X=[x 1..., x n] and Y=[y 1..., y n].If the covariance matrix of note X, Y is C, the eigenwert of C is λ 1and λ 21>=λ 2), corresponding proper vector is v 1and v 2, error ellipse is v 1and v 2in the coordinate system forming with for the center of circle, λ 1and λ 2be respectively the ellipse of major axis and minor axis, wherein
Step 5, calculating maximum error circle, set up the local map of expanding.With for the center of circle, λ 1for radius, set up the maximum error circle that error ellipse is corresponding.According to the actual connection of angle point, calculate the public outer tangent line of the maximum error circle that the actual angle point that is connected is corresponding, further calculate the point of contact of tangent line and maximum error circle, connect the actual point of contact that connects the corresponding maximum error circle of angle point and form the local map M that expands, the center of circle of all maximum error circles, point of contact form the local crucial point set im_point that expands map M.
Step 6, map to image space.The transverse and longitudinal coordinate that note im_point comprises is a little respectively P=(p 1..., p m), Q=(q 1..., q m), p maxand q maxthe maximal value that represents respectively transverse and longitudinal coordinate, p minand q minbe respectively the minimum value of transverse and longitudinal coordinate.A bit (p in im_point i, q i) by formula (2) and formula (3), map to a bit (h in image space i, k i):
h i = int [ η ( p max - p min ) · ( p i - p min ) p max - p min ] , - - - ( 2 )
k i = int [ η ( q max - q min ) · ( q i - q min ) q max - q min ] , - - - ( 3 )
Wherein, η is scale-up factor, and η is relevant to the yardstick of sonar data and image space.
Step 7, note (h i, k i) be the picture in the center of circle in im_point, r ifor corresponding radius of a circle, i=1 ..., j, j is center of circle number.With (h i, k i) be the center of circle, r ifor radius, determine border circular areas, making the pixel value in the definite region of the picture at point of contact in this border circular areas and im_point is 0, otherwise is 1.Remember that the region that in this bianry image space, pixel value is 0 is the part expansion map p_M in image space.
Step 8, location matches.Location matches mainly comprises two aspects: rotational invariance coupling and two-dimensional scan coupling.
Steps A, rotational invariance coupling.
Between steps A 1, note p_M horizontal and vertical direction pixel, the maximal value of distance is respectively h maxand v max, the present invention defines P c(int (h max/ 2), int (v max/ 2)) be the central point of p_M.With P cfor true origin, along the horizontal and vertical direction of p_M, set up coordinate system Σ c;
Steps A 2, make D=even ((h max/ 2) 2+ (v max/ 2) 2) 1/2, " even " represents ((h max/ 2) 2+ (v max/ 2) 2) 1/2upwards get even number.At coordinate system Σ cin, with P cfor the center of circle, as radius, be respectively the concentric circles of r and r+ Δ r, form donut R r.R rthe number of pixels comprising is designated as N r, the pixel count wherein being occupied by p_M is designated as U r, its ratio is designated as v (r)=U r/ N r, be called effective duty cycle corresponding to radius r.According to this, calculate initial point to all effective duty cycles within the scope of D/2, form effective duty cycle vector V=[v (0), v (Δ r) ... v (r) ... v (D/2)] t.
Steps A 3, the effective duty cycle vector of establishing current and historical p_M are respectively V l=[v l(0), v l(Δ r) ... v l(r) ... v l(D/2)] tand V d=[v d(0), v d(Δ r) ... v d(r) ... v d(D/2)] t, V land V dmatching rate be:
p rpt = Σ r = 0 D / 2 { [ v l ( r ) - υ l ] [ v d ( r ) - υ d ] } { Σ r = 0 D / 2 [ v l ( r ) - υ l ] 2 Σ r = 0 D / 2 [ v d ( r ) - υ d ] 2 } - - - ( 4 )
Wherein, υ l = Σ r = 0 D / 2 v l ( r ) / ( 1 + D / 2 Δr ) , υ d = Σ r = 0 D / 2 v d ( r ) / ( 1 + D / 2 Δr ) .
Step B, two-dimensional scan coupling.
Step B1, the establishment of coordinate system method of mentioning according to steps A 1, set up respectively the coordinate system of current p_M and historical p_M;
Step B2, remember that the pixel count of picture horizontal and vertical direction in current bianry image space is N, the pixel count of the capable and i row of i that current p_M occupies is N respectively riand N ci, the effective duty cycle of i row and column is designated as respectively w ri=N ri/ N and w ci=N ci/ N.Generally, due to current p_M and historical p_M towards difference, the dutycycle of same sequence number ranks is generally different.For this reason, the present invention realizes the coupling of historical p_M and current p_M according to following rule:
2.1 fixing current p_M, remember that the effective duty cycle of its i row and column is with calculate the effective duty cycle of its all ranks, generate the effective duty cycle vector of current p_M with W c l = [ w c 1 l , w c 2 l , . . . , w cN l ] T ;
2.2 with 2.1, remember that the effective duty cycle vector of historical p_M is respectively with W c d = [ w c 1 d , w c 2 d , . . . , w cN d ] T ;
2.3 utilize formula (4) to calculate respectively the canonical correlation coefficient of two dimensions with .Historical p_M is rotated counterclockwise 1 degree, recalculates the canonical correlation coefficient of current p_M and historical p_M, is designated as with .Successively, calculate respectively 360 groups of canonical correlation coefficients in two dimension one-periods, generate two dimensions related coefficient vector separately: with λ c = [ λ c 0 , λ c 1 , . . . , λ c 359 ] T .
If the current p_M of step B3 and historical p_M represent identical local environment, related coefficient vector λ rand λ cthe some continuous element of middle correspondence will surpass the threshold value λ setting th; Otherwise, λ rand λ celement be all less than λ thor only have some discrete elements to surpass threshold value.Get λ rand λ cin comprise surpass threshold value λ ththe average of element, be designated as respectively with the matching rate of current p_M and historical p_M
Step 9, images match.Utilize SIFT operator to extract respectively the point of interest of current p_M and historical p_M, and carry out images match, note matching rate is P p, P in the present invention pmake a comment or criticism really the point of interest logarithm of coupling and the ratio of all match interest point logarithms.
Step 10, matching rate merge.In the present invention, stipulate that final matching rate is P f=α p rpt+ β p 2d+ γ p p, wherein α, β and γ are respectively the weight of each matching rate, meet alpha+beta+γ=1.The value of α, β and γ determines according to the matching precision of each method in practical application.If P fdo not meet threshold value requirement, store current p_M; Otherwise, can utilize match information to carry out the tasks such as localization for Mobile Robot, map building and path planning.
The present invention has following beneficial effect
1, by sonar range data-mapping to image space, utilize image processing techniques to process sonar range data, provide a kind of new sonar data to process thinking.
2, by setting up the local map of expanding, improved the robustness that sonar data is processed.
3, proposed three kinds of sonar data image conversion matching process, three kinds of methods complement each other, and have improved matching precision.
4, the image conversion of sonar data is processed and can be effectively applied in localization for Mobile Robot, map building and path planning process, improves precision and the robustness of robot navigation's task.
Accompanying drawing explanation
Fig. 1 is sonar data image conversion Processing Algorithm process flow diagram;
Fig. 2 (a), Fig. 2 (b) are respectively data set S and S 0, Fig. 2 (b) has shown error ellipse and maximum error circle simultaneously;
Fig. 3 (a), Fig. 3 (b) are respectively rotational invariance matching process schematic diagram and rotational invariance matching method matches result schematic diagram;
Fig. 4 (a), Fig. 4 (b) are respectively two-dimensional scan matching process schematic diagram and two-dimensional scan matching method matches result schematic diagram;
Fig. 5 (a), Fig. 5 (b) are respectively current p_M and historical p_M interest point extraction result schematic diagram;
Fig. 6 (a) is the schematic diagram of current and historical p_M SIFT interest points matching effect, and Fig. 6 (b) is that current p_M point of interest is from the schematic diagram of matching effect.
Embodiment
The present embodiment is implemented take invention technical scheme under prerequisite, provided detailed embodiment and process, but practical range of the present invention is not limited to following embodiment.
The present embodiment utilizes 16 sonar sensors image data under corridor-office environment of the Pioneer3-DX of robot equipment, utilize Visual Studio2008, OpenCV-1.0.0 and Matlab R2009a hybrid programming are realized the image conversion of sonar data and are processed, and algorithm flow as shown in Figure 1.The concrete execution step of the present invention is as follows:
(1) utilize robot sonar to gather current local environment data, the sonar data point of separated ra=5000mm, separating resulting, as shown in Fig. 2 (a), filters the singular value in sonar data, sets up metric space S 0, as shown in Fig. 2 (b);
(2) in the present embodiment, get N=10, s=1, X t=[x 1, x 2..., x 10], Y t=[y 1, y 2..., y 10].First, by formula (1), calculate X tand Y tcovariance matrix C t, then, utilize the eig () function providing in Matlab, calculate C tproper vector and individual features value, further calculate EVR=λ min/ λ max.
(3) edge with step-length s moving window, be less than N to data in window.Repeating step (2), calculates the EVR that all windows are corresponding, forms current distance space S 0eVR curve.
(4) more adjacent EVR value, calculates the peak value of current EVR curve.Suppose that O is one of them peak value, the sonar data point that O is corresponding is (x, y).Get and take (x, y) as the center of circle, the data point of the n in the circle that r' is radius, its coordinate is designated as X=[x 1..., x n] and Y=[y 1..., y n].Calculate the covariance matrix C of X, Y, the eigenvalue λ of C 1and λ 2and corresponding proper vector v 1and v 2.At v 1and v 2in the coordinate system forming, with for the center of circle, λ 1and λ 2for major axis and minor axis are determined error ellipse, wherein with for the center of circle, λ 1for radius, determine maximum error circle.
(5) according to step (4), calculate the maximum error circle that all peak values of current EVR curve are corresponding.According to the actual connection of angle point, calculate the public outer tangent line of the maximum error circle that the actual angle point that is connected is corresponding, further calculate the point of contact of tangent line and maximum error circle, build the local map M that expands, determine the crucial point set im_point of M (as shown in Fig. 3 (a) and Fig. 4 (a)).
(6) the transverse and longitudinal coordinate that note im_point comprises is a little respectively P=(p 1..., p m) and Q=(q 1..., q m), determine p maxand q max, and p minand q min.By formula (2) and formula (3) by the point (p in im_point i, q i) map to the point (h in image space i, k i), η=0.02 in the present invention.
(7) determine (h i, k i), r i, i=1 ..., j.With (h i, k i) be the center of circle, r ifor radius is determined border circular areas.Order is 0 by the pixel value in the definite region of the picture at this border circular areas and im_point point of contact, otherwise is 1, determines that map p_M (as shown in Fig. 5 (a) and Fig. 5 (b)) is expanded in the part in image space.
(8) location matches.
Steps A, rotational invariance coupling.
The maximal value h of distance between steps A 1, calculating p_M horizontal and vertical direction pixel maxand v max, further calculate the center point P of p_M c(int (h max/ 2), int (v max/ 2)), set up coordinate system Σ c, as shown in Fig. 3 (a);
Steps A 2, first calculate D=even ((h max/ 2) 2+ (v max/ 2) 2) 1/2, then calculate the effective duty cycle vector V of current and historical p_M l=[v l(0), v l(Δ r) ... v l(r) ... v l(D/2)] tand V d=[v d(0), v d(Δ r) ... v d(r) ... v d(D/2)] t, by formula (4), calculate V land V dmatching rate p rpt(as shown in Fig. 3 (b)).
Step B, two-dimensional scan coupling.
Step B1, the establishment of coordinate system method proposing according to steps A 1, set up respectively the coordinate system of current p_M and historical p_M, as shown in Fig. 4 (a);
Step B2, calculate the effective duty cycle vector of current p_M with
Step B3, calculate the effective duty cycle vector of historical p_M with
Step B4, utilize formula (4) to calculate the canonical correlation coefficient vector of two dimensions with λ c = [ λ c 0 , λ c 1 , . . . , λ c 359 ] T .
Step B5, calculating P 2d(as shown in Fig. 4 (b)).
(9) utilize SIFT operator to extract respectively the point of interest (as shown in Fig. 5 (a) and Fig. 5 (b)) of current p_M and historical p_M, and carry out images match (as shown in Fig. 6 (a) and Fig. 6 (b)), note matching rate is P p.
(10) calculate final matching rate P f=α p rpt+ β p 2d+ γ p p.In the present invention, test of many times result shows, the precision of three kinds of matching algorithms is just followed successively by: two-dimensional scan matching algorithm, and rotational invariance matching algorithm and image matching algorithm, so in this experiment, get α=0.33, β=0.37, γ=0.30.If P fdo not meet threshold value requirement, store current p_M, otherwise, can utilize match information to carry out the tasks such as localization for Mobile Robot, map building and path planning.

Claims (11)

1. an image conversion disposal route for sonar data, is characterized in that specifically comprising the following steps:
(1) filter due to sonar blind area or exceed the measured value that sonar to measure scope produces;
(2) by the method for cluster, filter singular value and the discrete point in sonar data;
(3) calculate the EVR curve of sonar to measure value, by detecting the mode of current EVR peak of curve, obtain the environment angle point information that sonar data comprises;
(4) error ellipse of each Corner Feature that calculating sonar data comprises;
(5) calculate the maximum error circle of each error ellipse, build the current local map M that expands;
(6) set up the mapping relations in the current local map M of expansion and bianry image space;
(7) M is mapped to bianry image space, map p_M is expanded in the part in design of graphics image space;
(8) utilize rotational invariance matching process and two-dimensional scan matching process to realize the local coupling of expanding map;
(9) utilize SIFT operator extraction point of interest, and carry out images match;
(10) be rotated the fusion of unchangeability matching process, two-dimensional scan matching process and image matching method matching result.
2. the method for claim 1, is characterized in that step (1) comprising: note sonar data is (x, y, θ, ra), (x wherein, y) be coordinates of targets, θ is the orientation of the relative robot of target, and ra is that target is to the distance of robot, due to the impact of measuring blind area or exceeding sonar to measure scope, can produce undesirable data, be labeled as ra=R, the maximum measuring distance that wherein R is sonar sensor, the point of removing all ra=R, the data set after filtration is designated as S.
3. the method for claim 1, it is characterized in that step (2) comprising: calculate in sonar data collection S Euclidean distance a little between any two, and according to the size of coordinate and distance, the point in S is classified, add up each classification and comprise number a little, when the number of certain classification mid point is less than threshold value Num, remove institute in respective class a little, left point forms metric space S 0.
4. the method for claim 1, is characterized in that step (3) comprising: definition length is N, and sliding step is s, and glide direction is moving window, edge direction, from S 0in get successively N point, its coordinate is designated as respectively X t=[x 1, x 2..., x n] and Y t=[y 1, y 2..., y n], by:
C t = cov ( X t , X t ) cov ( X t , Y t ) cov ( Y t , X t ) cov ( Y t , Y t )
Calculate X tand Y tcovariance matrix C t, wherein with be respectively X tand Y tthe average of element, note C tthe ratio EVR=λ of eigenwert min/ λ max, edge with step-length s moving window, be less than N to data in window, calculate current EVR curve, by the mode of more adjacent EVR value, calculate the peak value of current EVR curve, the corresponding angle point of each peak value.
5. the method for claim 1, is characterized in that the described error of calculation ellipse of step (4) specifically comprises following content:
Suppose that O is one of them peak value, the sonar data point that O is corresponding is (x, y), get and take (x, y) as the center of circle, and the data point of the n in the circle that r' is radius, its coordinate is designated as X=[x 1..., x n] and Y=[y 1..., y n], the covariance matrix C of calculating X and Y, the eigenwert of C is λ 1and λ 21>=λ 2), corresponding proper vector is v 1and v 2, error ellipse is v 1and v 2in the coordinate system forming with for the center of circle, λ 1, λ 2be respectively the ellipse of major axis and minor axis, wherein
6. the method for claim 1, is characterized in that the current local map M that expands of the described structure of step (5), specifically comprises following content:
With for the center of circle, λ 1for radius, build the maximum error circle that error ellipse is corresponding, according to the actual connection of angle point, calculate the public outer tangent line of the maximum error circle that the actual angle point that is connected is corresponding, further calculate the point of contact of tangent line and maximum error circle, the point of contact that connects the corresponding maximum error circle of the actual angle point that is connected forms the local map M that expands, and the center of circle of all maximum error circles, point of contact form the local crucial point set im_point that expands map M.
7. the method for claim 1, is characterized in that the described mapping relations of setting up M and bianry image space of step (6) comprise:
Horizontal ordinate and ordinate that note im_point comprises are a little respectively P=(p 1..., p m) and Q=(q 1..., q m), p maxand q maxbe respectively the maximal value of transverse and longitudinal coordinate, p minand q minbe respectively the minimum value of transverse and longitudinal coordinate, a bit (p in im_point i, q i) pass through formula h i = int [ η ( p max - p min ) · ( p i - p min ) p max - p min ] With k i = int [ η ( q max - q min ) · ( q i - q min ) q max - q min ] Map to a bit (h in image space i, k i), wherein, η is scale-up factor.
8. the method for claim 1, is characterized in that the described part in image space set up of step (7) expands map and comprise following content:
In note im_point, the picture in the center of circle and corresponding radius are respectively (h i, k i) and r i, i=1 ..., j, j is center of circle number, with (h i, k i) be the center of circle, r ifor radius, determine border circular areas, making the pixel value in the definite region of the picture at point of contact in this border circular areas and im_point is 0, otherwise is 1, remembers that region that in this bianry image space, pixel value is 0 is that map p_M is expanded in part in image space.
9. the method for claim 1, is characterized in that step (8) comprising:
(8.1) utilize the current p_M of rotational invariance matching method matches and historical p_M, specifically comprise following content:
(i) calculate the maximal value h of distance between p_M horizontal and vertical direction pixel maxand v max, further calculate the center point P of p_M c(int (h max/ 2), int (v max/ 2)), along horizontal and vertical direction, set up coordinate system Σ c, at coordinate system Σ cin, with P cfor the center of circle, as radius, be respectively the concentric circles of r and r+ Δ r, form donut R r, R rthe number of pixels comprising is designated as N r, the pixel count wherein being occupied by p_M is designated as U r, its ratio is designated as v (r)=U r/ N r, be called effective duty cycle corresponding to radius r;
(ii) calculate D=even ((h max/ 2) 2+ (v max/ 2) 2) 1/2, " even " represents ((h max/ 2) 2+ (v max/ 2) 2) 1/2upwards get even number, then calculate the effective duty cycle vector V of current and historical p_M land V d, and pass through calculate V land V dmatching rate p rpt, wherein, V l=[v l(0), v l(Δ r) ... v l(r) ... v l(D/2)] t, V d=[v d(0), v d(Δ r) ... v d(r) ... v d(D/2)] t, υ l = Σ r = 0 D / 2 v l ( r ) / ( 1 + D / 2 Δr ) , υ d = Σ r = 0 D / 2 v d ( r ) / ( 1 + D / 2 Δr ) ;
(8.2) utilize the current p_M of two-dimensional scan matching method matches and historical p_M, specifically comprise following content:
(i), according to the establishment of coordinate system method in (8.1), set up respectively the coordinate system of current p_M and historical p_M;
(ii) pixel count of remembering picture horizontal and vertical direction in current bianry image space is N, and the pixel count of the capable and i row of the i that occupied by p_M is N respectively riand N ci, the effective duty cycle of i row and column is designated as respectively w ri=N ri/ N and w ci=N ci/ N, the effective duty cycle that calculates respectively all ranks of current p_M forms effective duty cycle vector with , in like manner, calculate the effective duty cycle vector of historical p_M with
(iii) calculate respectively canonical correlation coefficient and the final matching degree P2d of current and historical p_M ranks effective duty cycle vector.
10. the method for claim 1, is characterized in that step (9) comprises following content: utilize SIFT operator to extract respectively the point of interest of current p_M and historical p_M, carry out images match, calculate matching rate P p.
11. the method for claim 1, is characterized in that step (10) comprising:
Calculate final matching rate P f=α p rpt+ β p 2d+ γ p p, alpha+beta+γ=1 wherein, α, the value of β and γ is determined according to the precision of each matching process, if P fdo not meet threshold value requirement, store current p_M, otherwise, can utilize match information to carry out localization for Mobile Robot, map building and path planning task.
CN201410210477.2A 2014-05-16 2014-05-16 A kind of image conversion processing method of sonar data Expired - Fee Related CN103983270B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410210477.2A CN103983270B (en) 2014-05-16 2014-05-16 A kind of image conversion processing method of sonar data

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410210477.2A CN103983270B (en) 2014-05-16 2014-05-16 A kind of image conversion processing method of sonar data

Publications (2)

Publication Number Publication Date
CN103983270A true CN103983270A (en) 2014-08-13
CN103983270B CN103983270B (en) 2016-09-28

Family

ID=51275334

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410210477.2A Expired - Fee Related CN103983270B (en) 2014-05-16 2014-05-16 A kind of image conversion processing method of sonar data

Country Status (1)

Country Link
CN (1) CN103983270B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106197421A (en) * 2016-06-24 2016-12-07 北京工业大学 A kind of forward position impact point for moving robot autonomous exploration generates method
CN108268427A (en) * 2018-01-10 2018-07-10 中国地质大学(武汉) A kind of free kick goal probability analysis method, equipment and storage device
CN108764080A (en) * 2018-05-17 2018-11-06 中国电子科技集团公司第五十四研究所 A kind of unmanned plane vision barrier-avoiding method based on cloud space binaryzation

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103247040A (en) * 2013-05-13 2013-08-14 北京工业大学 Layered topological structure based map splicing method for multi-robot system
CN103278170A (en) * 2013-05-16 2013-09-04 东南大学 Mobile robot cascading map building method based on remarkable scenic spot detection

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103247040A (en) * 2013-05-13 2013-08-14 北京工业大学 Layered topological structure based map splicing method for multi-robot system
CN103278170A (en) * 2013-05-16 2013-09-04 东南大学 Mobile robot cascading map building method based on remarkable scenic spot detection

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
CHOI J ET AL.: "Autonomous topological modeling of a home environment and topological localization using a sonar grid map", 《AUTONOMOUS ROBOTS》 *
CHOI Y-H ET AL.: "A line feature based SLAM with low grade range sensors using geometric constraints and active exploration for mobile robot", 《AUTONOMOUS ROBOTS》 *
李润伟: "基于声纳的室内环境栅格地图创建方法的研究", 《中国优秀硕士学位论文全文数据库信息科技辑》 *
高丽华等: "一种基于声纳信息的移动机器人地图创建方法", 《制造业自动化》 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106197421A (en) * 2016-06-24 2016-12-07 北京工业大学 A kind of forward position impact point for moving robot autonomous exploration generates method
CN106197421B (en) * 2016-06-24 2019-03-22 北京工业大学 A kind of forward position target point generation method independently explored for mobile robot
CN108268427A (en) * 2018-01-10 2018-07-10 中国地质大学(武汉) A kind of free kick goal probability analysis method, equipment and storage device
CN108764080A (en) * 2018-05-17 2018-11-06 中国电子科技集团公司第五十四研究所 A kind of unmanned plane vision barrier-avoiding method based on cloud space binaryzation
CN108764080B (en) * 2018-05-17 2021-10-01 中国电子科技集团公司第五十四研究所 Unmanned aerial vehicle visual obstacle avoidance method based on point cloud space binarization

Also Published As

Publication number Publication date
CN103983270B (en) 2016-09-28

Similar Documents

Publication Publication Date Title
CN103236064B (en) A kind of some cloud autoegistration method based on normal vector
CN103727930B (en) A kind of laser range finder based on edge matching and camera relative pose scaling method
CN104240297A (en) Rescue robot three-dimensional environment map real-time construction method
CN112184736B (en) Multi-plane extraction method based on European clustering
CN103295239A (en) Laser-point cloud data automatic registration method based on plane base images
CN103617328A (en) Airplane three-dimensional attitude computation method
CN112362072A (en) High-precision point cloud map creation system and method in complex urban area environment
CN114998276B (en) Robot dynamic obstacle real-time detection method based on three-dimensional point cloud
Qian et al. Robust visual-lidar simultaneous localization and mapping system for UAV
CN105137412A (en) Accurate fitting method of line segment features in 2D laser radar distance image
CN112949407A (en) Remote sensing image building vectorization method based on deep learning and point set optimization
CN113177593A (en) Fusion method of radar point cloud and image data in water traffic environment
CN103983270A (en) Graphic sonar data processing method
Aissou et al. Building roof superstructures classification from imbalanced and low density airborne LiDAR point cloud
CN102184536B (en) Method and system for extracting straight line and/or line segment end points from image
CN109583626B (en) Road network topology reconstruction method, medium and system
CN103700095A (en) Log end surface image partitioning algorithm for improving active contour model based on circle constraint
CN111275748A (en) Point cloud registration method based on laser radar in dynamic environment
CN113791400B (en) Stair parameter autonomous detection method based on laser radar
CN113554705B (en) Laser radar robust positioning method under changing scene
CN115453570A (en) Multi-feature fusion mining area dust filtering method
CN111932612B (en) Intelligent vehicle vision positioning method and device based on second-order hidden Markov model
CN111239761B (en) Method for indoor real-time establishment of two-dimensional map
Fang et al. ContinuityLearner: Geometric continuity feature learning for lane segmentation
CN113240721A (en) Navigation vanishing point detection method applied to complex scene of roadway

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20160928