CN101413806B - Mobile robot grating map creating method of real-time data fusion - Google Patents
Mobile robot grating map creating method of real-time data fusion Download PDFInfo
- Publication number
- CN101413806B CN101413806B CN2008101435378A CN200810143537A CN101413806B CN 101413806 B CN101413806 B CN 101413806B CN 2008101435378 A CN2008101435378 A CN 2008101435378A CN 200810143537 A CN200810143537 A CN 200810143537A CN 101413806 B CN101413806 B CN 101413806B
- Authority
- CN
- China
- Prior art keywords
- grid
- mobile robot
- probability
- value
- sonar
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
Images
Landscapes
- Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)
Abstract
The invention provides a building method of a grid map by a mobile robot based on real-time data fusion. The method comprises the following steps: artificially dividing an environment into a plurality of grids with the same size, and obtaining the distance information by sonar sensors arranged on the front end of the mobile robot; extracting measured values of three sonar sensors which are closest to a currently computed grid unit at the same time, respectively explaining a single sonar data by a fuzzy logic and a probability theory so as to obtain a group of characteristic vectors which are subject to the data fusion as input vectors of a neural network, wherein, output values of the neural network comprises idle state, busy state and undefined state of the grid; finally updating the state of the grid by the Bayes rule. The mobile robot of the invention obtains the environmental information by sonar ranging finders and completes the environmental modeling, thus providing reliable basis for the subsequent autonomous navigation of the mobile robot.
Description
Technical field
The invention belongs to robot and field of artificial intelligence, relate to the mobile robot grating map creating method that a kind of real time data merges.
Background technology
Over nearly 20 years, artificial intelligence technology and fast development of computer technology, autonomous intelligence mobile robot's research has obtained great concern.Its requirement can be obtained ambient condition information automatically by self contained sensor such as stadimeter, video camera, infrared etc., sets up space environment model and identification self current location, moves along the active path of planning automatically, thereby finishes particular task.Intelligent mobile robot is widely used in industries such as industrial or agricultural, communications and transportation, military affairs, health care at present, solves work problem and the human hard work of replacement under the hazardous environment.For improving ability to work and the range of application of mobile robot under circumstances not known, for being fit to the map of robot " understanding ", promptly map building is a crucial difficult problem of being badly in need of solution environment representation.
Mobile robot's working environment generally can be reduced to two dimensional model, and is described with taking grid.This model is divided into the several rules grid with robot place space environment, determines environment representation by the state (barrier, free time, the unknown) that extracts each grid cell.This model representation is directly perceived, is easy to create and safeguard.In circumstances not known, any prior imformation is not understood by robot, for example size of environment scale and barrier, shape, position etc., and do not exist in the environment such as the artificial object of reference of setting such as road sign, beacon.Map is created by robot can only depend on the information that its sensor obtains, as mileage gauge, sonar, laser or the like.Setting up the process of map, in fact is exactly robot according to the perception of sensor independently to the process of its environment of an activation modeling.Because the restriction of sensor self, there is uncertainty in various degree in perception information, need handle perception information usually again.Information processing in the map building can reduce following three problems:
(1) how the uncertainty of perception information is described;
(2) how to create map, not only will reflect perception information in the map, also will reflect the uncertainty of information according to uncertainty description to information;
(3) when same objective has been had new perception information, how to handle old information and fresh information, promptly upgrade map.
In Mobile Robotics Navigation, sonar sensor is owing to its cheapness, be simple and easy to characteristics such as usefulness, convenient data processing, therefore obtain widespread use in the mobile robot field, but also there are defectives such as multipath reflection, direct reflection, angular accuracy are low in sonar sensor itself, and there is bigger uncertainty in its perception information.In the research of map building, fuzzy logic and probability theory are two kinds of representative methods that are used to describe and handle uncertain information at present.The former is uncertain by sky, non-dummy section and state in the ambiguity in definition set representations environment, and each grid cell is calculated corresponding degree of membership according to measurement result; The latter provides the probability that it occupies for barrier to each grid cell, carries out information fusion according to bayes rule.Analysis shows that with experiment higher based on the accuracy of map that probability method produces, profile is clear, but too responsive to error message, the False Rate height; Fuzzy logic method then has higher robustness, and is still stable when the uncertain degree of information is high, but precision is low, and the map of generation is unintelligible.
Summary of the invention
Technical matters to be solved of the present invention provides the mobile robot grating map creating method that a kind of real time data merges, mobile robot of the present invention utilizes the sonar ranging instrument to obtain environmental information, finish environmental modeling, thereby provide reliable foundation for the follow-up independent navigation of mobile robot.
For solving the problems of the technologies described above, the technical solution adopted in the present invention is:
The mobile robot grating map creating method that a kind of real time data merges is characterized in that, may further comprise the steps:
1) with local grid map coordinate (i, j) initialization, and barrier that obtains by a plurality of sensors on the mobile robot and the distance value between the mobile robot;
2) 3 distance values with the numerical value minimum in the described distance value are arranged as one-dimension array with fuzzy logic and probability theory explanation; { O, E, U} represent the taking of grid in the map, free time and unknown state promptly each distance value to be set up 3 fuzzy sets; Ambiguity in definition vector T={ μ
O, μ
E, μ
URepresent that each grid is in the confidence level of three state, grid belongs to the degree of membership of three state and is 1;
μ
o(r, s), μ
E(r is that (i j) takies membership function and idle membership function apart from sensor distance r at the grid of zones of different to grid g s); μ
O, E(α, the β) membership function of the different beam axis line angle of expression, μ
O, E(r) membership function of expression different measuring distance;
The occupied membership function μ of each grid cell
o(r, α), the idle membership function μ of grid
E(r, α) and the uncertain membership function μ of trellis states could
U(r, α) by following calculating gained:
μ
o(r,α)=M
0[μ
o(r,s)+μ
o,E(α,β)+μ
o,E(r)]/3;
μ
E(r,α)=[μ
E(r,s)+μ
O,E(α,β)+μ
O,E(r)]/3;
μ
U(r,α)=1-μ
O(r,α)-μ
E(r,α);
Wherein, M
0The occupied maximum likelihood of expression grid cell, value is less than 1; Wherein s is the sensor distance measured value, and Δ s is the estimation of error scope to obstacle distance S, and α is the angle of r with respect to the beam axis line; β represents the tapered width half-angle of wave beam; R represents the maximum detectable range of sensor;
Grid g (i, j) be occupied or idle probability is used P (taking) and P (free time) expression respectively, computing method are as follows:
P (taking)=1-P (free time);
Corresponding 5 the element { μ of each distance value
o(r, α), μ
E(r, α), μ
U(r, α), P (taking); P (free time) }, for 3 distance values, then described one-dimension array has 15 elements.
3) with described one-dimension array through Processing with Neural Network, be output as conditional probability vector O=[O
Occ, O
Emp, O
Uce], represent respectively grid cell g (x, y) corresponding three kinds may states the probable value of (barrier, dummy section, uncertain).
4) according to coordinate transform, with grid cell g (i, j) from local coordinate system project to the global coordinate system correspondence grid cell g (x, y); Utilize the Bayesian probability method to realize that respectively the probability of three kinds of possibility states upgrades then, according to the maximal value rule, the maximal value of getting the possibility state is the degree of confidence of current grid.
5) coordinate is updated to next grid cell, if trellis states could has not been upgraded, returns step 1); If trellis states could has been upgraded, then grating map creating is finished.
Described neural network interpretation model comprises three layers altogether: input layer, hidden layer and output layer; Wherein the input vector of input layer is described one-dimension array; The transport function of described hidden layer neuron adopts S type tan; The transport function that described output layer adopts is a S type logarithmic function, and node is output as the probable value of [0,1] scope.
Described Bayesian probability method is:
For grid cell m
I, jThree kinds may states initial probability be 1/3, then pass through measurement data sequence S=(s
(1)..., s
(T)) after, s wherein
(n)Represent three and the maximally related sensor measurement value sequence of current computation grid, corresponding integrated probability is as follows respectively:
In the formula, P (takies
t), the P (free time
t), P is (uncertain
t) be the final trellis states could value behind consideration historical information and the current information, O
Occ, O
Emp, O
UceBe current neural Network Data Fusion output, P (takies
T-1), the P (free time
T-1), P is (uncertain
T-1) be last one constantly final trellis states could value, i.e. prior probability, when t=1, P (takies
T-1), the P (free time
T-1), P is (uncertain
T-1) value be the initial probability 1/3 of agreement;
At last P (is taken
t), the P (free time
t), P is (uncertain
t) get maximal value, be the degree of confidence of current grid.
Described sensor is 8.
The position with respect to the robot forward direction of 8 sensors is: ± 10 °, ± 30 °, ± 50 ° and ± 90 °.
As improvement, get M
0=0.95.
The beneficial effect that the present invention had:
The present invention proposes the mobile robot grating map creating method that a kind of real time data merges, and is used to describe the sky and the non-dummy section of mobile robot's environment of living in.At first artificially environment is divided into the grid of several sizes such as grade, obtains range information by the sonar sensor that is installed in mobile robot's front end; Extract and the measured value of nearest three sonar sensors of current computation grid cell distance at synchronization, adopt fuzzy logic and probability theory to explain single sonar data respectively, promptly represent the taking of certain grid, empty and uncertain with fuzzy logic, represent to occur in certain grid probability and the empty probability that barrier occupies with probability theory according to range information.Can obtain an eigenvectors thus, and carry out data fusion as the input vector of neural network, the output valve of neural network is grid free time, occupied or nondeterministic statement.Adopt Bayes rule to upgrade the state of grid at last.Mobile robot of the present invention utilizes the sonar ranging instrument to obtain environmental information, finishes environmental modeling, thereby provides reliable foundation for the follow-up independent navigation of mobile robot.
Compared with prior art, advantage of the present invention just is:
1. because the training of neural network based on sample, has adaptivity soon to new environment;
2. fuzzy logic, probability theory are explained that respectively the result of sonar data merges, effectively combine fuzzy logic interpretation model robustness height, advantage that probability theory interpretation model precision is high;
3. can handle a plurality of sensor informations simultaneously, consider the influence of a plurality of sensors to the measurement result of same grid, associating ambient sensors information can generate more accurate result, can solve direct reflection and other uncertain problem well.
Description of drawings
Fig. 1 grating map creating method schematic diagram of the present invention;
The configuration of Fig. 2 mobile robot forward direction sonar ring;
Fig. 3 sonar sensor model;
Fig. 4 neural Network Data Fusion is created grating map;
Fig. 5 training sample environment;
Fig. 6 takies grid and obtains the sample sample.
Embodiment
The invention will be further described below in conjunction with accompanying drawing.
Embodiment 1:
The present invention proposes the mobile robot grating map creating method that a kind of real time data merges, and its systematic schematic diagram as shown in Figure 1.The sonar data uncertainty that adopts neural network that probability theory, fuzzy logic are explained is carried out information fusion, and has considered the influence of the spatial coherence of a plurality of sonars to same trellis states could, to set up the environment grating map.
S0, S1 among Fig. 1 ..., S7 is the sonar sensor distance measurement value that is installed in mobile robot's front end, O
Emp{ x, y}, O
Occ{ x, y}, O
Uce{ x, y} are respectively the idle probability of grid, barrier probability and the nondeterministic statement probability of current calculating.
The mobile robot adopts the ultrasonic ranging sensor to finish environmental modeling, and its front end is equipped with eight sonar ranging sensors.The position of mobile robot's sonar sensor is among Fig. 2: ± 10 °, ± 30 °, ± 50 ° and ± 90 °, be used to survey the information of barrier on the direction separately.
When the mobile robot explores in the ergodic process at environment, once gather sonar sensor measurement data sequence and corresponding robot global positional information.Be that local reference frame is set up at the center with the machine people then, create local map based on the sonar sensor measurement data.In local map building, a certain particular grid unit cell (i to local map, j), three at first relevant continuous sonar sensors according to the spatial positional information choice direction of corresponding this geometric units correspondence, use fuzzy logic (fuzzy set { take, sky, uncertain } is described) and probability theory (probability { barrier respectively with three sonar sensor ranging information, empty } describe) explain and can obtain 15 decryptions, with this input as neural network.Through (as shown in Figure 4) after the interpretation process of neural network, its corresponding neural network is output as conditional probability vector O=[O
Occ, O
Emp, O
Uce], represent respectively grid cell g (x, y) corresponding three kinds may states the probable value of (barrier, dummy section, uncertain).On this basis, according to coordinate transform, with grid cell g (i, j) from local coordinate system project to global coordinate system to deserved grid cell g (x, y).Utilize the Bayesian probability model to realize that respectively the probability of three kinds of possibility states upgrades then.
Sonar sensor model, fuzzy logic explain that sonar uncertainty models, probability theory explanation sonar uncertainty models, neural network fusion establishment grating map algorithm and Bayes upgrade grid and set forth respectively in 1,2,3,4,5 trifles.
1, sonar sensor model
Illustrate: single Shu Shengna basic model visual field determines that by β and R β represents the tapered width half-angle, and R represents maximum detectable range.The visual field can project on the regular net, is free time or occupied information because each unit of grid all records corresponding locus, takies grid so be called, and also claims to take grid.As shown in Figure 3, the visual field can be divided into three zones.
Area I: coherent element may be occupied;
Area I I: coherent element may be empty;
Area I II: coherent element situation the unknown;
For given range reading, " free time " of area I I has bigger possibility than area I " occupied ".No matter be " free time " or " occupied ", than more accurate towards the data of two edge directions, partly cause is that a limit along barrier may produce direct reflection or cause other apart from the perception mistake acoustic beam along the data of sound wave axis direction.
Though the sensor model among Fig. 3 is a kind of general choice, but how is model conversion to be very different aspect the confidence data, introduce fuzzy logic, the probability theory transfer process of being correlated with below respectively, be convenient to follow-up neural network and carry out data fusion.
2, fuzzy logic is explained the sonar uncertainty models
Illustrate: fuzzy logic explains that the basic thought of sonar uncertainty models is to set up 3 fuzzy sets { O, E, U} represent the taking of all grids in the map, free time and unknown state.Ambiguity in definition vector T={ μ
O, μ
E, μ
URepresent that each grid is in the reliability of three state, grid belongs to the degree of membership of three state and is 1.
Algorithm:
In Fig. 3, establishing the sonar range measured value is s, and Δ s is the estimation of error scope to obstacle distance s.(i j) represents with r to the distance of sensor, is α with respect to the angle of beam axis line for the arbitrary grid cell g in the beam coverage.In following formula (1)~(4), μ
o(r, s), μ
E(r is s) for to take membership function and idle membership function apart from sensor distance r at the grid of zones of different.μ
O, E(α, the β) membership function of the different beam axis line angle of expression, μ
O, E(r) membership function of expression different measuring distance.
The occupied membership function μ of each grid cell
o(r, α), the idle membership function μ of grid
E(r, α) and the uncertain membership function μ of trellis states could
U(r α) can calculate gained by formula (5)~(7).
μ
o(r,α)=M
0[μ
o(r,s)+μ
o,E(α,β)+μ
o,E(r)]/3 (5)
μ
E(r,α)=[μ
E(r,s)+μ
o,E(α,β)+μ
O,E(r)]/3 (6)
μ
U(r,α)=1-μ
O(r,α)-μ
E(r,α) (7)
M in the formula (5)
0The occupied maximum likelihood of expression grid cell because the occupied possibility of grid can not be 100%, is therefore got M
0=0.95.
3, probability theory is explained the sonar uncertainty models
Illustrate: represent the occupied or idle condition of grid with probability function.
Algorithm:
Sonar can only be observed an incident: (i is occupied or idle j) to element g.This can be written as H={ and takies, the free time }.The true probability that takes place of H incident is represented with P (H):
0≤P(H)≤1
A critical nature of probability is: if know P (H), so H do not have probability of happening P (just known also that H) this can be expressed as:
1-P(H)=P(-H)
(H) probability of form is called unconditional probability, a prior imformation only is provided, and does not consider sensor reading s for P (H) and P.Concerning robot, can according to sensor reading zoning g (i, j) function idle or occupied probability is more useful, this probability is conditional probability.When P (H|s) is exactly the concrete reading s of given sensor, the probability of the actual generation of H incident.Conditional probability also has such character: P (H|s)+P (H|s)=1.
For each grid cell in Fig. 3 area I:
P (free time)=1-P (taking) (9)
The implication of r, α and fuzzy logic are explained consistent in the sonar uncertainty models, M
1Represent that occupied unit reading makes never that to take degree of confidence be 100%, gets M
1=0.98.
For each grid cell among Fig. 3 area I I:
P (taking)=1-P (free time) (10)
Different with the area I grid cell, the idle probability of area I I element can reach 1.
4, neural Network Data Fusion is created grating map
Illustrate: in grating map creating, the measurement data of sonar sensor must be explained and be mapped as relevant position unit g (x, degree of confidence y).Yet there are problems such as multipath reflection, direct reflection, angular accuracy be low in sonar sensor, is difficult to set up precise math model and is used to explain sonar data.Because the multilayer neural network after the training can approach any probability distribution, so can utilize the neural network after the training to realize of the mapping of sonar to measure data to the grid probability.
Algorithm:
1) neural network structure
As shown in Figure 4, the neural network interpretation model that proposes of the present invention comprises three layers altogether: input layer, hidden layer and output layer.Below go through the design and the realization of each layer.
(1) input layer
The input vector of neural network comprises 15 elements.For given grid cell g (i, j), select with robot center and grid cell centerline direction maximally related about totally three 3 sonar to measure data that the sonar sensor synchronization gets access to.Then, receiving the fuzzy logic interpretation model of sensor according to monophone calculates each sonar to measure data (totally three groups of fuzzy sets is represented the state of grid, i.e. T for i, explanation j) to grid cell g
1={ μ
O1, μ
E1, μ
U1,
T
2={ μ
O2, μ
E2, μ
U3, T
3={ μ
O3, μ
E3, μ
U3; Receiving the probability theory interpretation model of sensor according to monophone calculates each sonar to measure data (totally three groups of probable values is represented the state of grid, i.e. P for i, state j) to grid cell g
1={ P
O1, P
E1, P
2={ P
O2, P
E2, P
3={ P
O3, P
E3.With these six groups of data totally 15 elements as the input of neural network, as follows:
ρ={μ
O1,μ
E1,μ
U1,μ
O2,μ
E2,μ
U2,μ
O3,μ
E3,μ
U3,P
O1,P
E1,P
O2,P
E2P
O3P
E3}
T
(2) hidden layer
The neuron number of hidden layer is designed to 31, and the neuron number of hidden layer is not what fix, adjusts as required in hands-on.The transport function of hidden layer neuron adopts S type tan.
(3) output layer
Output layer has 3 nodes, and output valve is respectively O=[O
Occ, O
Emp, O
Uce], three kinds of expression grid cell correspondence may state.Wherein first output node represents that grid cell is that the probable value of seizure condition is O
Occ, second output node represents that grid cell is that the probable value of idle condition is O
Emp, the 3rd output node represents that grid cell is that the probable value of nondeterministic statement is O
UceThe transport function that this network layer adopted is a S type logarithmic function, and node output still is the probable value of [0,1] scope.
2) neural metwork training
Neural metwork training is the very important step of neural network, in case network training is finished, robustness and adaptivity owing to neural network self promptly can be used in the multiple different environment.Below introduce in detail and explain related training sample and the training algorithm of network about sonar.
(1) training sample
According to neural network model shown in Figure 4, the form of training data sample is as follows:
<μ
O1,μ
E1,μ
U1,μ
O2,μ
E2,μ
U2,μ
O3,μ
E3,μ
U3,P
O1,P
E1,P
O2,P
E2,P
O3,P
E3,O
occ,O
emp,O
uce>
Wherein<μ
O1, μ
E1, μ
U1,<μ
O2, μ
E2, μ
U2,<μ
O3, μ
E3, μ
U3Represent respectively robot center and grid cell centerline direction maximally related about totally three 3 sonar to measure data that the sonar sensor synchronization obtains, receive the fuzzy logic interpretation model of sensor according to monophone and calculate each sonar to measure data grid cell g (i, explanation j);<P
O1, P
E1,<P
O2, P
E2,<P
O3, P
E3Explain the explanation of 3 sonar to measure data that above-mentioned three sensor synchronizations got access to probability theory respectively; Grid cell g (i, desired output usefulness<O j)
Occ, O
Emp, O
UceExpression, correspond respectively to take, the output of idle and nondeterministic statement, its possible output valve be [1,0,0], [0,1,0], [0,0,1].
The gatherer process of training sample is as follows.Robot is put into (being robot location, peripheral obstacle location aware) under the known indoor environment, and by the robot rectilinear motion, rotatablely move, continuous several times is gathered robot pose and sonar sensor measurement data at random, and<O
Occ, O
Emp, O
UceCalculate gained cell (i, the corresponding output of real space state j) for considering grid cell and barrier spatial information.
Fig. 5 is the hands-on environment representation.The sonar ranging scope that adopts is R=300cm, field angle β=15 °, and tolerance is 15cm, grid size is decided to be 10cm * 10cm.
Below by obtaining of example explanation sample.
Take the zone that grid has covered one 30 * 24 unit as shown in Figure 6, also just can be expressed as one 30 * 24 two-dimensional array, consider particular grid g[7] [11] (among figure shown in the bullet), the robot location is at g[27] [11] (among figure shown in the black point), i.e. r=200cm.Synchronization obtains the measurement data of 8 sonars, with particular grid g[7] the maximally related sonar data in [11] position for be positioned among Fig. 2 ± 10 °, 30 ° sonar data (also can replace the position at 30 ° sonar fetched data-30 ° sonar data with the position because this moment grid and this two sensors equidistant).
● for the position is 10 ° sonar, and resultant range information is s=210cm, α=-10 °, wherein | α |≤| β |, r≤s ± tolerance, grid cell is in the zone of sonar model angle, in the range reading upper range, belong in the measured value coverage simultaneously.
Applying mechanically formula (1)~(7), can to obtain this trellis states could that this sonar data is represented with fuzzy logic be T
1=0.367,0.475,0.158}; Because be in area I, applying mechanically this trellis states could of representing with probability theory formula (8)~(9) is P
1=0.327,0.673}.
● for the position is-10 ° sonar, and resultant range information is 205cm, α=10 °, wherein | α |≤| β |, r≤s ± tolerance, grid cell in the range reading upper range, belong in the measured value coverage in the zone of sonar model angle simultaneously.Applying mechanically formula (1)~(7), can to obtain this trellis states could that this sonar data is represented with fuzzy logic be T
2=0.475,0.367,0.158}; Because be in area I, applying mechanically this trellis states could of representing with probability theory formula (8)~(9) is P
2=0.327,0.673}.
● for the position is 30 ° sonar, and resultant range information is 110cm, α=-30 °.Wherein | α | | β |, r<s+ tolerance, though in the range reading upper range, grid cell is outside the zone of sonar model angle, so do not belong in the measured value coverage, the fuzzy logic of this trellis state represent with to(for) this sonar data is T
3=0,0, and 1}, this trellis states could of representing with probability theory is P
3=0,1}.
● this grid is for taking, so O=[1 is arranged, 0,0 in the reality].
Obtain a sample data like this, by T
1, T
2, T
3, P
1, P
2, P
3, O forms, and is as follows:
<0.367,0.475,0.158,0.475,0.367,0.158.0,0,1,0.327,0.673,0.327,0.673,0,1,1,0,0〉all the other samples can produce by same procedure.
(2) neural network BP training algorithm
In this paper neural metwork training, adopt classical Levenberg-Marquardt algorithm that neural network is trained, learning rate is 0.01, selects for use general approximate mean square deviation function as performance index function, when error is 2.5 * 10
-5Shi Xunlian finishes.
5, Bayes upgrades grid
After neural metwork training finished, the mobile robot walked traversal along the barrier edge lines in space environment, obtain space environment information, and utilize neural network that the sensor array of being gathered is made an explanation.Therefore to same grid cell, may there be different a plurality of explanations constantly.For obtaining to explain more accurately, need carry out integrated to these data.For avoiding complexity of calculation, guarantee that the enhancement formula of map building algorithm is handled, integration mode still adopts Bayesian (Bayes) integrated model.
Because this neural network model is exported the probability about three kinds of states of grid cell simultaneously, therefore in the map building process, carry out integrated to three kinds of state probability historical datas of grid cell respectively.For grid cell m
I, jThree kinds may states initial probability be 1/3, then pass through measurement data sequence S=(s
(1)..., s
(T))Back (s wherein
(n)Represent three and the maximally related sensor measurement value sequence of current computation grid), corresponding integrated probability is as follows respectively:
In the formula, P (takies
t), the P (free time
t), P is (uncertain
t) consider the final trellis states could value behind historical information and the current information, O
Occ, O
Emp, O
UceBe current neural Network Data Fusion output, P (takies
T-1), the P (free time
T-1), P is (uncertain
T-1) be last one constantly final trellis states could value, promptly prior probability (calculates) when t=1 for the first time, and P (takies
T-1), the P (free time
T-1), P is (uncertain
T-1) value be the initial probability 1/3 of agreement.
At last P (is taken
t), the P (free time
t), P is (uncertain
t) get maximal value, be the degree of confidence of current grid.
Claims (5)
1. the mobile robot grating map creating method that real time data merges is characterized in that, may further comprise the steps:
1) with local grid map coordinate (i, j) initialization, and barrier that obtains by a plurality of sensors on the mobile robot and the distance value between the mobile robot;
2) 3 distance values with the numerical value minimum in the described distance value are arranged as one-dimension array with fuzzy logic and probability theory explanation; { O, E, U} represent the taking of grid in the map, free time and unknown state promptly each distance value to be set up 3 fuzzy sets; Ambiguity in definition vector T={ μ
O, μ
E, μ
URepresent that each grid is in the confidence level of three state, grid belongs to the degree of membership of three state and is 1;
μ
o(r, s), μ
E(r is that (i j) takies membership function and idle membership function apart from sensor distance r at the grid of zones of different to grid g s); μ
O, E(α, the β) membership function of the different beam axis line angle of expression, μ
O, E(r) membership function of expression different measuring distance;
The occupied membership function μ of each grid cell
o(r, α), the idle membership function μ of grid
E(r, α) and the uncertain membership function μ of trellis states could
U(r, α) by following calculating gained:
μ
0(r,α)=M
0[μ
0(r,s)+μ
0,E(α,β)+μ
0,E(r)]/3;
μ
E(r,α)=[μ
E(r,s)+μ
O,E(α,β)+μ
O,E(r)]/3;
μ
U(r,α)=1-μ
O(r,α)-μ
E(r,α);
Wherein, M
0The occupied maximum likelihood of expression grid cell, value is less than 1; Wherein s is the sensor distance measured value, and Δ s is the estimation of error scope to obstacle distance S, and α is the angle of r with respect to the beam axis line; β represents the tapered width half-angle of wave beam; R represents the maximum detectable range of sensor;
Grid g (i, j) be occupied or idle probability is used P (taking) and P (free time) expression respectively, computing method are as follows:
P (taking)=1-P (free time);
Corresponding 5 the element { μ of each distance value
o(r, α), μ
E(r, α), μ
U(r, α), P (taking); P (free time) }, for 3 distance values, then described one-dimension array has 15 elements;
3) with described one-dimension array through Processing with Neural Network, be output as conditional probability vector O=[O
Occ, O
Emp, O
Uce], represent grid cell g (i, j) probable value of corresponding barrier, dummy section, uncertain these three kinds possibility states respectively;
4) according to coordinate transform, with grid cell g (i, j) from local coordinate system project to the global coordinate system correspondence grid cell g (x, y); Utilize the Bayesian probability method to realize that respectively the probability of three kinds of possibility states upgrades then, according to the maximal value rule, the maximal value of getting the possibility state is the degree of confidence of current grid;
5) coordinate is updated to next grid cell, if trellis states could has not been upgraded, returns step 1); If trellis states could has been upgraded, then grating map creating is finished;
Described neural network comprises three layers altogether: input layer, hidden layer and output layer; Wherein the input vector of input layer is an one-dimension array; The transport function of hidden layer neuron adopts S type tan; The transport function that output layer adopts is a S type logarithmic function, and node is output as the probable value of [0,1] scope.
2. the mobile robot grating map creating method that real time data according to claim 1 merges is characterized in that described Bayesian probability method is:
For grid cell m
I, jThree kinds may states initial probability be 1/3, then pass through measurement data sequence S=(s
(1)..., s
(T)) after, s wherein
(n)Represent three and the maximally related sensor measurement value sequence of current computation grid, corresponding integrated probability is as follows respectively:
In the formula, P (takies
t), the P (free time
t), P is (uncertain
t) be the final trellis states could value behind consideration historical information and the current information, O
Occ, O
Emp, O
UceBe current neural Network Data Fusion output, P (takies
T-1), the P (free time
T-1), P is (uncertain
T-1) be last one constantly final trellis states could value, i.e. prior probability, when t=1, P (takies
T-1), the P (free time
T-1), P is (uncertain
T-1) value be the initial probability 1/3 of agreement;
At last P (is taken
t), the P (free time
t), P is (uncertain
t) get maximal value, be the degree of confidence of current grid.
3. the mobile robot grating map creating method that real time data according to claim 1 merges is characterized in that described sensor is 8.
4. the mobile robot grating map creating method that real time data according to claim 3 merges is characterized in that the position with respect to the robot forward direction of described 8 sensors is: ± 10 °, ± 30 °, ± 50 ° and ± 90 °.
5. the mobile robot grating map creating method that merges according to each described real time data of claim 1 to 4 is characterized in that, gets M
0=0.95.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2008101435378A CN101413806B (en) | 2008-11-07 | 2008-11-07 | Mobile robot grating map creating method of real-time data fusion |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2008101435378A CN101413806B (en) | 2008-11-07 | 2008-11-07 | Mobile robot grating map creating method of real-time data fusion |
Publications (2)
Publication Number | Publication Date |
---|---|
CN101413806A CN101413806A (en) | 2009-04-22 |
CN101413806B true CN101413806B (en) | 2011-05-25 |
Family
ID=40594447
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN2008101435378A Expired - Fee Related CN101413806B (en) | 2008-11-07 | 2008-11-07 | Mobile robot grating map creating method of real-time data fusion |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN101413806B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110651315A (en) * | 2017-06-01 | 2020-01-03 | 三菱电机株式会社 | Map processing device, map processing method, and map processing program |
Families Citing this family (43)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101650568B (en) * | 2009-09-04 | 2010-12-29 | 湖南大学 | Method for ensuring navigation safety of mobile robots in unknown environments |
CN101907461B (en) * | 2010-06-24 | 2012-07-04 | 西安电子科技大学 | Measuration data correlation method for passive multisensor based on angle cotangent value |
CN101944240B (en) * | 2010-08-20 | 2012-02-15 | 浙江大学 | Fusion method of multi-robot three-dimensional geometrical map |
CN102156476B (en) * | 2011-04-14 | 2013-12-18 | 山东大学 | Intelligent space and nurse robot multi-sensor system and information fusion method of intelligent space and nurse robot multi-sensor system |
CN102521653B (en) * | 2011-11-23 | 2015-01-21 | 河海大学常州校区 | Biostimulation neural network device and method for jointly rescuing by multiple underground robots |
CN102915465B (en) * | 2012-10-24 | 2015-01-21 | 河海大学常州校区 | Multi-robot combined team-organizing method based on mobile biostimulation nerve network |
CN103472823B (en) * | 2013-08-20 | 2015-11-18 | 苏州两江科技有限公司 | A kind of grating map creating method of intelligent robot |
US9802309B2 (en) * | 2014-07-24 | 2017-10-31 | X Development Llc | Methods and systems for generating instructions for a robotic system to carry out a task |
CN105511485B (en) * | 2014-09-25 | 2018-07-06 | 科沃斯机器人股份有限公司 | For the creation method of self-movement robot grating map |
CN104268422A (en) * | 2014-10-10 | 2015-01-07 | 浪潮集团有限公司 | Ultrasonic range sensor array error correction method adopting k-nearest neighbors method |
US10093021B2 (en) * | 2015-12-02 | 2018-10-09 | Qualcomm Incorporated | Simultaneous mapping and planning by a robot |
CN105573318B (en) * | 2015-12-15 | 2018-06-12 | 中国北方车辆研究所 | environment construction method based on probability analysis |
CN105425803B (en) * | 2015-12-16 | 2020-05-19 | 纳恩博(北京)科技有限公司 | Autonomous obstacle avoidance method, device and system |
CN105571596B (en) * | 2016-01-18 | 2018-09-18 | 福州华鹰重工机械有限公司 | More vehicle environmental heuristic approach and device |
CN107037806B (en) * | 2016-02-04 | 2020-11-27 | 科沃斯机器人股份有限公司 | Self-moving robot repositioning method and self-moving robot adopting same |
CN105955272B (en) * | 2016-05-23 | 2019-07-26 | 上海钛米机器人科技有限公司 | The fusion method of the more time-of-flight sensors of service robot |
US9996944B2 (en) * | 2016-07-06 | 2018-06-12 | Qualcomm Incorporated | Systems and methods for mapping an environment |
CN108240807B (en) * | 2016-12-27 | 2023-06-02 | 法法汽车(中国)有限公司 | Method for estimating space occupation |
CN108481321B (en) * | 2017-01-09 | 2020-07-28 | 广东宝乐机器人股份有限公司 | Robot movement control method and robot |
CN107024210A (en) * | 2017-03-09 | 2017-08-08 | 深圳市朗空亿科科技有限公司 | A kind of Indoor Robot barrier-avoiding method, device and navigation system |
CN109421056A (en) * | 2017-08-25 | 2019-03-05 | 科沃斯机器人股份有限公司 | Self-movement robot |
CN108459599B (en) * | 2017-12-21 | 2020-08-07 | 华为技术有限公司 | Motion path planning method and device |
CN108663041B (en) * | 2018-02-09 | 2020-04-24 | 意诺科技有限公司 | Method and device for drawing navigation map |
CN108709562B (en) * | 2018-04-28 | 2020-07-03 | 北京机械设备研究所 | Method for constructing rolling grid map of mobile robot |
CN108802741B (en) * | 2018-06-22 | 2022-05-17 | 西安电子科技大学 | Mobile robot sonar data fusion method based on DSmT theory |
CN108919300B (en) * | 2018-07-17 | 2022-07-08 | 重庆大学 | Mixed map creating method for warehouse channel scene |
CN108983777B (en) * | 2018-07-23 | 2021-04-06 | 浙江工业大学 | Autonomous exploration and obstacle avoidance method based on self-adaptive front exploration target point selection |
CN111382768B (en) | 2018-12-29 | 2023-11-14 | 华为技术有限公司 | Multi-sensor data fusion method and device |
CN109855634B (en) * | 2019-01-22 | 2021-02-12 | 上海岚豹智能科技有限公司 | Image processing method and device for grid map |
JP7439822B2 (en) * | 2019-03-06 | 2024-02-28 | ソニーグループ株式会社 | Map creation device, map creation method, and program |
CN109975816B (en) * | 2019-03-11 | 2023-03-24 | 武汉理工大学 | Sensor information fusion method of miniature underwater robot |
CN109991605A (en) * | 2019-04-03 | 2019-07-09 | 吉林大学 | A kind of multi-radar data fusion method of intelligent network connection automotive environment sensory perceptual system |
EP3739361A1 (en) * | 2019-05-13 | 2020-11-18 | Aptiv Technologies Limited | Method and system for fusing occupancy maps |
EP3745299A1 (en) * | 2019-05-31 | 2020-12-02 | Infineon Technologies AG | Neural network device and method using a neural network for sensor fusion |
CN110703747B (en) * | 2019-10-09 | 2021-08-03 | 武汉大学 | Robot autonomous exploration method based on simplified generalized Voronoi diagram |
CN113865598A (en) * | 2020-06-30 | 2021-12-31 | 华为技术有限公司 | Positioning map generation method, positioning method and positioning device |
CN113093221A (en) * | 2021-03-31 | 2021-07-09 | 东软睿驰汽车技术(沈阳)有限公司 | Generation method and device of grid-occupied map |
CN113866758B (en) * | 2021-10-08 | 2023-05-26 | 深圳清航智行科技有限公司 | Scene monitoring method, system, device and readable storage medium |
CN114264780A (en) * | 2021-12-27 | 2022-04-01 | 四川天奥空天信息技术有限公司 | Air-space-ground integration-based greenhouse gas monitoring method and system |
CN114967728B (en) * | 2022-02-25 | 2023-03-31 | 珠海紫燕无人飞行器有限公司 | Unmanned aerial vehicle cluster collaborative search method |
CN115016464B (en) * | 2022-05-13 | 2024-09-27 | 中山大学·深圳 | Micro-nano robot path planning and automatic navigation method and system |
CN115290098B (en) * | 2022-09-30 | 2022-12-23 | 成都朴为科技有限公司 | Robot positioning method and system based on variable step length |
CN115407344B (en) * | 2022-11-01 | 2023-01-17 | 小米汽车科技有限公司 | Grid map creation method, device, vehicle and readable storage medium |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101033971A (en) * | 2007-02-09 | 2007-09-12 | 中国科学院合肥物质科学研究院 | Mobile robot map creation system and map creation method |
CN101093503A (en) * | 2006-06-20 | 2007-12-26 | 三星电子株式会社 | Method, apparatus, and medium for building grid map in mobile robot and method, apparatus, and medium for cell decomposition that uses grid map |
-
2008
- 2008-11-07 CN CN2008101435378A patent/CN101413806B/en not_active Expired - Fee Related
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101093503A (en) * | 2006-06-20 | 2007-12-26 | 三星电子株式会社 | Method, apparatus, and medium for building grid map in mobile robot and method, apparatus, and medium for cell decomposition that uses grid map |
CN101033971A (en) * | 2007-02-09 | 2007-09-12 | 中国科学院合肥物质科学研究院 | Mobile robot map creation system and map creation method |
Non-Patent Citations (1)
Title |
---|
余洪山.移动机器人地图创建和自主探索方法研究.《中国博士学位论文全文数据库 信息科技辑》.2008,(第8期),正文第60-65页. * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110651315A (en) * | 2017-06-01 | 2020-01-03 | 三菱电机株式会社 | Map processing device, map processing method, and map processing program |
CN110651315B (en) * | 2017-06-01 | 2021-09-07 | 三菱电机株式会社 | Map processing device, map processing method, and computer-readable storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN101413806A (en) | 2009-04-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN101413806B (en) | Mobile robot grating map creating method of real-time data fusion | |
Sünderhauf | Switchable Constraints for Robust Simultaneous Localization and Mapping and Satellite-Based Localization | |
CN105843223B (en) | A kind of mobile robot three-dimensional based on space bag of words builds figure and barrier-avoiding method | |
Moravec | Sensor fusion in certainty grids for mobile robots | |
Fox | Markov localization-a probabilistic framework for mobile robot localization and navigation. | |
Zhang et al. | Visual SLAM for underwater vehicles: A survey | |
Chatterjee et al. | A neuro-fuzzy assisted extended Kalman filter-based approach for simultaneous localization and mapping (SLAM) problems | |
CN107544501A (en) | A kind of intelligent robot wisdom traveling control system and its method | |
US20230243658A1 (en) | Systems, Methods and Devices for Map-Based Object's Localization Deep Learning and Object's Motion Trajectories on Geospatial Maps Using Neural Network | |
Ma et al. | A review of terrain aided navigation for underwater vehicles | |
Yu et al. | A deep-learning-based strategy for kidnapped robot problem in similar indoor environment | |
CN111429515A (en) | Learning method of robot obstacle avoidance behavior based on deep learning | |
CN112857370A (en) | Robot map-free navigation method based on time sequence information modeling | |
CN114397894B (en) | Mobile robot target searching method imitating human memory | |
Nguyen et al. | MoDeT: a low-cost obstacle tracker for self-driving mobile robot navigation using 2D-laser scan | |
Karpov et al. | Multi-robot exploration and mapping based on the subdefinite models | |
Jiang et al. | Intelligent plant cultivation robot based on key marker algorithm using visual and laser sensors | |
CN117760444A (en) | Mobile robot navigation system with vision and multi-line laser fusion | |
Murtra et al. | Action evaluation for mobile robot global localization in cooperative environments | |
Li et al. | Robot map building from sonar sensors and DSmT | |
Badalkhani et al. | Multi-robot SLAM in dynamic environments with parallel maps | |
Gasós et al. | Using fuzzy sets to represent uncertain spatial knowledge in autonomous robots | |
Gan et al. | 3d path planning for a rotary wing uav using a gaussian process occupancy map | |
Sünderhauf | Simultaneous Localization and Mapping | |
CN116698017B (en) | Object-level environment modeling method and system for indoor large-scale complex scene |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
C17 | Cessation of patent right | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20110525 Termination date: 20121107 |