CN109765901A - Dynamic cost digital map navigation method based on line laser and binocular vision - Google Patents

Dynamic cost digital map navigation method based on line laser and binocular vision Download PDF

Info

Publication number
CN109765901A
CN109765901A CN201910119415.3A CN201910119415A CN109765901A CN 109765901 A CN109765901 A CN 109765901A CN 201910119415 A CN201910119415 A CN 201910119415A CN 109765901 A CN109765901 A CN 109765901A
Authority
CN
China
Prior art keywords
map
robot
coordinate
coordinate system
dimensional
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910119415.3A
Other languages
Chinese (zh)
Inventor
毕盛
张粤
董敏
洪瀚思
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
South China University of Technology SCUT
Original Assignee
South China University of Technology SCUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China University of Technology SCUT filed Critical South China University of Technology SCUT
Priority to CN201910119415.3A priority Critical patent/CN109765901A/en
Publication of CN109765901A publication Critical patent/CN109765901A/en
Pending legal-status Critical Current

Links

Landscapes

  • Manipulator (AREA)
  • Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)

Abstract

The dynamic cost digital map navigation method based on line laser and binocular vision that the invention discloses a kind of, comprising steps of 1) establishing two-dimensional grid cartographic model;2) mobile robot pose model is established;3) laser radar data model is established;4) calibration binocular camera and acquisition three-dimensional coordinate;5) dynamic cost map is created using the three-dimensional coordinate that step 4) obtains;6) path planning is carried out using dwa algorithm and avoidance is navigated.The present invention can create cost map, the barrier that perception two-dimensional laser radar cannot detect by binocular ranging;It can be under known environment two-dimensional grid map environment, use the barrier for being above and below two-dimensional laser Radar Plane in binocular camera detection environment, and barrier is mapped in environmental map, pass through dwa algorithm, the guidance path line of avoiding obstacles is obtained, mobile robot can be made to arrive at the destination during the motion along the path for avoiding obstacle.

Description

Dynamic cost digital map navigation method based on line laser and binocular vision
Technical field
The present invention relates to the technical field of robot navigation's avoidance, refer in particular to a kind of based on line laser and binocular vision Dynamic cost digital map navigation method.
Background technique
As intelligent robot technology is flourished, robot has in production and life to be more and more widely used. According to the difference of application environment, robot is divided into plurality of classes, such as industrial robot often has multiple joints And mechanical arm, the efficient work such as assembly, carrying is carry in factory;And service robot product, including it is wheeled Mobile robot, anthropomorphic robot etc. can be used in household or the environment of commercialization, be developed for different using purpose Different function, intelligent mobile trolley may be implemented unmanned, and singing and dancing entertainment interactive etc. may be implemented in anthropomorphic robot; Unmanned plane is frequently utilized for military surveillance, remote sensing mapping etc.;It is more that there are also underwater robot, nanometer robot, agricultural robots etc. The specialized robot of classification suffers from its more single-minded purposes.Wherein with the laser sensing for pushing mobile robot development The continuous development and breakthrough of device technology, laser sensor are widely used in mobile robot field.In robot pair The exploration of circumstances not known creates on map, and navigation avoidance, and laser has the critical role that do not replace.At present on the market It is widely used and mainly two-dimensional laser at low cost, but its flatness makes robot probe flat less than lower than robot The barrier in face and the hanging barrier for being higher than laser plane.
In order to make up deficiency of the single two-dimensional laser in mobile robot, more and more sensors are added to robot On, wherein just there is camera.It can be with the function of ranging, naturally it is contemplated that using binocular camera using binocular camera The barrier that two-dimensional laser can not detect is detected, to realize that mobile robot can avoid in actual scene that laser can not The barrier detected.
In robot Real Time Obstacle Avoiding algorithm, dynamic cost map is most important, and dynamic cost map denotation is sensing The barrier that device real-time detection arrives, in conjunction with the static map of foundation, so that it may detect the barrier letter around mobile robot Breath, then in conjunction with path planning algorithm, calculates the shortest path that can be arrived at the destination.
And for robot software's system, the robot system frame of mainstream is based on ROS (The Robot at present Operating System) system, which is provided with the data-interface of various mainstream sensors, also provides the control to robot Interface processed can complete the experiment of all kinds of humanoid robots under true environment.And ROS system provides many basic functions Packet, allows us further to do experiment and research on the basic framework of robot.
Summary of the invention
It is an object of the invention to overcome the shortcomings of that two-dimensional laser sensor can only detect single plane barrier, propose A kind of dynamic cost digital map navigation method based on line laser and binocular vision, cannot be detected when navigating avoidance it is short, or The hanging barrier of person can detecte these barriers using the binocular camera demarcated and calculate these barriers and arrives The distance of head plane is imaged, then we are mapped to two-dimensional dynamic generation plus the three-dimensional generated by camera when navigating avoidance Valence map layer manages the barrier of the detection of binocular vision by this layer, as regularly updated the barrier of vision in map Barrier expired in map is deleted in position.In conjunction with the dynamic cost map of two laser, it is really generation that we, which can obtain, The information of three-dimensional barrier, allows mobile robot to avoid the barrier that laser radar can't detect in boundary.
To achieve the above object, technical solution provided by the present invention are as follows: the dynamic generation based on line laser and binocular vision Valence digital map navigation method, comprising the following steps:
1) two-dimensional grid cartographic model is established
The environmental map model using two-dimensional grid map as Environment Obstacles information is established, world coordinate system and map are established The transformational relation of coordinate system;Wherein, model indicates Environment Obstacles information using two-dimensional grid map, and two-dimensional grid cartographic model exists It is saved in the form of gray scale picture in system, and has recorded the height height of map picture in a model, map picture Corresponding world coordinates (the X of width width and map lower left corner pixellowerleft,Ylowerleft), it defines in map picture Top left corner pixel is the origin (0,0) of map coordinates system, and row coordinate is with map picture from top to bottom for positive direction, and column coordinate is with ground Figure picture is from left to right positive direction, defines the origin of world coordinates in map reference Place, resolution are map resolution ratio, and the x-axis direction of world coordinate system corresponds to the side that column coordinate increases in map coordinates system To the y-axis direction of world coordinate system corresponds to the direction that row coordinate reduces in map coordinates system, and world coordinate system is with the rice in reality For unit, each pixel has corresponded to the square two-dimensional surface region that a length and width in real world are 5cm, if pixel value For white, then it represents that barrier is not present in the region, if pixel value is black, representing the region, there are barriers;It uses Gmapping builds static map of the nomography building based on laser radar;
2) mobile robot pose model is established
It establishes using mobile robot center as the robot coordinate system of origin, establishes the robot position in two-dimensional coordinate plane Appearance model, and the pose of robot in the environment is indicated in the form of coordinate system transformational relation;Wherein, the robot pose mould For type using robot center as origin, robot center to robot positive direction is x-axis, and it is straight that right hand plane is established as unit of rice Angular coordinate system, as robot coordinate system, and indicate with Pose (x, y, θ) pose of robot, x in formula, y represent robot The coordinate of world coordinate system locating for the origin of coordinate system, θ indicate world coordinate system x-axis direction to robot coordinate system x-axis side To angle, with counterclockwise for angle increase direction, then the coordinate in robot coordinate system corresponds in world coordinate system Coordinate conversion is as follows:
In formula, x, y, θ are robot pose Pose (x, y, θ);wx,wyFor the coordinate of world coordinate system;rx,ryFor machine The coordinate of people's coordinate system;
3) laser radar data model is established
According to the riding position and direction of two-dimensional laser radar, laser radar data model is established, and is built according to step 1) The data protocol of robot pose model and laser radar that vertical cartographic model and step 2) are established, realizes laser radar The barrier data measured are simultaneously mapped in environmental map by measurement to environment distance;Wherein, the laser radar number It is as follows according to model form:
Two-dimensional laser radar can scan the complaint message in a plane within the scope of 360 degree, two-dimensional laser radar scanning Range can be up to 8 meters, and angular resolution is 1 degree, and in the data model of two-dimensional laser radar, transmitting data every time will be transmitted 360 floating datas counterclockwise transmit the distance of obstacle laser on every 1 angular direction since 0 degree of direction of laser radar The distance of radar center, as unit of rice, and recording the obstacle distance radar center distance on i degree direction is ρi, when two Dimension laser radar is centrally mounted at (0.08,0) in robot coordinate system, 0 degree of direction of laser radar and robot coordinate The y-axis direction of system it is parallel and towards it is identical when, when robot is in position and posture Pose (x, y, θ), radar data ρiMapping The coordinate into world coordinate system:
In formula, Xworld, Yworld are the coordinate in world coordinate system;
Also, by radar data ρiThe coordinate being finally mapped in map coordinates system:
In formula, height represents map picture row coordinate, and width represents map picture column coordinate, and resolution is represented Map resolution ratio;
4) calibration binocular camera and acquisition three-dimensional coordinate
The photo that 10 groups of different two cameras are clapped using scaling board, by position of the angle point in image for finding scaling board Set, come calculate two cameras internal reference and outer ginseng, the re-projection of 4 × 4 binocular camera is obtained eventually by combined calibrating Matrixcx,cyIt is left camera as planar central coordinate, c'x,c'yIt is right camera as in plane Heart coordinate;TxFor the centre distance of two cameras, f is camera focus;Then it is controlled using Block- matching Stereo Matching Algorithm Disparity map, that is, depth map of two cameras gives parallax d and picture planar point (x, y), passes through formulaAcquire the three-dimensional coordinate P=(X/W, Y/W, Z/W) under camera coordinate system;
5) dynamic cost map is created using the three-dimensional coordinate that step 4) obtains
According to the definition of binocular vision system, using the optical center of the left camera in binocular camera as origin, optical axis direction For z-axis, base direction is x-axis from left to right, as unit of rice, establishes right hand rectangular coordinate system in space, referred to as binocular coordinate system, On obtaining image after the three-dimensional coordinate P=(x', y', z') of each pixel, what y' was indicated is exactly elevation information, and x' is corresponded to Robot two-dimensional coordinate plane-y-axis, z' corresponding is x-axis direction, is screened according to y' lower than robot and is higher than robot Barrier, it is then compressed to the two-dimensional surface where robot, remembers that the coordinate under robot coordinate system is r (x, y), tool Body formula has r (x, y)=(z' ,-x'), obtains the world coordinate system coordinate (r for being compressed to robot two-dimensional surfacex,ry) after, root According to the pose (robot of current robotx,roboty, θ), it is multiplied i.e. by rigid body translation matrix with plane coordinatesObtain coordinate (w of the binocular coordinate system in world coordinate systemx,wy);It is creating When building map, it is at the center of map coordinates system, according to world coordinate system that robot initial pose, which is the origin of world coordinate system, With the relationship of map coordinates system, pose of the barrier under binocular coordinate system in map coordinates system is further calculated out, then The barrier is added in grating map, and adds expansion radius, and this makes it possible to being created that the cost based on binocular vision Figure;It can similarly be created that the cost map of two-dimensional laser using identical method, then merge two cost maps, it will be able to Obtain the cost map for the barrier that laser and vision are seen in current environment;It can be kept away using the cost map robot Barrier navigates to the target point in map;
6) finally it is exactly to select a target point, passes through the part dwa obstacle avoidance algorithm and carry out avoidance navigation: a, in robot control The spatially carry out speed dispersion sampling of system;B, for each sample rate, look at using mobile one section of the sample rate away from What can occur from after;C, each track of forward simulation is evaluated, evaluation criterion is: close to barrier, by close-target close to complete Office path and speed;The illegal route is abandoned during evaluation;D, choose the track of highest scoring and send corresponding speed To robot;E, step a-d is constantly repeated, until arriving at the destination.
Compared with prior art, the present invention have the following advantages that with the utility model has the advantages that
The method of the present invention can create cost map, the obstacle that perception two-dimensional laser radar cannot detect by binocular ranging Object;It can be swashed under known environment two-dimensional grid map environment using two dimension is above and below in binocular camera detection environment The barrier of optical radar plane, and barrier is mapped in environmental map, by dwa algorithm, obtain avoiding obstacles Guidance path line can be such that mobile robot arrives at the destination during the motion along the path for avoiding obstacle.
Detailed description of the invention
Fig. 1 is two-dimensional grid map schematic diagram.
Fig. 2 is the corresponding true environment schematic diagram of map.
Fig. 3 is mobile robot hardware platform schematic diagram.
Fig. 4 is robot coordinate system's definition and 0 degree of direction schematic diagram of laser radar.
Fig. 5 is binocular calibration exemplary diagram.
Fig. 6 is the mounting means schematic diagram of laser radar and binocular camera.
Fig. 7 is actual environment obstacle schematic diagram.
Fig. 8 is the dynamic cost map that binocular camera is merged with laser radar.
Fig. 9 is the guidance path line schematic diagram being calculated in robot system.
Figure 10 avoids obstacle according to guidance path line for robot and reaches purpose result schematic diagram.
Specific embodiment
The present invention is further explained in the light of specific embodiments.
Dynamic cost digital map navigation method provided by the present invention based on line laser and binocular vision, including following step It is rapid:
1) it establishes two-dimensional grid cartographic model: establishing the environmental map mould using two-dimensional grid map as Environment Obstacles information Type establishes the transformational relation of world coordinate system and map coordinates system.Wherein, model indicates Environment Obstacles using two-dimensional grid map Information, two-dimensional grid cartographic model is saved in the form of gray scale picture in systems, and has recorded map picture in a model Height height, the width width of map picture and the corresponding world coordinates (X of map lower left corner pixellowerleft, Ylowerleft), the origin (0,0) that top left corner pixel in map picture is map coordinates system is defined, row coordinate is with map picture from upper To it is lower be positive direction, column coordinate, from left to right for positive direction, defines the origin of world coordinates in map reference with map picturePlace, resolution are map resolution ratio, and the x-axis direction of world coordinate system is accordingly The direction that column coordinate increases in figure coordinate system, the y-axis direction of world coordinate system correspond to the side that row coordinate reduces in map coordinates system To, world coordinate system as unit of the rice in reality, each pixel corresponded to a length and width in real world be 5cm just Rectangular two-dimensional surface region, if pixel value is white, then it represents that barrier is not present in the region, if pixel value is black, generation There are barriers in the table region.The static map based on laser radar can be constructed by building nomography using gmapping.
2) it establishes mobile robot pose model: establishing using mobile robot center as the robot coordinate system of origin, build Robot pose model in vertical two-dimensional coordinate plane, and indicate in the form of coordinate system transformational relation robot in the environment Pose.Wherein, the robot pose model, using robot center as origin, robot center to robot positive direction is x Axis is established right hand plane right-angle coordinate, as robot coordinate system as unit of rice, and indicates machine with Pose (x, y, θ) The pose of device people, wherein x, y represent the coordinate of world coordinate system locating for the origin of robot coordinate system, and θ indicates world coordinates It is angle of the x-axis direction to robot coordinate system's x-axis direction, with the direction increased counterclockwise for angle, then robot coordinate system In coordinate correspond to the coordinate in world coordinate system conversion it is as follows:
In formula, x, y, θ are robot pose Pose (x, y, θ);wx,wyFor the coordinate of world coordinate system;rx,ryFor machine The coordinate of people's coordinate system.
3) it establishes laser radar data model: according to the riding position and direction of two-dimensional laser radar, establishing laser radar Data model, and the robot pose model and laser radar that the cartographic model and step 2) established according to step 1) are established Data protocol, realize laser radar to the measurement of environment distance and the barrier data measured are mapped to environmental map work as In.Wherein, the laser radar data model form is as follows:
Two-dimensional laser radar can scan the complaint message in a plane within the scope of 360 degree, general two-dimensional laser thunder Up to scanning range up to 8 meters, angular resolution is 1 degree, and in the data model of two-dimensional laser radar, transmitting data every time will be passed Defeated 360 floating datas, since 0 degree of direction of laser radar, the distance of obstacle counterclockwise transmitted on every 1 angular direction swashs The distance at optical radar center, as unit of rice, and recording the obstacle distance radar center distance on i degree direction is ρi, when Two-dimensional laser radar center is mounted at (0.08,0) in robot coordinate system, and 0 degree of direction of laser radar and robot are sat Mark system y-axis direction it is parallel and towards it is identical when, when robot is in position and posture Pose (x, y, θ), radar data ρiIt reflects It is mapped to coordinate in world coordinate system:
In formula, Xworld, Yworld are the coordinate in world coordinate system;
Also, by radar data ρiThe coordinate being finally mapped in map coordinates system,
In formula, height represents map picture row coordinate, and width represents map picture column coordinate, and resolution is represented Map resolution ratio.
4) calibration binocular camera is with acquisition three-dimensional coordinate: the photo of 10 groups of different two cameras is clapped using scaling board, By finding position of the angle point in image of scaling board, come calculate two cameras internal reference and outer ginseng, eventually by joint mark Surely the re-projection matrix of 4 × 4 binocular camera is obtainedcx,cyIt is left camera as plane Centre coordinate, c'x,c'yIt is right camera as planar central coordinate;TxFor the centre distance of two cameras, f is camera focus;Then The disparity map i.e. depth map of two cameras in left and right is obtained using Block- matching Stereo Matching Algorithm, parallax d is given and a picture is flat Face two-dimensional points (x, y), we can pass through formulaAcquire the three-dimensional coordinate P=(X/ under camera coordinate system W,Y/W,Z/W)。
5) three-dimensional coordinate obtained using step 4) creates dynamic cost map: according to the definition of binocular vision system, with The optical center of left camera in binocular camera is origin, and optical axis direction is z-axis, and base direction is x-axis from left to right, is with rice Unit establishes right hand rectangular coordinate system in space, referred to as binocular coordinate system, the three-dimensional coordinate P of each pixel on obtaining image After=(x', y', z'), what y' was indicated is exactly elevation information, x' correspond to robot two-dimensional coordinate plane-y-axis, z' is corresponding It is x-axis direction, can be screened according to y' lower than robot and higher than the barrier of robot, it is then compressed to robot The two-dimensional surface at place remembers that the coordinate under robot coordinate system is r (x, y), and specific formula has r (x, y)=(z' ,-x'), obtains It is compressed to the world coordinate system coordinate (r of robot two-dimensional surfacex,ry) after, according to the pose (robot of current robotx, roboty, θ), it is multiplied i.e. by rigid body translation matrix with plane coordinatesIt obtains double Coordinate (w of the mesh coordinate system in world coordinate systemx,wy), when creating map, robot initial pose is world coordinate system Origin, according to the relationship of world coordinate system and map coordinates system, can be further calculated out double at the center of map coordinates system Pose of the barrier in map coordinates system under mesh coordinate system, then we add the barrier in grating map, and add Add expansion radius, can thus be created that the cost map based on binocular vision.It can similarly be created using identical method Then the cost map of two-dimensional laser out merges two cost maps, so that it may obtain laser and vision in current environment and be seen The cost map of the barrier arrived.Using the cost map robot can avoidance navigate to the target point in map.
6) finally it is exactly to select a target point, passes through the part dwa obstacle avoidance algorithm and carry out avoidance navigation: a, in robot control The spatially carry out speed dispersion sampling of system.B, for each sample rate, look at a bit of using sample rate movement After what can occur.C, evaluate each track of forward simulation, evaluation criterion such as: close to barrier, by close-target close to the overall situation Path and speed;The illegal route is abandoned during evaluation.D, choose the track of highest scoring and send corresponding speed to Robot.E, step a-d is constantly repeated, until arriving at the destination.
Case study on implementation: the shifting based on binocular camera Yu two-dimensional laser radar is completed on KOBUKI mobile robot platform Mobile robot navigation.
(1) two-dimensional grid cartographic model
Two-dimensional grid map as shown in Figure 1, storage information include the line number of every row number of pixels and map picture in total, Each pixel represents the plane domain of a length and width 5cm in true environment in map, represents the region using black picture element and deposits In obstacle, white pixel represents the region, and there is no obstacles.
In this example, it is highly 60 pixels that map picture width, which is 50 pixels, represents 2.5 meters one piece wide, long 3 meters of reality Place is tested, as shown in Figure 2.
(2) mobile robot platform model
In the selection of mobile robot hardware platform, KOBUKI General Mobile robot base has been used, and LSLIDAR2D laser radar and two 300,000 common pixel camera heads are as sensor acquisition range information, an X86 platform The terminal of 4GB memory is as robot server-side, as shown in Figure 3.
To the setting of the mounting means of sensor and robot coordinate as shown in figure 4, being with robot center in this example Origin establishes right hand plane coordinate system using robot positive direction as x-axis, and 0 degree of direction of laser radar is put and machine The y-axis of device people's coordinate system is parallel.
(3) for the scaling board of binocular calibration, mean square error, mark calibration binocular camera and acquisition three-dimensional coordinate: are demarcated It is as shown in Figure 5 to determine result Q.
(4) it is led based on binocular dynamic cost map and dynamic Laser cost map and combining for two-dimensional grid static map Boat positioning.
In the implementation case: on robot platform, the binocular camera form fixed with laser radar is as shown in Figure 6. True environment is as shown in Figure 7.And the obstacle information of global context is indicated using two-dimensional grid map in robot system, make Barrier is detected simultaneously with binocular camera and two-dimensional laser radar, and is not interfere with each other, local environment complaint message is mapped To in environmental map, the Global obstacle information within the scope of 5 meters of robot periphery is replaced using local disturbance's information, such as Fig. 8 Shown, the short chest of white is less than the obstacle of laser radar scanning plane in Fig. 7, it is detected to obtain by binocular camera, and And it is mapped in the map of Fig. 8.Guidance path shown in Fig. 9 has so been obtained by dwa algorithm.According to the road of navigation programming Diameter eventually arrives at terminal shown in Fig. 10.
Therefore, in conclusion the method for the present invention can create cost map by binocular ranging, two-dimensional laser radar is perceived The barrier that cannot be detected;It can be under known environment two-dimensional grid map environment, using high in binocular camera detection environment In with the barrier that is lower than two-dimensional laser Radar Plane, and barrier is mapped in environmental map, by dwa algorithm, is obtained To the guidance path line of avoiding obstacles, mobile robot can be made to reach mesh along the path for avoiding obstacle during the motion Ground, have actual promotional value, be worthy to be popularized.
Embodiment described above is only the preferred embodiments of the invention, and but not intended to limit the scope of the present invention, therefore All shapes according to the present invention change made by principle, should all be included within the scope of protection of the present invention.

Claims (1)

1. the dynamic cost digital map navigation method based on line laser and binocular vision, which comprises the following steps:
1) two-dimensional grid cartographic model is established
The environmental map model using two-dimensional grid map as Environment Obstacles information is established, world coordinate system and map reference are established The transformational relation of system;Wherein, model indicates Environment Obstacles information using two-dimensional grid map, and two-dimensional grid cartographic model is in system In saved in the form of gray scale picture, and have recorded the height height of map picture, the width of map picture in a model Corresponding world coordinates (the X of width and map lower left corner pixellowerleft,Ylowerleft), define upper left in map picture Angle pixel is the origin (0,0) of map coordinates system, and row coordinate is with map picture from top to bottom for positive direction, and column coordinate is with Map Piece is from left to right positive direction, defines the origin of world coordinates in map referencePlace, Resolution is map resolution ratio, and the x-axis direction of world coordinate system corresponds to the direction that column coordinate increases in map coordinates system, generation The y-axis direction of boundary's coordinate system corresponds to the direction that row coordinate reduces in map coordinates system, and world coordinate system is with the meter Wei Dan in reality Position, each pixel have corresponded to the square two-dimensional surface region that a length and width in real world are 5cm, if pixel value is white Color, then it represents that barrier is not present in the region, if pixel value is black, representing the region, there are barriers;It uses Gmapping builds static map of the nomography building based on laser radar;
2) mobile robot pose model is established
It establishes using mobile robot center as the robot coordinate system of origin, establishes the robot pose mould in two-dimensional coordinate plane Type, and the pose of robot in the environment is indicated in the form of coordinate system transformational relation;Wherein, the robot pose model with Robot center is origin, and robot center to robot positive direction is x-axis, and right hand flat square seat is established as unit of rice System, as robot coordinate system are marked, and indicates the pose of robot with Pose (x, y, θ), x in formula, y represent robot coordinate The coordinate of world coordinate system locating for the origin of system, θ indicate world coordinate system x-axis direction to robot coordinate system's x-axis direction Angle, with the direction increased counterclockwise for angle, then the coordinate in robot coordinate system corresponds to the coordinate in world coordinate system It converts as follows:
In formula, x, y, θ are robot pose Pose (x, y, θ);wx,wyFor the coordinate of world coordinate system;rx,ryFor robot coordinate The coordinate of system;
3) laser radar data model is established
According to the riding position and direction of two-dimensional laser radar, laser radar data model is established, and established according to step 1) The data protocol of robot pose model and laser radar that cartographic model and step 2) are established realizes laser radar to ring The barrier data measured are simultaneously mapped in environmental map by the measurement of border distance;Wherein, the laser radar data mould Type form is as follows:
Two-dimensional laser radar can scan the complaint message in a plane within the scope of 360 degree, two-dimensional laser radar scanning range Can be up to 8 meters, angular resolution is 1 degree, and in the data model of two-dimensional laser radar, will be transmitted by transmitting data every time by 360 Floating data counterclockwise transmits the distance of obstacle laser radar on every 1 angular direction since 0 degree of direction of laser radar The distance at center, as unit of rice, and recording the obstacle distance radar center distance on i degree direction is ρi, when two dimension swashs Optical radar is centrally mounted at (0.08,0) in robot coordinate system, the y in 0 degree of direction and robot coordinate system of laser radar Axis direction it is parallel and towards it is identical when, when robot is in position and posture Pose (x, y, θ), radar data ρiIt is mapped to the world Coordinate in coordinate system:
In formula, Xworld, Yworld are the coordinate in world coordinate system;
Also, by radar data ρiThe coordinate being finally mapped in map coordinates system:
In formula, height represents map picture row coordinate, and width represents map picture column coordinate, and resolution represents map Resolution ratio;
4) calibration binocular camera and acquisition three-dimensional coordinate
The photo that 10 groups of different two cameras are clapped using scaling board is come by finding position of the angle point in image of scaling board Calculate two cameras internal reference and outer ginseng, obtain the re-projection matrix of 4 × 4 binocular camera eventually by combined calibratingcx,cyIt is left camera as planar central coordinate, c'x,c'yFor right camera as planar central is sat Mark;TxFor the centre distance of two cameras, f is camera focus;Then left and right two is obtained using Block- matching Stereo Matching Algorithm Disparity map, that is, depth map of camera gives parallax d and picture planar point (x, y), passes through formula Acquire the three-dimensional coordinate P=(X/W, Y/W, Z/W) under camera coordinate system;
5) dynamic cost map is created using the three-dimensional coordinate that step 4) obtains
According to the definition of binocular vision system, using the optical center of the left camera in binocular camera as origin, optical axis direction z Axis, base direction is x-axis from left to right, as unit of rice, establishes right hand rectangular coordinate system in space, referred to as binocular coordinate system, It obtains on image after the three-dimensional coordinate P=(x', y', z') of each pixel, what y' was indicated is exactly elevation information, and x' corresponds to machine Device people's two-dimensional coordinate plane-y-axis, z' corresponding is x-axis direction, is screened according to y' lower than robot and higher than robot Then it is compressed to the two-dimensional surface where robot by barrier, remember that the coordinate under robot coordinate system is r (x, y), specifically Formula has r (x, y)=(z' ,-x'), obtains the world coordinate system coordinate (r for being compressed to robot two-dimensional surfacex,ry) after, according to Pose (the robot of current robotx,roboty, θ), it is multiplied i.e. by rigid body translation matrix with plane coordinatesObtain coordinate (w of the binocular coordinate system in world coordinate systemx,wy);It is creating When building map, it is at the center of map coordinates system, according to world coordinate system that robot initial pose, which is the origin of world coordinate system, With the relationship of map coordinates system, pose of the barrier under binocular coordinate system in map coordinates system is further calculated out, then The barrier is added in grating map, and adds expansion radius, and this makes it possible to being created that the cost based on binocular vision Figure;It can similarly be created that the cost map of two-dimensional laser using identical method, then merge two cost maps, it will be able to Obtain the cost map for the barrier that laser and vision are seen in current environment;It can be kept away using the cost map robot Barrier navigates to the target point in map;
6) be finally exactly to select a target point, pass through the part dwa obstacle avoidance algorithm carry out avoidance navigation: a, robot control Spatially carry out speed dispersion sampling;B, for each sample rate, look at using after the mobile a distance of the sample rate What can occur;C, each track of forward simulation is evaluated, evaluation criterion is: close to barrier, leans on close-target close to global road Diameter and speed;The illegal route is abandoned during evaluation;D, choose the track of highest scoring and send corresponding speed to machine Device people;E, step a-d is constantly repeated, until arriving at the destination.
CN201910119415.3A 2019-02-18 2019-02-18 Dynamic cost digital map navigation method based on line laser and binocular vision Pending CN109765901A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910119415.3A CN109765901A (en) 2019-02-18 2019-02-18 Dynamic cost digital map navigation method based on line laser and binocular vision

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910119415.3A CN109765901A (en) 2019-02-18 2019-02-18 Dynamic cost digital map navigation method based on line laser and binocular vision

Publications (1)

Publication Number Publication Date
CN109765901A true CN109765901A (en) 2019-05-17

Family

ID=66456832

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910119415.3A Pending CN109765901A (en) 2019-02-18 2019-02-18 Dynamic cost digital map navigation method based on line laser and binocular vision

Country Status (1)

Country Link
CN (1) CN109765901A (en)

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110108269A (en) * 2019-05-20 2019-08-09 电子科技大学 AGV localization method based on Fusion
CN110502019A (en) * 2019-09-06 2019-11-26 北京云迹科技有限公司 A kind of barrier-avoiding method and device of Indoor Robot
CN110609559A (en) * 2019-10-25 2019-12-24 江苏恒澄交科信息科技股份有限公司 Improved DWA dynamic window method for unmanned ship path following and obstacle avoidance
CN110726409A (en) * 2019-09-09 2020-01-24 杭州电子科技大学 Map fusion method based on laser SLAM and visual SLAM
CN111469127A (en) * 2020-04-14 2020-07-31 北京海益同展信息科技有限公司 Cost map updating method and device, robot and storage medium
CN111726591A (en) * 2020-06-22 2020-09-29 珠海格力电器股份有限公司 Map updating method, map updating device, storage medium and electronic equipment
CN111743464A (en) * 2020-07-06 2020-10-09 追创科技(苏州)有限公司 Obstacle avoidance method and device based on line laser
CN111830977A (en) * 2020-07-02 2020-10-27 中国兵器科学研究院 Autonomous navigation software framework and navigation method for mobile robot
CN112129281A (en) * 2019-06-25 2020-12-25 南京航空航天大学 High-precision image navigation positioning method based on local neighborhood map
CN112180943A (en) * 2020-10-19 2021-01-05 山东交通学院 Underwater robot navigation obstacle avoidance method based on visual image and laser radar
CN112286194A (en) * 2020-10-29 2021-01-29 广东杜尼智能机器人工程技术研究中心有限公司 Cost map area division method
CN112294197A (en) * 2020-11-04 2021-02-02 深圳市普森斯科技有限公司 Sweeping control method of sweeper, electronic device and storage medium
CN112363498A (en) * 2020-10-19 2021-02-12 山东交通学院 Underwater robot intelligent motion control method based on laser radar
CN112378408A (en) * 2020-11-26 2021-02-19 重庆大学 Path planning method for realizing real-time obstacle avoidance of wheeled mobile robot
CN112447065A (en) * 2019-08-16 2021-03-05 北京地平线机器人技术研发有限公司 Trajectory planning method and device
CN112462768A (en) * 2020-11-25 2021-03-09 深圳拓邦股份有限公司 Mobile robot navigation map creating method and device and mobile robot
CN112489131A (en) * 2020-12-26 2021-03-12 上海有个机器人有限公司 Method, device, medium and robot for constructing cost map based on road surface detection
CN112486172A (en) * 2020-11-30 2021-03-12 深圳市普渡科技有限公司 Road edge detection method and robot
CN112506196A (en) * 2020-12-07 2021-03-16 合肥工业大学 Robot obstacle avoidance method and system based on priori knowledge
CN112505723A (en) * 2021-02-03 2021-03-16 之江实验室 Three-dimensional map reconstruction method based on navigation point selection
CN112817315A (en) * 2020-12-31 2021-05-18 江苏集萃智能制造技术研究所有限公司 Obstacle avoidance method and system for unmanned cleaning vehicle in dynamic environment
CN113239520A (en) * 2021-04-16 2021-08-10 大连海事大学 Near-water-bottom three-dimensional underwater terrain environment modeling method
CN113520246A (en) * 2021-07-30 2021-10-22 珠海一微半导体股份有限公司 Mobile robot compensation cleaning method and system
CN113607162A (en) * 2021-10-09 2021-11-05 创泽智能机器人集团股份有限公司 Path planning method and device based on three-dimensional map
CN113608532A (en) * 2021-07-29 2021-11-05 深圳市眼科医院 Automatic obstacle avoidance system for blind people going out and scooter thereof
CN114194685A (en) * 2021-12-23 2022-03-18 山东新华医疗器械股份有限公司 Stacking AGV control system, method and device
WO2022127541A1 (en) * 2020-12-16 2022-06-23 北京极智嘉科技股份有限公司 Robot and localization method
CN115406445A (en) * 2022-08-18 2022-11-29 四川华丰科技股份有限公司 Multi-sensor data fusion processing method and robot obstacle avoidance method
CN115775309A (en) * 2023-01-31 2023-03-10 中国电子科技集团公司第二十八研究所 Fire power range marking method and device for three-dimensional space

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103400392A (en) * 2013-08-19 2013-11-20 山东鲁能智能技术有限公司 Binocular vision navigation system and method based on inspection robot in transformer substation
CN108151728A (en) * 2017-12-06 2018-06-12 华南理工大学 A kind of half dense cognitive map creation method for binocular SLAM
US20180181141A1 (en) * 2016-12-28 2018-06-28 H.P.B Optoelectronic Co., Ltd Robot surveillance system
CN108214487A (en) * 2017-12-16 2018-06-29 广西电网有限责任公司电力科学研究院 Based on the positioning of the robot target of binocular vision and laser radar and grasping means
CN108663681A (en) * 2018-05-16 2018-10-16 华南理工大学 Mobile Robotics Navigation method based on binocular camera Yu two-dimensional laser radar
CN108839055A (en) * 2018-05-31 2018-11-20 芜湖星途机器人科技有限公司 Robot with independent navigation
CN108932736A (en) * 2018-05-30 2018-12-04 南昌大学 Two-dimensional laser radar Processing Method of Point-clouds and dynamic robot pose calibration method
CN108983783A (en) * 2018-08-03 2018-12-11 上海有个机器人有限公司 A kind of robot movement speed control method and system, storage medium and terminal
CN109118542A (en) * 2017-06-22 2019-01-01 百度在线网络技术(北京)有限公司 Scaling method, device, equipment and storage medium between laser radar and camera
CN109129507A (en) * 2018-09-10 2019-01-04 北京联合大学 A kind of medium intelligent introduction robot and explanation method and system
CN109270534A (en) * 2018-05-07 2019-01-25 西安交通大学 A kind of intelligent vehicle laser sensor and camera online calibration method

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103400392A (en) * 2013-08-19 2013-11-20 山东鲁能智能技术有限公司 Binocular vision navigation system and method based on inspection robot in transformer substation
US20180181141A1 (en) * 2016-12-28 2018-06-28 H.P.B Optoelectronic Co., Ltd Robot surveillance system
CN109118542A (en) * 2017-06-22 2019-01-01 百度在线网络技术(北京)有限公司 Scaling method, device, equipment and storage medium between laser radar and camera
CN108151728A (en) * 2017-12-06 2018-06-12 华南理工大学 A kind of half dense cognitive map creation method for binocular SLAM
CN108214487A (en) * 2017-12-16 2018-06-29 广西电网有限责任公司电力科学研究院 Based on the positioning of the robot target of binocular vision and laser radar and grasping means
CN109270534A (en) * 2018-05-07 2019-01-25 西安交通大学 A kind of intelligent vehicle laser sensor and camera online calibration method
CN108663681A (en) * 2018-05-16 2018-10-16 华南理工大学 Mobile Robotics Navigation method based on binocular camera Yu two-dimensional laser radar
CN108932736A (en) * 2018-05-30 2018-12-04 南昌大学 Two-dimensional laser radar Processing Method of Point-clouds and dynamic robot pose calibration method
CN108839055A (en) * 2018-05-31 2018-11-20 芜湖星途机器人科技有限公司 Robot with independent navigation
CN108983783A (en) * 2018-08-03 2018-12-11 上海有个机器人有限公司 A kind of robot movement speed control method and system, storage medium and terminal
CN109129507A (en) * 2018-09-10 2019-01-04 北京联合大学 A kind of medium intelligent introduction robot and explanation method and system

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
ROBOTDREAMSKY: "ROS机器人导航", 《ROBOTDREAMSKY HTTPS://WENKU.BAIDU.COM/VIEW/B70E5289B7360B4C2F3F6497.HTML?FROM=SEARCH》 *
冯里千: "移动机器人自主探索环境定位导航技术的研究与实现", 《 中优秀国硕士学位论文全文数据库,信息科技辑》 *

Cited By (43)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110108269B (en) * 2019-05-20 2023-01-17 电子科技大学 AGV positioning method based on multi-sensor data fusion
CN110108269A (en) * 2019-05-20 2019-08-09 电子科技大学 AGV localization method based on Fusion
CN112129281B (en) * 2019-06-25 2022-02-22 南京航空航天大学 High-precision image navigation positioning method based on local neighborhood map
CN112129281A (en) * 2019-06-25 2020-12-25 南京航空航天大学 High-precision image navigation positioning method based on local neighborhood map
CN112447065B (en) * 2019-08-16 2022-04-26 北京地平线机器人技术研发有限公司 Trajectory planning method and device
CN112447065A (en) * 2019-08-16 2021-03-05 北京地平线机器人技术研发有限公司 Trajectory planning method and device
CN110502019A (en) * 2019-09-06 2019-11-26 北京云迹科技有限公司 A kind of barrier-avoiding method and device of Indoor Robot
CN110726409A (en) * 2019-09-09 2020-01-24 杭州电子科技大学 Map fusion method based on laser SLAM and visual SLAM
CN110726409B (en) * 2019-09-09 2021-06-22 杭州电子科技大学 Map fusion method based on laser SLAM and visual SLAM
CN110609559A (en) * 2019-10-25 2019-12-24 江苏恒澄交科信息科技股份有限公司 Improved DWA dynamic window method for unmanned ship path following and obstacle avoidance
CN111469127A (en) * 2020-04-14 2020-07-31 北京海益同展信息科技有限公司 Cost map updating method and device, robot and storage medium
CN111469127B (en) * 2020-04-14 2022-06-07 京东科技信息技术有限公司 Cost map updating method and device, robot and storage medium
CN111726591A (en) * 2020-06-22 2020-09-29 珠海格力电器股份有限公司 Map updating method, map updating device, storage medium and electronic equipment
CN111726591B (en) * 2020-06-22 2021-11-23 珠海格力电器股份有限公司 Map updating method, map updating device, storage medium and electronic equipment
CN111830977A (en) * 2020-07-02 2020-10-27 中国兵器科学研究院 Autonomous navigation software framework and navigation method for mobile robot
CN111743464A (en) * 2020-07-06 2020-10-09 追创科技(苏州)有限公司 Obstacle avoidance method and device based on line laser
CN112363498A (en) * 2020-10-19 2021-02-12 山东交通学院 Underwater robot intelligent motion control method based on laser radar
CN112180943A (en) * 2020-10-19 2021-01-05 山东交通学院 Underwater robot navigation obstacle avoidance method based on visual image and laser radar
CN112286194B (en) * 2020-10-29 2022-05-17 广东杜尼智能机器人工程技术研究中心有限公司 Cost map area division method
CN112286194A (en) * 2020-10-29 2021-01-29 广东杜尼智能机器人工程技术研究中心有限公司 Cost map area division method
CN112294197A (en) * 2020-11-04 2021-02-02 深圳市普森斯科技有限公司 Sweeping control method of sweeper, electronic device and storage medium
CN112462768A (en) * 2020-11-25 2021-03-09 深圳拓邦股份有限公司 Mobile robot navigation map creating method and device and mobile robot
CN112462768B (en) * 2020-11-25 2024-03-29 深圳拓邦股份有限公司 Mobile robot navigation map creation method and device and mobile robot
CN112378408A (en) * 2020-11-26 2021-02-19 重庆大学 Path planning method for realizing real-time obstacle avoidance of wheeled mobile robot
WO2022111723A1 (en) * 2020-11-30 2022-06-02 深圳市普渡科技有限公司 Road edge detection method and robot
CN112486172A (en) * 2020-11-30 2021-03-12 深圳市普渡科技有限公司 Road edge detection method and robot
CN112506196A (en) * 2020-12-07 2021-03-16 合肥工业大学 Robot obstacle avoidance method and system based on priori knowledge
WO2022127541A1 (en) * 2020-12-16 2022-06-23 北京极智嘉科技股份有限公司 Robot and localization method
CN112489131B (en) * 2020-12-26 2024-04-05 上海有个机器人有限公司 Method, device, medium and robot for constructing cost map based on pavement detection
CN112489131A (en) * 2020-12-26 2021-03-12 上海有个机器人有限公司 Method, device, medium and robot for constructing cost map based on road surface detection
CN112817315A (en) * 2020-12-31 2021-05-18 江苏集萃智能制造技术研究所有限公司 Obstacle avoidance method and system for unmanned cleaning vehicle in dynamic environment
CN112505723B (en) * 2021-02-03 2024-01-23 之江实验室 Three-dimensional map reconstruction method based on navigation point selection
CN112505723A (en) * 2021-02-03 2021-03-16 之江实验室 Three-dimensional map reconstruction method based on navigation point selection
CN113239520A (en) * 2021-04-16 2021-08-10 大连海事大学 Near-water-bottom three-dimensional underwater terrain environment modeling method
CN113239520B (en) * 2021-04-16 2023-09-08 大连海事大学 Near-water three-dimensional underwater topography environment modeling method
CN113608532A (en) * 2021-07-29 2021-11-05 深圳市眼科医院 Automatic obstacle avoidance system for blind people going out and scooter thereof
CN113520246A (en) * 2021-07-30 2021-10-22 珠海一微半导体股份有限公司 Mobile robot compensation cleaning method and system
CN113607162A (en) * 2021-10-09 2021-11-05 创泽智能机器人集团股份有限公司 Path planning method and device based on three-dimensional map
CN114194685A (en) * 2021-12-23 2022-03-18 山东新华医疗器械股份有限公司 Stacking AGV control system, method and device
CN115406445A (en) * 2022-08-18 2022-11-29 四川华丰科技股份有限公司 Multi-sensor data fusion processing method and robot obstacle avoidance method
CN115406445B (en) * 2022-08-18 2024-05-17 四川华丰科技股份有限公司 Multi-sensor data fusion processing method and robot obstacle avoidance method
CN115775309A (en) * 2023-01-31 2023-03-10 中国电子科技集团公司第二十八研究所 Fire power range marking method and device for three-dimensional space
CN115775309B (en) * 2023-01-31 2023-04-04 中国电子科技集团公司第二十八研究所 Fire power range marking method and device for three-dimensional space

Similar Documents

Publication Publication Date Title
CN109765901A (en) Dynamic cost digital map navigation method based on line laser and binocular vision
CN108663681B (en) Mobile robot navigation method based on binocular camera and two-dimensional laser radar
CN111156998B (en) Mobile robot positioning method based on RGB-D camera and IMU information fusion
EP3321888B1 (en) Projected image generation method and device, and method for mapping image pixels and depth values
AU2018295572B2 (en) Real time position and orientation tracker
Zhang et al. Intelligent collaborative localization among air-ground robots for industrial environment perception
JP4794019B2 (en) Apparatus and method for providing a three-dimensional map representation of a region
CN105354875B (en) A kind of indoor environment is two-dimentional with the construction method and system of three-dimensional conjunctive model
Maier et al. Real-time navigation in 3D environments based on depth camera data
CN104932001B (en) A kind of 3D nuclear radiation environments in real time rebuild monitoring system
Rovira-Más et al. Creation of three-dimensional crop maps based on aerial stereoimages
Bruno et al. Augmented reality visualization of scene depth for aiding ROV pilots in underwater manipulation
CN114529585A (en) Mobile equipment autonomous positioning method based on depth vision and inertial measurement
CN110202577A (en) A kind of autonomous mobile robot that realizing detection of obstacles and its method
Jensen et al. Laser range imaging using mobile robots: From pose estimation to 3D-models
Maurer et al. Geo-referenced 3d reconstruction: Fusing public geographic data and aerial imagery
CN116957360A (en) Space observation and reconstruction method and system based on unmanned aerial vehicle
Wang et al. Micro aerial vehicle navigation with visual-inertial integration aided by structured light
Lu et al. 3-D location estimation of underwater circular features by monocular vision
Allotta et al. Thesaurus project: Design of new autonomous underwater vehicles for documentation and protection of underwater archaeological sites
Brujic-Okretic et al. Remote vehicle manoeuvring using augmented reality
Bai et al. Research on key technologies of meal delivery robot based on the fusion of lidar and machine vision
Zhou et al. A multi-binocular camera-based localization method for amphibious spherical robots
Orlov et al. Machine vision system for autonomous agricultural vehicle
Amano et al. Full color 3D point clouds from bird's-eye view using multi-view laser scanner and quadcopter

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20190517