CN104848851B - Intelligent Mobile Robot and its method based on Fusion composition - Google Patents
Intelligent Mobile Robot and its method based on Fusion composition Download PDFInfo
- Publication number
- CN104848851B CN104848851B CN201510289939.9A CN201510289939A CN104848851B CN 104848851 B CN104848851 B CN 104848851B CN 201510289939 A CN201510289939 A CN 201510289939A CN 104848851 B CN104848851 B CN 104848851B
- Authority
- CN
- China
- Prior art keywords
- mrow
- intelligent mobile
- mobile robot
- composition
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/20—Instruments for performing navigational calculations
Abstract
The invention discloses a kind of Intelligent Mobile Robot based on Fusion composition and its method, including robot moving platform, there are industrial computer, odometer, tachogenerator, vision collecting device and laser radar on the robot moving platform;The industrial computer is connected respectively with climb displacement device, vision collecting device and laser radar;Climb displacement device is connected with tachogenerator.Beneficial effect of the present invention:Visual pattern processing positioning of the present invention, laser radar synchronous superposition are two kinds of decision positioning methods, climb displacement device reckoning is a kind of relative positioning mode, it is successfully realized collective effect of two kinds of absolute fix modes in composition, relative positioning mode is corrected by absolute fix mode, the continuous correction of climb displacement device is realized.
Description
Technical field
The present invention is under the jurisdiction of localization for Mobile Robot and field of navigation technology, more particularly to it is a kind of based on odometer, vision,
The Intelligent Mobile Robot and method of the multisensor Data Fusion technology composition such as laser radar.
Background technology
With Intelligent Mobile Robot in recent years deeper into application, complicated substation and power transformation patrol task
Navigation mode and navigation and positioning accuracy for crusing robot propose higher requirement, in order to preferably adapt to complicated
Substation, the need for meeting complicated substation inspection task, increasing airmanship input is tested and put into should
With laser navigation technology is one of them.In robotic laser navigation procedure, realize that autonomous positioning and high accuracy navigation are to patrol
The important prerequisite that robot completes patrol task is examined, and realizes positioning and the key of high accuracy navigation and is to set up complete and accurate
Substation map, therefore, transformer station's map structuring for Intelligent Mobile Robot carry out laser navigation have it is important
Meaning.The method for substation map structuring mainly has following several at present:
1st, substation map structuring is carried out using 3D (three-dimensional) laser scanner.This method is by transformer station
The interval multiple positions of selection carry out whole station multidrop environment data acquisition.In each position, 3D laser scanners carry out level
360 ° of three-dimensional environment data acquisitions.After whole station multidrop environment data acquisition terminates, use environment data processing software is to collecting
Multidrop environment data carry out splicing reproduction processes, it is established that transformer station's three-dimensional environment map.On the basis of three-dimensional environment map
It is upper to be cut, cut height and make every effort to identical with the height of Intelligent Mobile Robot laser radar, the Final finishing output of cutting
The available two-dimentional substation map of Intelligent Mobile Robot location navigation.Although this method builds figure precision height, and environment is again
Existing effect is good, but this method cost is high, needs to be cut repeatedly for the uneven region of local transformer station's physical features, and right
It is difficult in the modification of local map.
2nd, using the method for carrying out transformer station's map structuring based on odometer, gyroscope, laser radar.This method is machine
People is roamed in whole station, and carrying out environmental data by the odometer, gyroscope, laser radar that are installed on robot body adopts
Collection.After data environment data acquisition terminates, data are entered into row format and change edlin of going forward side by side, final output two dimension substation
Map.Although this method scheme has certain feasibility, environmental data collecting, location navigation are realized in same robot
It is upper integrated, but the program is belonged to relative positioning mode and there is error, environmental data collecting process due to odometer and gyroscope
Middle error cannot be corrected always, can have larger accumulated error for a long time or after large scale reckoning, annular is returned
Road can not be closed, and this is accomplished by the processing of later stage environmental data, in addition to Data Format Transform and editor, in order to correct accumulation
The map distortion that error band comes, in addition it is also necessary to introduce complicated ICP (closest approach loop iteration matching algorithm), spring model etc. and calculate
Method, scheme is complicated, and a large amount of careful map modification adjustment work are also carried out in addition, and normative reference is obscured, and workload is big and deposits
The problems such as accuracy of map after modification is not high, reproduction effects are bad.
The content of the invention
The purpose of the present invention is exactly to solve the above problems, it is proposed that one kind is based on multisensor Data Fusion technology structure
The Intelligent Mobile Robot and its method of figure, by constantly carrying out feedback compensation to odometer, effectively reduce odometer
For a long time or large scale reckoning cumulative errors, solve the loop checking installation that odometer cumulative errors cause and do not close, map
The problems such as deformation.
To achieve these goals, the present invention is adopted the following technical scheme that:
A kind of Intelligent Mobile Robot based on Fusion composition, including robot moving platform, institute
Stating on robot moving platform has industrial computer, odometer, tachogenerator, vision collecting device and laser radar;
Industrial computer:It is carried to the computer on robot moving platform;
Climb displacement device:Receive tachogenerator signal, continuous output device people current positional information and course angle
Information, and described information is sent to industrial computer by communication line;
Vision collecting device:For gathering the mark line image on Intelligent Mobile Robot Roam Path, and by logical
Letter circuit is sent to industrial computer;
Laser radar:For realizing the scanning to Intelligent Mobile Robot operation area environmental data, and by scanning figure
As being sent to industrial computer by communication line;
The industrial computer is connected respectively with climb displacement device, vision collecting device and laser radar;Climb displacement device
It is connected with tachogenerator.
The vision collecting device includes video camera, and camera lens are parallel to the ground, is distributed around camera lens
There is polishing LED light array, video camera is connected by netting twine with industrial computer.
On the Intelligent Mobile Robot Roam Path of layout area, tag line is set, the tag line is that color is bright
The aobvious colored colour band or line for being different from road surface color.
A kind of patterning process of the Intelligent Mobile Robot based on Fusion composition, including following step
Suddenly:
(1) before Intelligent Mobile Robot patterning process starts, the substation inspection machine in target pattern region
Tag line is set on people's Roam Path;
(2) set up directly as origin (0,0,0) using the position that Intelligent Mobile Robot climb displacement device is started working
Angular coordinate system, climb displacement device is patrolled by receiving progress reckoning to tachogenerator signal, and exporting t transformer station
Examine the current positional information of robot and course angle information;
Meanwhile, the mark line image on vision collecting device collection Intelligent Mobile Robot mobile platform Roam Path,
Handled by visual pattern and identify course angle information of the t Intelligent Mobile Robot mobile platform relative to tag line;
Environment around Laser Radar Scanning robot moving platform Roam Path, obtains different laser radar datas,
The data include the distance of sampling anglec of rotation pip corresponding with the angle, and scan data is transmitted to work by Ethernet
Control machine;
(3) environmental data that t is collected according to laser radar uploads the positional information of t with climb displacement device
Local composition is carried out to scanning area with course angle information, and the position for exporting Intelligent Mobile Robot by synchronous positioning is believed
Breath and course angle information;
(4) in the course angle information for the Intelligent Mobile Robot for obtaining synchronous localization process and step (2) by regarding
Feel that the course angle information that image procossing is identified is weighted filtering process, the course angle letter for the Du Genggao that established trust by filtering
Breath;
(5) positional information obtained synchronous localization process and weighted filtering processing after course angle feedback of the information to go
Journey computing device realizes the correction to climb displacement device;
(6) repeated the above steps (2)~(5), and local composition is simultaneously carried out the global composition of fusion output by loop iteration.
In the step (2), climb displacement device to tachogenerator signal by receiving the method for carrying out reckoning
For:
Wherein, Sr (t-1), Sl (t-1) are respectively robot moving platform right wheel and revolver when the t-1 moment is to t
Between be spaced the distance passed by, d is robot moving platform wheelspan;[X (t), Y (t), W (t)] is t robot location's information
With course angle information.
In the step (2), handled by visual pattern and identify that t Intelligent Mobile Robot mobile platform is relative
It is in the specific method of the course angle information of tag line::
(2-1) image calibration:The calibrating parameters obtained using the camera calibration stage are demarcated to each two field picture, to disappear
The pattern distortion brought except camera lens;
(2-2) carries out color model transformation to chromatic image, is HSI model images RGB model conversions;
(2-3) target image is split, and selects region-of-interest in HSI model images, passes through the H and S of determination threshold value pair
Image is split, and extracts Characteristic Contrast image;
(2-4) Morphological scale-space, is measured and is extracted to characteristics of image by picture structure element, by corrosion and it is swollen
Swollen morphological method is to image procossing, to facilitate the identification and analysis to feature;
(2-5) extracts target signature, determines often row target signature center point coordinate by rim detection, passes through particle analysis
The angle of tag line and image vertical central axis line is calculated, course angle information w of the robot in working region is global is determined
(t) vision positioning, that is, is realized.
Setting identification line image processing time interval need to be shifted to an earlier date for visual pattern processing, vision collecting device is every setting
Time interval is performed once, is handled and obtained and output device people course angle information by visual pattern.
The specific method of the step (3) is:
3-1) the point set data that laser radar is collected are transformed under rectangular coordinate system;
Clustering processing 3-2) is carried out to the point set under rectangular coordinate system using clustering distance threshold value index;
3-3) point set after cluster is carried out curve fitting using least square method, reference substance characteristic straight line side is obtained
Journey, determines the center point coordinate of each reference substance characteristic straight line;
3-4) adjacent feature straight line is matched, by the local feature offset between each reference substance characteristic straight line
Distribution average computation is carried out, the optimal offset of consecutive frame feature is drawn;
Previous frame data 3-5) are added into a little optimal offset obtained in the previous step, local map is obtained;Repeat
Above-mentioned processing procedure;
3-6) by loop iteration, new laser data point set is matched with legacy data, synchronized update positional information and
Course angle information.
The step 3-2) specific method be:
The distance between point set consecutive points are calculated by order, judged whether within clustering distance threshold value indication range,
Consecutive points are clustered if in the range of;The isolated point beyond each cluster areas is will be independent of to remove.
The step 3-4) specific method be:
Front and rear adjacent each reference substance characteristic straight line central point distance of two frames is compared with characteristic distance threshold value, if adjacent spy
Levy straight central point coordinates and be then considered the adjacent same reference substance characteristic straight line of two frames less than characteristic distance threshold value;Joined with a later frame
The center point coordinate for examining thing characteristic straight line subtracts former frame correspondence reference substance feature point coordinates, obtains local feature offset, will
Local feature offset between each reference substance characteristic straight line of consecutive frame carries out distribution average computation, show that consecutive frame feature is optimal
Offset.
The beneficial effects of the invention are as follows:
Visual pattern processing positioning of the present invention, laser radar synchronous superposition are two kinds of decision positioning methods,
Climb displacement device reckoning is a kind of relative positioning mode, is successfully realized two kinds of absolute fix mode being total in composition
Same-action, relative positioning mode is corrected by absolute fix mode, realizes the continuous correction of climb displacement device.
The present invention effectively reduces climb displacement device in length by continuously carrying out feedback compensation to climb displacement device
Time or large scale reckoning cumulative errors, solve the loop checking installation that climb displacement device cumulative errors cause and do not close,
The problems such as map deformation.
The present invention is that scheme is simple, and cost is low, and feasibility is good.It is distributed with around visual image acquisition device camera lens
Polishing LED, solves the strong and weak influence to visual pattern processing of illumination, is adapted to night-environment and carries out map structuring.
Brief description of the drawings
Fig. 1 is the Intelligent Mobile Robot structural representation of the invention based on multisensor Data Fusion technology composition;
Fig. 2 is the Intelligent Mobile Robot flow chart of work methods of the invention based on multi-technical fusion composition;
Wherein, 1, robot moving platform, 2, industrial computer, 3, odometer, 4, vision collecting device, 5, laser radar, 6,
Driving wheel, 7, tag line.
Embodiment:
The present invention will be further described with embodiment below in conjunction with the accompanying drawings:
In Fig. 1, based on the Intelligent Mobile Robot of multisensor Data Fusion technology composition, it is moved including robot
Platform 1, there is an industrial computer 2 on the platform, odometer 3, vision collecting device 4, laser radar 5, the driving wheel 6 of mobile platform,
There is tag line 7 on ground.
Described robot moving platform 1, the embodiment is wheel moving platform form, if being characterized in, mobile platform has
Dry driving wheel 6;
Described industrial computer 2, is characterized in it being the computer being carried on robot moving platform;
Described odometer 3, is characterized in receiving tachogenerator signal, and be connected with industrial computer by communication line;
Described vision collecting device 4, is characterized in, the device is installed on robot moving platform, possesses core and regard
Feel acquisition elements video camera, camera lens are parallel to the ground, and video camera is connected by netting twine with industrial computer 2, in the embodiment
Mako G-032 industrial cameras are used, the image vertical central axis line of video camera is parallel with tag line 7, in camera lens week
Enclose and polishing LED array is distributed with;
The laser radar 5, is characterized in being installed on the front end or rear end of robot moving platform, and laser radar passes through logical
Letter line is connected with industrial computer 2;
Described tag line 7, its feature is temporarily arranged on layout area robot Roam Path before composition progress, is marked
Knowing line has the course angle information global relative to layout area determined;
Below in conjunction with the accompanying drawings 2 pairs the present invention relates to the substation inspection based on multisensor Data Fusion technology composition
The method and step of robot is described further.
[1] in target pattern region, before robot patterning process starts, in layout area robot Roam Path
On be provided with tag line, the tag line has the course information global relative to layout area determined.In the present embodiment, mark
Line is the colored colour band or line that a kind of color is clearly distinguishable from road surface.
[2] rectangular coordinate system, start machine are set up using the position of robot odometer start-up operation as origin (0,0,0)
People's system, starts laser radar, starts composition processing routine.The front of robot is x-axis direction in the embodiment, with x-axis
It is vertical and with axle into the direction of right-handed coordinate system be y-axis.
[3] odometer carries out reckoning by being received to tachogenerator signal, and exports t substation inspection machine
The current positional information of device people and course angle information, this is a kind of relative positioning mode;
Odometer by tachogenerator signal receive carry out reckoning method be:
Wherein, Sr (t-1), Sl (t-1) are respectively robot moving platform right wheel and revolver when the t-1 moment is to t
Between be spaced the distance passed by, d is robot moving platform wheelspan;[X (t), Y (t), W (t)] is t robot location's information
With course angle information.
Meanwhile, the mark line image on vision collecting device collection Intelligent Mobile Robot mobile platform Roam Path,
Handled by visual pattern and identify course angle information of the t Intelligent Mobile Robot mobile platform relative to tag line,
This is a kind of absolute fix mode;
Environment around Laser Radar Scanning robot moving platform Roam Path, obtains different laser radar datas,
The data include the distance of sampling anglec of rotation pip corresponding with the angle, and scan data is transmitted to work by Ethernet
Control machine;
Odometer reckoning, visual pattern processing, the scanning of laser radar environmental data, it is synchronous to carry out.In the embodiment,
Odometer carries out reckoning by being received to tachogenerator signal, and constantly the positional information of output device people's t and
Course angle information.Visual pattern processing module is to the mark line image on robot moving platform Roam Path in the embodiment
Handled, and identify the angle of t tag line feature and image vertical central axis line, the angle is this robot at moment
Course angle of the mobile platform relative to tag line.
Key step is as follows:
(1) image calibration.The calibrating parameters obtained using the camera calibration stage are demarcated to each two field picture, to eliminate
The pattern distortion that camera lens are brought.
(2) color model transformation is carried out to chromatic image.What it is due to the system needs identification is coloured sign chromatape
Or line, so needing to extract by the color model to interesting image.The color graphics processing space that the system is used
Model is HSI models, and the model is changed by illumination condition is influenceed small, and H, which represents tone, S and represents saturation degree, I, represents brightness.
RGB model conversions are HSI model images.
(3) target image is split.RIO (region-of-interest) is selected in HSI model images, by the H for learning stage determination
Image is split with S threshold value, Characteristic Contrast image is extracted.
(4) Morphological scale-space.Characteristics of image is measured and extracted by picture structure element, by corroding and expanding
Etc. morphological method to image procossing to facilitate the identification and analysis to feature.
(5) target's feature-extraction and parameter are calculated.The method that target's feature-extraction is used is to use 8 connected domains to enter image
Row scanning.After feature extraction comes out, often row target signature center point coordinate is determined by rim detection, passes through particle analysis meter
The angle of tag line and image vertical central axis line is calculated, because tag line is true in the global course information of robot work region
It is fixed, this at moment robot relative to tag line course angle have calculated that just can determine that robot in working region it is global in
Course information w (t), that is, realize vision positioning, and this is a kind of absolute fix mode.
It is computationally intensive because visual pattern processing module needs to take more resource, so needing to preset vision
Image processing time interval, the time interval can be 5~10 times of odometer reckoning cycle, visual pattern processing module
Each time interval is performed once, and by output device people course angle information w (t), vision collecting device is taken the photograph in the embodiment
Polishing LED array is distributed with around camera, the strong and weak influence to visual pattern processing of illumination can be solved.Swashing in the embodiment
Optical radar environmental data is scanned, and is that laser radar carries out rotation 190 degree or 270 degree or 360 degree of scannings, is obtained different laser thunders
Up to data, the data include the distance of sampling anglec of rotation pip corresponding with the angle, and scan data is passed through into Ethernet
Transmit to industrial computer.
[4] the composition module inside industrial computer uses SLAM (synchronous superposition method) method according to laser thunder
Scanning area is patterned up to the environmental data collected.
The specific steps of SLAM methods:
(1) laser radar data coordinate system is changed.Under being polar coordinate system due to the point set data obtained from laser radar
Data (d, w), in order to be handled under the same coordinate system, so needing to pass through (d*cosw, d*sinw) by laser radar point set data
It is transformed under rectangular coordinate system.
(2) clustering processing is carried out to the point set under rectangular coordinate system using clustering distance threshold value index.Calculated by order
The distance between adjacent point set, judges whether within cluster threshold range, is clustered consecutive points if in the range of.
For being removed independently of the isolated point beyond each cluster areas.
(3) point set after cluster is carried out curve fitting.Using least square method to the point set march after cluster
Line is fitted, and is obtained reference substance characteristic straight line equation, is determined the center point coordinate of each reference substance characteristic straight line.
(4) adjacent feature straight line is matched, by the local feature offset between each reference substance characteristic straight line
Distribution average computation is carried out, the optimal offset of consecutive frame feature is drawn.Will front and rear adjacent each reference substance characteristic straight line center of two frames
Point distance is compared with characteristic distance threshold value, and phase is considered if adjacent feature straight central point coordinates is less than characteristic distance threshold value
The same reference substance characteristic straight line of adjacent two frames;Former frame correspondence reference is subtracted with the center point coordinate of a later frame reference substance characteristic straight line
Thing feature point coordinates, obtains local feature offset, by the local feature offset between each reference substance characteristic straight line of consecutive frame
Distribution average computation is carried out, the optimal offset of consecutive frame feature is drawn.
(5) by the offset added a little in previous step of previous frame data, local map is obtained;Repeat the above
Process;
(6) by loop iteration, circulation pipe is by new laser data point set and old Data Matching, synchronized update position successively
Confidence ceases and course angle information (Xt, Yt, Wt).
[5] the synchronous positioning position informations of SLAM and course angle information enter with visual pattern processing module positioning course angle information
Row weighting is handled:
The formula of weighted filtering processing is as follows:
W'(t)=c*W (t)+(1-c) * w (t)+Pe
W (t) is by the synchronous course angle information for being positioned at the t robot that map structuring is obtained, w (t) in above formula
The course angle information of the t robot obtained for visual pattern processing.C is weight coefficient, passes through the basis in debugging process
Experience is set, and Pe is penalty coefficient because in climb displacement device and visual pattern processing procedure data exist it is asynchronous,
A certain amount of penalty coefficient is added in updating formula, this penalty coefficient will be with reference to systematic sampling speed and calculating speed
It is fixed, correct obtained course angle W'(t) complete one during the reckoning at odometer new cycle t+1 moment is input to
The odometer correction in individual cycle.
[6] the coordinate information Wt ' after the positional information (Xt, Yt) after synchronous localization process and weighting processing is fed back
The correction to odometer is realized to odometer.
[7] repeated the above steps [2]~[6], and local map is simultaneously carried out fusion output global map by loop iteration.
Although above-mentioned the embodiment of the present invention is described with reference to accompanying drawing, not to present invention protection model
The limitation enclosed, one of ordinary skill in the art should be understood that on the basis of technical scheme those skilled in the art are not
Need to pay various modifications or deform still within protection scope of the present invention that creative work can make.
Claims (7)
1. a kind of patterning process of the Intelligent Mobile Robot based on Fusion composition, it is characterized in that, use
Based on the Intelligent Mobile Robot of Fusion composition, including robot moving platform, the robot movement
There are industrial computer, odometer, tachogenerator, vision collecting device and laser radar on platform;
Industrial computer:It is carried to the computer on robot moving platform;
Climb displacement device:Tachogenerator signal, the current positional information of continuous output device people and course angle information are received,
And described information is sent to industrial computer by communication line;
Vision collecting device:For gathering the mark line image on Intelligent Mobile Robot Roam Path, and pass through order wire
Road is sent to industrial computer;
Laser radar:Lead to for realizing the scanning to Intelligent Mobile Robot operation area environmental data, and by scan image
Cross communication line and be sent to industrial computer;
The industrial computer is connected respectively with climb displacement device, vision collecting device and laser radar;Climb displacement device is with surveying
Fast sensor connection;
Method comprises the following steps:
(1) before Intelligent Mobile Robot patterning process starts, the Intelligent Mobile Robot in target pattern region overflows
Trip sets tag line on path;
(2) right angle is set up as origin (0,0,0) using the position that Intelligent Mobile Robot climb displacement device is started working to sit
Mark system, climb displacement device carries out reckoning by being received to tachogenerator signal, and exports t substation inspection machine
The current positional information of device people and course angle information;
Meanwhile, the mark line image on vision collecting device collection Intelligent Mobile Robot mobile platform Roam Path passes through
Visual pattern processing identifies course angle information of the t Intelligent Mobile Robot mobile platform relative to tag line;
Environment around Laser Radar Scanning robot moving platform Roam Path, obtains different laser radar datas, the number
According to the distance of sampling anglec of rotation pip corresponding with the angle is included, and scan data is transmitted to industry control by Ethernet
Machine;
(3) environmental data that t is collected according to laser radar uploads the positional information and boat of t with climb displacement device
Carry out local composition to scanning area to angle information, and by synchronous positioning export Intelligent Mobile Robot positional information and
Course angle information;
(4) the course angle information for the Intelligent Mobile Robot for obtaining synchronous localization process in step (2) with passing through vision figure
As the course angle information that processing is identified is weighted filtering process, by the course angle information for filtering the Du Genggao that establishes trust;
(5) the course angle feedback of the information after the positional information obtained synchronous localization process and weighted filtering processing is to stroke counter
Calculate correction of the device realization to climb displacement device;
(6) repeated the above steps (2)~(5), and local composition is simultaneously carried out the global composition of fusion output by loop iteration.
2. a kind of composition side of the Intelligent Mobile Robot based on Fusion composition as claimed in claim 1
Method, it is characterized in that, in the step (2), climb displacement device carries out reckoning by being received to tachogenerator signal
Method is:
<mrow>
<mi>X</mi>
<mrow>
<mo>(</mo>
<mi>t</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<mi>X</mi>
<mrow>
<mo>(</mo>
<mi>t</mi>
<mo>-</mo>
<mn>1</mn>
<mo>)</mo>
</mrow>
<mo>+</mo>
<mfrac>
<mrow>
<mi>S</mi>
<mi>r</mi>
<mrow>
<mo>(</mo>
<mi>t</mi>
<mo>-</mo>
<mn>1</mn>
<mo>)</mo>
</mrow>
<mo>+</mo>
<mi>S</mi>
<mi>l</mi>
<mrow>
<mo>(</mo>
<mi>t</mi>
<mo>-</mo>
<mn>1</mn>
<mo>)</mo>
</mrow>
</mrow>
<mn>2</mn>
</mfrac>
<mo>*</mo>
<mi>c</mi>
<mi>o</mi>
<mi>s</mi>
<mrow>
<mo>(</mo>
<mi>W</mi>
<mo>(</mo>
<mrow>
<mi>t</mi>
<mo>-</mo>
<mn>1</mn>
</mrow>
<mo>)</mo>
<mo>)</mo>
</mrow>
</mrow>
<mrow>
<mi>Y</mi>
<mrow>
<mo>(</mo>
<mi>t</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<mi>Y</mi>
<mrow>
<mo>(</mo>
<mi>t</mi>
<mo>-</mo>
<mn>1</mn>
<mo>)</mo>
</mrow>
<mo>+</mo>
<mfrac>
<mrow>
<mi>S</mi>
<mi>r</mi>
<mrow>
<mo>(</mo>
<mi>t</mi>
<mo>-</mo>
<mn>1</mn>
<mo>)</mo>
</mrow>
<mo>+</mo>
<mi>S</mi>
<mi>l</mi>
<mrow>
<mo>(</mo>
<mi>t</mi>
<mo>-</mo>
<mn>1</mn>
<mo>)</mo>
</mrow>
</mrow>
<mn>2</mn>
</mfrac>
<mo>*</mo>
<mi>s</mi>
<mi>i</mi>
<mi>n</mi>
<mrow>
<mo>(</mo>
<mi>W</mi>
<mo>(</mo>
<mrow>
<mi>t</mi>
<mo>-</mo>
<mn>1</mn>
</mrow>
<mo>)</mo>
<mo>)</mo>
</mrow>
</mrow>
1
<mrow>
<mi>W</mi>
<mrow>
<mo>(</mo>
<mi>t</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<mi>W</mi>
<mrow>
<mo>(</mo>
<mi>t</mi>
<mo>-</mo>
<mn>1</mn>
<mo>)</mo>
</mrow>
<mo>+</mo>
<mfrac>
<mrow>
<mi>S</mi>
<mi>r</mi>
<mrow>
<mo>(</mo>
<mi>t</mi>
<mo>-</mo>
<mn>1</mn>
<mo>)</mo>
</mrow>
<mo>-</mo>
<mi>S</mi>
<mi>l</mi>
<mrow>
<mo>(</mo>
<mi>t</mi>
<mo>-</mo>
<mn>1</mn>
<mo>)</mo>
</mrow>
</mrow>
<mi>d</mi>
</mfrac>
</mrow>
Wherein, Sr (t-1), Sl (t-1) be respectively robot moving platform right wheel and revolver at the t-1 moment between the t time
Every the distance passed by, d is robot moving platform wheelspan;[X (t), Y (t), W (t)] is t robot location information and boat
To angle information.
3. a kind of composition side of the Intelligent Mobile Robot based on Fusion composition as claimed in claim 1
Method, it is characterized in that, in the step (2), handled by visual pattern and identify t Intelligent Mobile Robot mobile platform
Specific method relative to the course angle information of tag line is::
(2-1) image calibration:The calibrating parameters obtained using the camera calibration stage are demarcated to each two field picture, to eliminate phase
The pattern distortion that machine camera lens is brought;
(2-2) carries out color model transformation to chromatic image, is HSI model images RGB model conversions;
(2-3) target image is split, and region-of-interest is selected in HSI model images, by the H and S of determination threshold value to image
Split, extract Characteristic Contrast image;
(2-4) Morphological scale-space, is measured and is extracted to characteristics of image by picture structure element, passes through what is corroded and expand
Morphological method is to image procossing, to facilitate the identification and analysis to feature;
(2-5) extracts target signature, determines often row target signature center point coordinate by rim detection, is calculated by particle analysis
Go out the angle of tag line and image vertical central axis line, determine course angle information w (t) of the robot in working region is global, i.e.,
Realize vision positioning.
4. a kind of composition side of the Intelligent Mobile Robot based on Fusion composition as claimed in claim 3
Method, it is characterized in that, need to shift to an earlier date setting identification line image processing time interval for visual pattern processing, vision collecting device every
Setting time interval is performed once, is handled and obtained and output device people course angle information by visual pattern.
5. a kind of composition side of the Intelligent Mobile Robot based on Fusion composition as claimed in claim 1
Method, it is characterized in that, the specific method of the step (3) is:
3-1) the point set data that laser radar is collected are transformed under rectangular coordinate system;
Clustering processing 3-2) is carried out to the point set under rectangular coordinate system using clustering distance threshold value index;
3-3) point set after cluster is carried out curve fitting using least square method, reference substance characteristic straight line equation is obtained, really
The center point coordinate of fixed each reference substance characteristic straight line;
3-4) adjacent feature straight line is matched, by being carried out to the local feature offset between each reference substance characteristic straight line
Average computation is distributed, the optimal offset of consecutive frame feature is drawn;
Previous frame data 3-5) are added into a little optimal offset obtained in the previous step, local map is obtained;Repeat above-mentioned
Processing procedure;
3-6) by loop iteration, new laser data point set is matched with legacy data, synchronized update positional information and course
Angle information.
6. a kind of composition side of the Intelligent Mobile Robot based on Fusion composition as claimed in claim 5
Method, it is characterized in that, the step 3-2) specific method be:
The distance between point set consecutive points are calculated by order, judged whether within clustering distance threshold value indication range, if
Then consecutive points are clustered in the range of;The isolated point beyond each cluster areas is will be independent of to remove.
7. a kind of composition side of the Intelligent Mobile Robot based on Fusion composition as claimed in claim 5
Method, it is characterized in that, the step 3-4) specific method be:
Front and rear adjacent each reference substance characteristic straight line central point distance of two frames is compared with characteristic distance threshold value, if adjacent feature is straight
Line center point coordinate is then considered the adjacent same reference substance characteristic straight line of two frames less than characteristic distance threshold value;Use a later frame reference substance
The center point coordinate of characteristic straight line subtracts former frame correspondence reference substance feature point coordinates, obtains local feature offset, will be adjacent
Local feature offset between each reference substance characteristic straight line of frame carries out distribution average computation, draws the optimal skew of consecutive frame feature
Amount.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510289939.9A CN104848851B (en) | 2015-05-29 | 2015-05-29 | Intelligent Mobile Robot and its method based on Fusion composition |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510289939.9A CN104848851B (en) | 2015-05-29 | 2015-05-29 | Intelligent Mobile Robot and its method based on Fusion composition |
Publications (2)
Publication Number | Publication Date |
---|---|
CN104848851A CN104848851A (en) | 2015-08-19 |
CN104848851B true CN104848851B (en) | 2017-08-18 |
Family
ID=53848677
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510289939.9A Active CN104848851B (en) | 2015-05-29 | 2015-05-29 | Intelligent Mobile Robot and its method based on Fusion composition |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104848851B (en) |
Families Citing this family (47)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106595630B (en) * | 2015-10-14 | 2019-09-24 | 山东鲁能智能技术有限公司 | It is a kind of that drawing system and method are built based on laser navigation Intelligent Mobile Robot |
CN106584451B (en) * | 2015-10-14 | 2019-12-10 | 国网智能科技股份有限公司 | automatic transformer substation composition robot and method based on visual navigation |
DE102015225472A1 (en) * | 2015-12-16 | 2017-06-22 | Robert Bosch Gmbh | Method and device for creating a map |
CN105865449B (en) * | 2016-04-01 | 2020-05-05 | 深圳市杉川机器人有限公司 | Hybrid positioning method of mobile robot based on laser and vision |
CN105700532B (en) * | 2016-04-19 | 2018-08-10 | 长沙理工大学 | The Intelligent Mobile Robot navigator fix control method of view-based access control model |
CN105953787A (en) * | 2016-04-27 | 2016-09-21 | 河北德普电器有限公司 | Robot navigation map generation system |
CN106052691B (en) * | 2016-05-17 | 2019-03-15 | 武汉大学 | Close ring error correction method in the mobile drawing of laser ranging |
CN106020188B (en) * | 2016-05-17 | 2018-10-30 | 杭州申昊科技股份有限公司 | A kind of Intelligent Mobile Robot recharging method based on laser navigation |
CN106052674B (en) * | 2016-05-20 | 2019-07-26 | 青岛克路德机器人有限公司 | A kind of SLAM method and system of Indoor Robot |
CN106123890A (en) * | 2016-06-14 | 2016-11-16 | 中国科学院合肥物质科学研究院 | A kind of robot localization method of Fusion |
CN106325269B (en) * | 2016-08-05 | 2019-06-11 | 上海电力学院 | Two-wheeled balance car course corrections device and method based on odometer and Magnetic Sensor |
CN106292660B (en) * | 2016-08-05 | 2019-06-11 | 上海电力学院 | Balance car course corrections device and method based on odometer and gray-scale sensor |
CN106352870B (en) * | 2016-08-26 | 2019-06-28 | 深圳微服机器人科技有限公司 | A kind of localization method and device of target |
CN106403942B (en) * | 2016-08-30 | 2022-04-29 | 全球能源互联网研究院 | Personnel indoor inertial positioning method based on substation field depth image identification |
CN106227220A (en) * | 2016-09-28 | 2016-12-14 | 关健生 | Independent navigation crusing robot based on Distributed Architecture |
CN106325277A (en) * | 2016-09-28 | 2017-01-11 | 关健生 | Distributed framework based autonomous navigation and inspection robot control method |
CN106525025B (en) * | 2016-10-28 | 2019-10-11 | 武汉大学 | A kind of Intelligent Mobile Robot route planning air navigation aid |
CN108121359A (en) * | 2016-11-29 | 2018-06-05 | 沈阳新松机器人自动化股份有限公司 | A kind of shopping robot |
CN107063264A (en) * | 2017-04-13 | 2017-08-18 | 杭州申昊科技股份有限公司 | A kind of robot map creating method suitable for extensive substation |
CN107346133B (en) * | 2017-07-04 | 2020-12-04 | 武汉视览科技有限公司 | Autonomous drawing construction method and device for indoor mobile robot |
CN109323701A (en) * | 2017-08-01 | 2019-02-12 | 郑州宇通客车股份有限公司 | The localization method and system combined based on map with FUSION WITH MULTISENSOR DETECTION |
CN116255992A (en) * | 2017-08-23 | 2023-06-13 | 北京图森未来科技有限公司 | Method and device for simultaneously positioning and mapping |
CN107450561A (en) * | 2017-09-18 | 2017-12-08 | 河南科技学院 | The autonomous path planning of mobile robot and obstacle avoidance system and its application method |
CN107677279B (en) * | 2017-09-26 | 2020-04-24 | 上海思岚科技有限公司 | Method and system for positioning and establishing image |
CN107765145B (en) * | 2017-09-26 | 2019-12-10 | 国网智能科技股份有限公司 | Automatic partial discharge detection device, system and method |
CN107907124B (en) * | 2017-09-30 | 2020-05-15 | 杭州迦智科技有限公司 | Positioning method based on scene recognition, electronic equipment, storage medium and system |
CN107808123B (en) * | 2017-09-30 | 2021-11-05 | 杭州迦智科技有限公司 | Image feasible region detection method, electronic device, storage medium and detection system |
CN108181636B (en) * | 2018-01-12 | 2020-02-18 | 中国矿业大学 | Environment modeling and map building device and method for petrochemical plant inspection robot |
CN108426580B (en) * | 2018-01-22 | 2021-04-30 | 中国地质大学(武汉) | Unmanned aerial vehicle and intelligent vehicle collaborative navigation method based on image recognition |
CN108469826B (en) * | 2018-04-23 | 2021-06-08 | 宁波Gqy视讯股份有限公司 | Robot-based map generation method and system |
CN108646265A (en) * | 2018-05-11 | 2018-10-12 | 北京华捷艾米科技有限公司 | Crusing robot and its method for inspecting, cruising inspection system |
CN111230858B (en) * | 2019-03-06 | 2022-11-22 | 南昌工程学院 | Visual robot motion control method based on reinforcement learning |
CN112013844B (en) * | 2019-05-31 | 2022-02-11 | 北京小米智能科技有限公司 | Method and device for establishing indoor environment map |
CN110200552B (en) * | 2019-06-20 | 2020-11-13 | 小狗电器互联网科技(北京)股份有限公司 | Method for removing distortion of measuring end point of laser radar and sweeper |
CN110262495B (en) * | 2019-06-26 | 2020-11-03 | 山东大学 | Control system and method capable of realizing autonomous navigation and accurate positioning of mobile robot |
CN112230256B (en) * | 2019-07-15 | 2024-04-09 | 苏州宝时得电动工具有限公司 | Autonomous robot, positioning calibration method and device thereof, and storage medium |
CN112540382B (en) * | 2019-09-07 | 2024-02-13 | 山东大学 | Laser navigation AGV auxiliary positioning method based on visual identification detection |
CN112945266A (en) * | 2019-12-10 | 2021-06-11 | 炬星科技(深圳)有限公司 | Laser navigation robot and odometer calibration method thereof |
CN111157757A (en) * | 2019-12-27 | 2020-05-15 | 苏州博田自动化技术有限公司 | Vision-based crawler speed detection device and method |
CN111694009B (en) * | 2020-05-07 | 2023-07-11 | 南昌大学 | Positioning system, method and device |
CN111638530B (en) * | 2020-05-27 | 2023-09-19 | 广州蓝胖子移动科技有限公司 | Fork truck positioning method, fork truck and computer readable storage medium |
CN115191884A (en) * | 2021-04-09 | 2022-10-18 | 美智纵横科技有限责任公司 | Movement control method and device, cleaning robot and storage medium |
CN113222900A (en) * | 2021-04-16 | 2021-08-06 | 深圳市安仕新能源科技有限公司 | Pole lug polarity detection method and detection device, terminal equipment and storage medium |
CN113708491A (en) * | 2021-08-20 | 2021-11-26 | 中国铁路北京局集团有限公司天津供电段 | Intelligent inspection system for traction substation |
CN113657333B (en) * | 2021-08-23 | 2024-01-12 | 深圳科卫机器人科技有限公司 | Guard line identification method, guard line identification device, computer equipment and storage medium |
CN113922270B (en) * | 2021-09-03 | 2023-03-14 | 东莞市简从科技有限公司 | Line patrol vehicle control method, line patrol vehicle control device, and storage medium |
CN114236564B (en) * | 2022-02-23 | 2022-06-07 | 浙江华睿科技股份有限公司 | Method for positioning robot in dynamic environment, robot, device and storage medium |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101702010A (en) * | 2009-10-20 | 2010-05-05 | 河北工业大学 | Magnetic field measuring device based on visual guiding robot |
WO2010130179A1 (en) * | 2009-05-15 | 2010-11-18 | 北京工业大学 | Flexible two-wheel self-balance robot system and motion control method thereof |
CN202153615U (en) * | 2011-07-12 | 2012-02-29 | 重庆大学 | Robot for transformer station device inspection tour |
CN102789234A (en) * | 2012-08-14 | 2012-11-21 | 广东科学中心 | Robot navigation method and robot navigation system based on color coding identifiers |
CN103234542A (en) * | 2013-04-12 | 2013-08-07 | 东南大学 | Combination vehicle curve driving track measurement method base on visual sense |
CN103292804A (en) * | 2013-05-27 | 2013-09-11 | 浙江大学 | Monocular natural vision landmark assisted mobile robot positioning method |
CN103644904A (en) * | 2013-12-17 | 2014-03-19 | 上海电机学院 | Visual navigation method based on SIFT (scale invariant feature transform) algorithm |
CN203719666U (en) * | 2013-11-21 | 2014-07-16 | 西安中科光电精密工程有限公司 | Combined navigation system |
CN104470685A (en) * | 2013-01-18 | 2015-03-25 | 艾罗伯特公司 | Mobile Robot Providing Environmental Mapping For Household Environmental Control |
-
2015
- 2015-05-29 CN CN201510289939.9A patent/CN104848851B/en active Active
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2010130179A1 (en) * | 2009-05-15 | 2010-11-18 | 北京工业大学 | Flexible two-wheel self-balance robot system and motion control method thereof |
CN101702010A (en) * | 2009-10-20 | 2010-05-05 | 河北工业大学 | Magnetic field measuring device based on visual guiding robot |
CN202153615U (en) * | 2011-07-12 | 2012-02-29 | 重庆大学 | Robot for transformer station device inspection tour |
CN102789234A (en) * | 2012-08-14 | 2012-11-21 | 广东科学中心 | Robot navigation method and robot navigation system based on color coding identifiers |
CN104470685A (en) * | 2013-01-18 | 2015-03-25 | 艾罗伯特公司 | Mobile Robot Providing Environmental Mapping For Household Environmental Control |
CN103234542A (en) * | 2013-04-12 | 2013-08-07 | 东南大学 | Combination vehicle curve driving track measurement method base on visual sense |
CN103292804A (en) * | 2013-05-27 | 2013-09-11 | 浙江大学 | Monocular natural vision landmark assisted mobile robot positioning method |
CN203719666U (en) * | 2013-11-21 | 2014-07-16 | 西安中科光电精密工程有限公司 | Combined navigation system |
CN103644904A (en) * | 2013-12-17 | 2014-03-19 | 上海电机学院 | Visual navigation method based on SIFT (scale invariant feature transform) algorithm |
Also Published As
Publication number | Publication date |
---|---|
CN104848851A (en) | 2015-08-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN104848851B (en) | Intelligent Mobile Robot and its method based on Fusion composition | |
CN104864889B (en) | A kind of robot odometer correction system and method for view-based access control model | |
CN108229366B (en) | Deep learning vehicle-mounted obstacle detection method based on radar and image data fusion | |
CN106919915B (en) | Map road marking and road quality acquisition device and method based on ADAS system | |
CN106651953B (en) | A kind of vehicle position and orientation estimation method based on traffic sign | |
CN105197252B (en) | A kind of SUAV landing method and system | |
CN111462135A (en) | Semantic mapping method based on visual S L AM and two-dimensional semantic segmentation | |
CN102435174B (en) | Method and device for detecting barrier based on hybrid binocular vision | |
CN115439424B (en) | Intelligent detection method for aerial video images of unmanned aerial vehicle | |
CN109472831A (en) | Obstacle recognition range-measurement system and method towards road roller work progress | |
CN109270534A (en) | A kind of intelligent vehicle laser sensor and camera online calibration method | |
CN106595630B (en) | It is a kind of that drawing system and method are built based on laser navigation Intelligent Mobile Robot | |
CN105930819A (en) | System for real-time identifying urban traffic lights based on single eye vision and GPS integrated navigation system | |
CN106584451B (en) | automatic transformer substation composition robot and method based on visual navigation | |
CN103925927B (en) | A kind of traffic mark localization method based on Vehicular video | |
CN107481292A (en) | The attitude error method of estimation and device of vehicle-mounted camera | |
CN104200086A (en) | Wide-baseline visible light camera pose estimation method | |
CN110379168A (en) | A kind of vehicular traffic information acquisition method based on Mask R-CNN | |
CN110675453B (en) | Self-positioning method for moving target in known scene | |
CN106996748A (en) | A kind of wheel footpath measuring method based on binocular vision | |
CN105444741A (en) | Double view window based route characteristic identifying, deviation measuring, and accurate positioning method | |
CN107703936A (en) | Automatic Guided Vehicle system and dolly localization method based on convolutional neural networks | |
CN113313047A (en) | Lane line detection method and system based on lane structure prior | |
CN112446915B (en) | Picture construction method and device based on image group | |
CN113673444B (en) | Intersection multi-view target detection method and system based on angular point pooling |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
EXSB | Decision made by sipo to initiate substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CP03 | Change of name, title or address |
Address after: 250101 Electric Power Intelligent Robot Production Project 101 in Jinan City, Shandong Province, South of Feiyue Avenue and East of No. 26 Road (ICT Industrial Park) Patentee after: National Network Intelligent Technology Co., Ltd. Address before: 250101 Shandong city of Ji'nan province high tech Zone (Lixia District) Xinluo Avenue No. 2008 silver bearing building B block 626 Patentee before: Shandong Luneng Intelligent Technology Co., Ltd. |
|
CP03 | Change of name, title or address |