A kind of water surface is unmanned to equip multi-source cognitive method and system
Technical field
The present invention relates to the technical field of the unmanned equipment Study of intelligent of the water surface, in particular to the unmanned equipment of a kind of water surface
Multi-source cognitive method and system.
Background technique
The unmanned equipment of the water surface is that have nonlinearity dynamic characteristic, can be in the case where unmanned intervene various multiple
The novel carriers that task is executed under miscellaneous unknown aquatic environment, have many advantages, such as small, intelligent, autonomy-oriented, Chang Beiyong
The task that danger coefficient is high, operating environment is severe is executed, is had in fields such as military combat, sea area patrol, islands and reefs supplies wide
General application demand.Since the intelligent realization process of the unmanned equipment of the water surface depends on the quality of its environment sensing ability first,
Good environment perception method and system can provide for the autonomous intelligence decision process of the unmanned equipment of the water surface and its important ring
Border prior information, to maintain the safety of its operation, accuracy and reliability.Therefore, it is more to study a kind of unmanned equipment of the water surface
Source cognitive method and system realize the unmanned equipment of the water surface for promoting the autonomous intelligence progress of the unmanned equipment of the water surface
Effective operation has great importance.
Summary of the invention
The purpose of the present invention is to overcome the shortcomings of the existing technology and deficiency, provides a kind of unmanned equipment multi-source sense of the water surface
Perception method and system.The present invention passes through building Surface Picture data set pair for the multi-source perception problems of the unmanned equipment of the water surface
Deeplab and Faster RCNN network model is trained, to realize the identification of water surface boundary line and water hazard object.Root
According to the combined calibrating between camera and three-dimensional laser radar as a result, the three dimensional point cloud that laser radar obtains is projected to phase
On the image that machine obtains, depth information is added to image, then final by the conversion of camera coordinates system-world coordinate system coordinate
The world coordinates information of barrier and water surface boundary line is obtained, and passes through the topic of ROS (Robot Operating System)
Communication mechanism to application module, believes the information real-time delivery to provide priori environment for the unmanned decision for equipping next step
Breath.The purpose of the present invention is realized by the following technical solution:
A kind of unmanned equipment multi-source cognitive method of the water surface comprising the steps of:
The sensing parameter of S1, the in real time unmanned equipment multi-source sensory perceptual system of the acquisition water surface, obtains the visual information of aquatic environment
And three-dimensional point cloud information;
S2, preparatory collected Surface Picture is manually demarcated, utilizes the data set pair manually demarcated
Deeplab model and Faster RCNN model are trained and save network model parameter;
S3, the Surface Picture inputted in real time is divided by background, land and water surface three classes by Deeplab model, according to
The circumference of water-surface areas extracts water surface boundary line;
S4, the prediction block that water hazard object is extracted by Faster RCNN network model, calculate separately ship and floating
Friendship and ratio between object prediction block and the water-surface areas of image, semantic segmentation network output, reject meaningless detection of obstacles
As a result;
S5, camera calibration is carried out, obtains camera internal reference and outer ginseng, then carry out the joint of three-dimensional laser radar and camera
Calibration obtains the coordinate transformation relation between radar and camera in conjunction with calibration result;
S6, the three dimensional point cloud that laser radar obtains is projected into the image that camera obtains according to coordinate transformation relation
On, to image add depth information, then by camera coordinates system-world coordinate system coordinate conversion finally obtain barrier and
The world coordinates of water surface boundary line.
Further, the step S1 specifically: the visual information for obtaining Surface Picture in real time using camera uses three
It ties up laser radar and carries out real time scan to fan-shaped region to before unmanned equipment, obtain the three-dimensional point cloud information of aquatic environment;
Further, the step S2 specifically: Pixel-level is demarcated as from top to bottom by preparatory collected Surface Picture
Background, land and water surface three classes, for the training of Deeplab network model.Barrier candidate frame in Surface Picture is demarcated as
Two class of ship and floating material, for the training of Faster RCNN network model, to construct Surface Picture data set.It will be used to instruct
Experienced image data set inputs Deeplab network and Faster RCNN network respectively, iterates to restraining and save network
The weight distribution and bias of model;
Further, the step S3 specifically: input real-time collected Surface Picture to trained Deeplab
Network, referring to fig. 2, original image input by multiple convolutional layer and pond layer obtain a characteristic image, in order to obtain and
The output image for exporting image same size, is amplified characteristic image by deconvolution, finally using the condition that is fully connected with
The ability of model capture details is improved on airport (CRF), ensure that the segmentation of the pixel scale of land and the water surface.For acquisition
Semantic segmentation is as a result, obtain the pixel coordinate value at water surface line of demarcation by image procossing, and by the pixel at water surface line of demarcation
Coordinate set is transferred to information aggregators.
Further, the step S4 specifically: input real-time collected Surface Picture to trained Faster
RCNN network passes sequentially through shared convolutional layer, RPN network, the pond ROI layer and full articulamentum, final output by propagated forward
Barrier existing for input picture is divided into two class of ship and floating material, calculates Faster by the object detection results of image
Friendship and ratio between the prediction block of RCNN network output and the water-surface areas of image, semantic segmentation network output, for being classified as
The prediction block of floating material, given threshold 0.8, the result lower than this threshold value will be given and reject;For being classified as the prediction of ship
Frame, given threshold 0.1, the result lower than this threshold value will be given and reject;;
Further, the step S5 is specially to use gridiron pattern standardization, is chosen respectively in different angle different location
Several angle points on gridiron pattern determine the coordinate in the camera coordinates system of these angle points, the coordinate in world coordinate system and thunder
Up to the coordinate in coordinate system, respective coordinates are substituted into the mathematical model of camera calibration and combined calibrating, simultaneous solution obtains phase
Three rotation parameters (spin matrix), three translation parameters (translation matrix) in machine-radar fix transfer equation and one
Scale factor and spin matrix and translation matrix in camera-world coordinates transfer equation, so that it is determined that coordinate conversion side
The concrete form of journey.
Further, the step S6 specifically: in information aggregators, according to laser radar coordinate system and camera
The point cloud coordinate that laser radar obtains is converted to camera coordinates, then passes through camera coordinates system and picture by the transfer equation of coordinate system
A cloud is projected to imaging plane by the transformational relation between plain coordinate system, so that image has depth information.Finally will
The pixel coordinate information and depth information of the water surface boundary line of the prediction block and Deeplab model output of Faster RCNN output
Combine generation three-dimensional coordinate, corresponding world coordinates is converted to according to the Camera extrinsic that camera calibration obtains, so that it is determined that
The specific location of barrier and water surface line of demarcation in world coordinate system.
A kind of unmanned equipment multi-source sensory perceptual system of the water surface based on ROS, including perceive and apply two parts:
Perception part establishes point cloud information by the Node Mechanism of ROS and handles node, Image Information Processing node and letter
Cease aggregators.Image Information Processing intra-node contains two convolutional network moulds of Faster RCNN and Deeplab model
Type, image is handled by convolutional neural networks can be with the pixel coordinate information of acquired disturbance object prediction block and water surface boundary line, should
Information is transferred to the processing that information aggregators wait next step by the topic subscribing mechanism of ROS;Point cloud information handles node
Point cloud information is converted into the standard coordinate format under laser radar coordinate system, and will point cloud coordinate by topic communication mechanism
Information is transferred to information aggregators.In information aggregators, according to the conversion side of laser radar coordinate system and camera coordinates system
A cloud coordinate is converted to camera coordinates by journey, then by the transformational relation between camera coordinates system and pixel coordinate system, by point
Cloud projects to imaging plane, so that image has depth information, thus to obtain the three-dimensional coordinate of image;It finally will be according to camera
Image three-dimensional coordinate is converted to corresponding world coordinates by outer ginseng, so that it is determined that barrier and water surface line of demarcation are sat in the world
Specific location in mark system.
Application obscure portions include ROS different type functional node, including avoidance node, tracking node and path planning node
Deng.Avoidance node obtains the world coordinates of barrier and water surface boundary line by topic that subscription information aggregators are issued
Information, and vector field histogram is established by VFH+ obstacle avoidance algorithm, present feasible avoidance side can be determined by the histogram
To.Tracking node obtains video sequence and forecasting-obstacle frame in image by subscribing to image topic and target detection topic
On pixel coordinate information, by manual frame select determine tracking target after, activate CF target tracking algorism, calculated by the tracking
Coordinate information of the target in each frame image can be selected by output box in real time after the characteristic matching and filtering processing of method, thus real
Existing following function.Path planning node subscribes to semantic segmentation topic and information and merges topic, by segmented image obtain the water surface and
Then barrier pixel coordinate obtains its rough world coordinates information further according to information fusion topic, can be with by the information
A local map is established, RRT searching algorithm is used on this map, obtains the feasible pass of current local map.
Compared with the prior art, the invention has the following advantages and beneficial effects:
The present invention realizes the extraction of water surface boundary line using Deeplab network model, with traditional sea horizon detection method phase
Than changed by aquatic environment influenced it is smaller, have better system generalization ability, both adapt to the sea with obvious linear feature
Antenna detection is also applied for the coastline Detection Method of seashore geometrical characteristic complexity;Hindered using Faster RCNN network model
Hinder the coarse extraction of object candidate frame, and is merged with the three dimensional point cloud that laser radar obtains in real time, it can be in sensing parameter
It is realized in the case where redundancy detection to barrier more accurately three-dimensional description;By the distributed communication mechanism of ROS, can protect
Card sensing fuse information is perceived system after updating at the first time and obtains and handled in real time;Pass through camera and three-dimensional laser
As a result, establishing the corresponding relationship between visual identity result and world coordinates, be that the water surface is unmanned fills combined calibrating between radar
Standby subsequent intelligent decision making provides prior information.Multi-source cognitive method proposed by the present invention and system realize the water surface nobody
Equip complete description to aquatic environment key message, be widely portable to the intelligent navigation of the unmanned equipment of the various waters surface with
Control.
Detailed description of the invention
Fig. 1 is a kind of method flow diagram of the unmanned equipment multi-source cognitive method of water surface;
Fig. 2 is the Deeplab network architecture based on VGG16 in embodiment;
Fig. 3 is the Faster RCNN network architecture based on AlexNet in embodiment;
Fig. 4 is a kind of unmanned equipment multi-source sensory perceptual system schematic diagram of water surface based on ROS.
Specific embodiment
Present invention will now be described in further detail with reference to the embodiments and the accompanying drawings, but embodiments of the present invention are not
It is limited to this.
Embodiment:
Referring to Fig. 1, the unmanned equipment multi-source cognitive method of the water surface is planted, comprising the following steps:
Step 10 acquires the sensing parameter of the unmanned equipment multi-source sensory perceptual system of the water surface in real time, obtains the vision of Surface Picture
The three-dimensional point cloud information of information and aquatic environment;
Step 20 manually demarcates preparatory collected Surface Picture, using the data set demarcated to Deeplab
Model and Faster RCNN model are trained and save network model parameter;
The Surface Picture inputted in real time is divided into background, land and water surface three classes by Deeplab model by step 30,
Water surface boundary line is extracted according to the circumference of water-surface areas;
Step 40 extracts the prediction block of water hazard object by Faster RCNN network model, calculates separately ship and drift
Friendship and ratio between floating object prediction block and the water-surface areas of image, semantic segmentation network output, reject meaningless obstacle quality testing
Survey result;
Step 50 carries out camera calibration, obtains camera internal reference and outer ginseng, then carries out the connection of three-dimensional laser radar and camera
Calibration is closed, obtains the coordinate transformation relation between radar and camera in conjunction with calibration result;
The three dimensional point cloud that laser radar obtains is projected to the figure that camera obtains according to coordinate transformation relation by step 60
As upper, depth information is added to image, then barrier is finally obtained by the conversion of camera coordinates system-world coordinate system coordinate
With the world coordinates of water surface boundary line.
Above-mentioned steps 20 are specifically included Pixel-level is demarcated as background, land to collected Surface Picture from top to bottom in advance
Three major class in ground and the water surface, for the training of Deeplab network model.Barrier candidate frame in Surface Picture is demarcated as ship
Only and two major class of floating material, trained for Faster RCNN network model, to construct Surface Picture data set.It will be used for
Trained image data set inputs Deeplab network and Faster RCNN network respectively, iterates to restraining and save net
The weight distribution and bias of network model;
Above-mentioned steps 30 specifically include input in real time collected Surface Picture to trained Deeplab network, referring to
Fig. 2, original image input extract characteristics of image by convolutional layer and obtain corresponding characteristic pattern, then pass through pond layer compression again
Characteristic pattern extracts main feature, after the feature extraction of multilayer convolutional layer and pond layer and the Feature Compression, can get deep layer
Secondary characteristic pattern.Deeplab is by being changed to no down-sampling pond layer for the pond layer of the 4th layer and layer 5, it is ensured that special
The size of sign figure remains unchanged, and at the same time, this two layers of subsequent convolutional layer of pond layer is changed to empty convolutional layer, to protect
The neuron receptive field of card Chi Huahou does not change.Finally, it is big that characteristic image is amplified to original input image by deconvolution
It is small, the ability of model capture details is then improved using the condition random field (CRF) being fully connected, and ensure that land and the water surface
Pixel scale segmentation.Semantic segmentation for acquisition is sat as a result, obtaining the pixel at water surface line of demarcation by image procossing
Scale value, and the pixel coordinate collection at water surface line of demarcation is transferred to information aggregators.
Deeplab network model is based on VGG16 and is constructed, and removes adopt under VGG16 most latter two pond layer first
Then the two subsequent convolution kernels of pond layer are changed to empty convolution by sample, finally replace three of VGG16 full articulamentums
For convolutional layer, the full convolutional coding structure of Deeplab model is realized.In order to obtain the output of size identical as original image, using deconvolution
Method deconvolution is carried out to the characteristic pattern that obtains behind pond and process of convolution, to obtain one big with input image size
Small identical segmented image finally carries out details optimization to land and water segmented image using full connection Stochastic Conditions field, to obtain
The fine segmented image of one water surface border-line edge.
Above-mentioned steps 40 specifically include input in real time collected Surface Picture to trained Faster RCNN network,
Faster RCNN network model is based on AlexNet convolutional neural networks and is constructed, specifically by Fast RCNN network and RPN net
Network two large divisions constitute, wherein the shared convolutional layer of Fast RCNN network and RPN network by AlexNet first five layer of convolution mind
It is constituted through network, the third pond layer of AlexNet is revised as the pond ROI layer, retains two layers of full articulamentum of AlexNet, will
The last layer Softmax classifier is revised as selecting the linear regressor of water hazard object for frame and for ship and floating material
Linear regressor+Softmax classifier the layer of classification.Referring to Fig. 3, Surface Picture extracts original image by shared convolutional layer first
The characteristic pattern of original image is sent into RPN network structure thereafter by the characteristic pattern of picture.For sharing the characteristic pattern of convolutional layer output,
Convolution sliding is carried out by the convolution kernel of 3*3 and generates sliding window, and generates 9 in the central point of each sliding window
Anchor point frame.According to the mapping relations between sliding window and original image characteristic pattern, each anchor point frame can be obtained from original image
Characteristic pattern, by these characteristic patterns by propagated forward enter full articulamentum generate feature vector.Then by feature vector
Softmax classifier and linear regressor are respectively fed to carry out target classification and positioning.Anchor point frame is simplified, selection region obtains
Dividing high anchor point frame is suggestion areas.By suggestion areas and original image characteristic pattern that RPN network exports while the pond ROI layer is inputted,
The characteristic pattern for extracting suggestion areas corresponding position enters full articulamentum by propagated forward and generates feature vector, finally by
Softmax classifier and linear regressor generate the target prediction frame after final territorial classification score and recurrence, thus by defeated
Enter barrier existing for image and is divided into two major class of ship and floating material.Calculate Faster RCNN network output prediction block with
Image, semantic divides the friendship between the water-surface areas of network output and ratio, for being classified as the prediction block of floating material, given threshold
It is 0.8, the result lower than this threshold value will be given and reject;For being classified as the prediction block of ship, given threshold 0.1 is lower than this
The result of threshold value will be given and reject;
Above-mentioned steps 50 are specifically included using gridiron pattern standardization, choose gridiron pattern respectively in different angle different location
On several angle points, determine the coordinate in the camera coordinates system of these angle points, the coordinate in world coordinate system and radar fix
Coordinate in system substitutes into respective coordinates in the mathematical model of camera calibration and combined calibrating, and simultaneous solution obtains camera-thunder
Up to three rotation parameters (both spin matrixs), three translation parameters (both translation matrix) and the ruler in coordinate transfer equation
The factor and spin matrix and translation matrix in camera-world coordinates transfer equation are spent, so that it is determined that coordinate transfer equation
Concrete form.
Above-mentioned steps 60 are specifically included in information aggregators, according to laser radar coordinate system and camera coordinates system
The point cloud coordinate that laser radar obtains is converted to camera coordinates, then passes through camera coordinates system and pixel coordinate system by transfer equation
Between transformational relation, a cloud is projected into imaging plane so that image have depth information.It is finally that Faster RCNN is defeated
The pixel coordinate information and depth information of the water surface boundary line of prediction block and the output of Deeplab model out combine generation
Three-dimensional coordinate is converted to corresponding world coordinates according to the Camera extrinsic that camera calibration obtains, so that it is determined that barrier and the water surface
Specific location of the line of demarcation in world coordinate system.
Referring to fig. 4, the unmanned equipment multi-source sensory perceptual system of a kind of water surface based on ROS, ROS message processing module include sense
Know and apply two parts.
Perception part establishes three nodes by the Node Mechanism of ROS, is point cloud information processing node, image respectively
Information processing node and information aggregators.Image Information Processing intra-node contains Faster RCNN and Deeplab mould
Two convolutional network models of type, image is handled by convolutional neural networks can be with acquired disturbance object prediction block and water surface boundary line
Pixel coordinate information, the information by the topic subscribing mechanism of ROS be transferred to information aggregators wait next step place
Reason;Point cloud information handles node and point cloud information is converted to the standard coordinate format under laser radar coordinate system, and passes through topic
A cloud coordinate information is transferred to information aggregators by communication mechanism.In information aggregators, according to laser radar coordinate system and
A cloud coordinate is converted to camera coordinates by the transfer equation of camera coordinates system, then by camera coordinates system and pixel coordinate system it
Between transformational relation, a cloud is projected into imaging plane so that image have depth information, thus to obtain image three-dimensional sit
Mark;Finally image three-dimensional coordinate will be converted into corresponding world coordinates according to Camera extrinsic, so that it is determined that barrier and water
The specific location in world coordinate system in face line of demarcation.
Application obscure portions include ROS different type functional node, including avoidance node, tracking node and path planning node
Deng.Avoidance node obtains the world coordinates of barrier and water surface boundary line by topic that subscription information aggregators are issued
Information, and vector field histogram is established by VFH+ obstacle avoidance algorithm, present feasible avoidance side can be determined by the histogram
To.Tracking node obtains video sequence and forecasting-obstacle frame in image by subscribing to image topic and target detection topic
On pixel coordinate information, by manual frame select determine tracking target after, activate CF target tracking algorism, calculated by the tracking
Coordinate information of the target in each frame image can be selected by output box in real time after the characteristic matching and filtering processing of method, thus real
Existing following function.Path planning node subscribes to semantic segmentation topic and information and merges topic, by segmented image obtain the water surface and
Then barrier pixel coordinate obtains its rough world coordinates information further according to information fusion topic, can be with by the information
A local map is established, RRT searching algorithm is used on this map, obtains the feasible pass of current local map.
The above embodiment is a preferred embodiment of the present invention, but embodiments of the present invention are not by above-described embodiment
Limitation, it is other it is any without departing from the spirit and principles of the present invention made by change, modification, substitution, combination, letter
Change, should be equivalent substitute mode, be included within the scope of the present invention.