CN112509056B - Dynamic battlefield environment real-time path planning system and method - Google Patents

Dynamic battlefield environment real-time path planning system and method Download PDF

Info

Publication number
CN112509056B
CN112509056B CN202011373402.8A CN202011373402A CN112509056B CN 112509056 B CN112509056 B CN 112509056B CN 202011373402 A CN202011373402 A CN 202011373402A CN 112509056 B CN112509056 B CN 112509056B
Authority
CN
China
Prior art keywords
real
map
frame
feature
current frame
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011373402.8A
Other languages
Chinese (zh)
Other versions
CN112509056A (en
Inventor
张西山
连光耀
李会杰
闫鹏程
孙江生
袁祥波
连云峰
代冬升
李雅峰
李万领
钟华
付久长
王子林
宋秦松
郭伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
32181 Troops of PLA
Original Assignee
32181 Troops of PLA
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 32181 Troops of PLA filed Critical 32181 Troops of PLA
Priority to CN202011373402.8A priority Critical patent/CN112509056B/en
Publication of CN112509056A publication Critical patent/CN112509056A/en
Application granted granted Critical
Publication of CN112509056B publication Critical patent/CN112509056B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • G06T7/85Stereo camera calibration
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/005Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 with correlation of navigation data from several sources, e.g. map or contour matching
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/20Instruments for performing navigational calculations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/05Geographic models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • G06T2207/10012Stereo images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Remote Sensing (AREA)
  • General Physics & Mathematics (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Geometry (AREA)
  • Automation & Control Theory (AREA)
  • Biophysics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Mathematical Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Graphics (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a dynamic battlefield environment real-time path planning system and a method, wherein the dynamic battlefield environment real-time path planning system comprises a map reconstruction module, a visual perception module, an intelligent decision module and a UI interaction module; the UI interaction module realizes mutual calling and information connection of parameters among different modules; the map reconstruction module takes an image shot by the unmanned aerial vehicle as input to generate a real-time digital earth surface model; the visual perception module is used for extracting geographic entities existing in images shot by the unmanned aerial vehicle, generating a grid map from the real-time digital earth surface model, and setting barrier marks on the grid map; the intelligent decision module is combined with a DSAS algorithm to find an optimal path which can bypass the barrier to reach the destination from the beginning of the environment, so that real-time path planning is realized. The invention can provide high-real-time command control information for the unmanned equipment in a dynamic battlefield environment, thereby improving the autonomy capability of the unmanned equipment.

Description

Dynamic battlefield environment real-time path planning system and method
Technical Field
The invention relates to the field of paths, in particular to a dynamic battlefield environment real-time path planning system and a dynamic battlefield environment real-time path planning method.
Background
The information plays an increasingly important role in daily life, the efficiency and the quality of work can be obviously improved, and the information is used as an intelligent control system of unmanned equipment and can greatly improve the flexibility of the system through a flexible unmanned flight platform. With the rapid development of electronic and computer technologies in recent years, low-cost unmanned aerial vehicles have become mature, high-speed image processing is gradually applied to various tasks, and the realization of efficient intelligent control through the low-cost unmanned aerial vehicles becomes a flexible and efficient means. However, the traditional unmanned aerial vehicle map software system does not have a real-time map building function and an autonomous capability, and is low in processing speed, so that a real-time path planning system with the real-time map building function and the high processing speed is required.
Disclosure of Invention
In order to solve the problems, the invention provides a dynamic battlefield environment real-time path planning system and a dynamic battlefield environment real-time path planning method, which can provide high-real-time command control information for unmanned equipment in a dynamic battlefield environment, so that the autonomy capability of the unmanned equipment is improved.
In order to realize the purpose, the technical scheme adopted by the invention is as follows:
a dynamic battlefield environment real-time path planning system, comprising: the system comprises a map reconstruction module, a visual perception module, an intelligent decision module and a UI interaction module;
the UI interaction module realizes mutual calling and information connection of parameters among different modules;
the map reconstruction module takes an image shot by the unmanned aerial vehicle as input to generate a real-time digital earth surface model;
the visual perception module is used for extracting geographic entities existing in images shot by the unmanned aerial vehicle, generating a grid map from the real-time digital earth surface model, and setting barrier marks on the grid map;
the intelligent decision module is combined with a DSAS algorithm to realize that an optimal path which can bypass the barrier is searched from the beginning of the environment to the end point, and real-time path planning is realized.
Optionally, the map reconstructing module includes: the visual SLAM takes aerial images shot by an unmanned aerial vehicle on line as input, generates 3D point cloud for each frame of image in real time, generates local DSM in real time by combining the 3D point cloud with texture information of the image, and fuses the local DSM to the global DSM to generate a real-time digital earth surface model.
Optionally, the visual SLAM system includes feature extraction, tracking, mapping, and looping;
the feature extraction is used for extracting feature points in the aerial image and calculating a descriptor;
the tracking completes the feature matching by calculating the distance between the feature point descriptors of the current frame and the previous frame, and then recovers the pose change of the current frame relative to the previous frame through the feature matching;
searching more feature matches in the current frame and the previous frames, recovering more 3D points through triangulation, and performing local BA optimization on the current frame, the commonly observed associated frame and the commonly observed map point to acquire more accurate camera pose and 3D map point coordinates;
the loop is used for detecting whether the camera returns to a place which is previously visited or not, and adding new constraint according to the detected loop to correct the corresponding camera pose and the 3D map point coordinate.
Optionally, the image shot by the unmanned aerial vehicle is used as input to generate a real-time digital earth surface model, a target detection algorithm is used to extract geographic entities existing in the image, a pre-trained AlexNet convolutional neural network is used to extract depth level features of the geographic entities, dimension reduction processing is performed on feature vectors of each geographic entity to generate a grid map, barrier marks are set at corresponding positions in the grid map according to the types of the ground objects, starting from the initial place of the environment and finding an optimal path capable of bypassing the barriers to reach the destination by combining with a DSAS algorithm are achieved, and real-time path planning is achieved.
Optionally, the generating the real-time digital earth surface model by using the image shot by the unmanned aerial vehicle as an input includes using a visual SLAM to use an aerial image shot by the unmanned aerial vehicle on line as an input, generating a 3D point cloud for each frame of the aerial image in real time, removing noise points of the 3D point cloud by iterative filtering, performing delaunay triangulation on feature points of the 3D point cloud after the noise points are filtered to generate a 2D mesh, projecting the generated 2D mesh into the 3D point cloud after the noise points are filtered to generate a 3D mesh, fusing texture information of the 3D mesh into the aerial image to generate a local DSM, dividing the local DSM into tiles, and fusing the local DSM into the global DSM by using a multiband algorithm to generate the real-time digital earth surface model.
Optionally, the extracting, by using the target detection algorithm, the geographic entity present in the image includes dividing the input image into S × S grids, each grid being responsible for detecting all objects falling into its own region, where each grid predicts 5 values, and predicts coordinates (x, y) of a center position of a prediction box of the obtained object, width and height (w, h) of the object, and confidence that the object belongs to a certain class, where if the prediction box contains a detection object, pr is 1, otherwise, pr is 0, and in the prediction stage, the probability of the class is a product of a class conditional probability and the predicted confidence:
Figure GDA0002929923620000031
optionally, the geographic entity comprises a relatively stationary geographic entity.
In the DSAS algorithm, besides coordinate information representing the space position of the unmanned aerial vehicle, pointers pointing to a father node and all child nodes of the unmanned aerial vehicle are also included, when a new obstacle in front of the current flight path is detected, the DSAS algorithm deletes the flight path nodes threatening the coverage area and all descendant nodes of the flight path nodes, reinserts the father nodes into an OPEN table, calculates the cost value of a new expansion node, and continues the searching process until a new optimal flight path of the unmanned aerial vehicle is obtained
Compared with the prior art, the invention has the technical progress that:
compared with the traditional computer vision method, the method has the advantages that the processing capacity is high, the real-time processing capacity is realized, more than 30frames of images can be processed per second, the map data can reach more than ten million point clouds, and high real-time command control information can be provided for the unmanned equipment in a dynamic battlefield environment, so that the autonomous capacity of the unmanned equipment is improved.
Drawings
The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this specification, illustrate embodiments of the invention and together with the description serve to explain the principles of the invention and not to limit the invention.
In the drawings:
FIG. 1 is a block diagram of the present invention.
Fig. 2 is a diagram of a descriptor of SIFT keypoints.
FIG. 3 is a schematic diagram of a matrix in BundleAdjustment.
FIG. 4 is a flow chart of the present invention.
Fig. 5 is a specific network structure diagram of the YOLO model.
Fig. 6 is a specific network flow diagram of the YOLO model.
Detailed Description
These several specific embodiments may be combined with each other below, and details of the same or similar concepts or processes may not be repeated in some embodiments. Embodiments of the present invention will be described below with reference to the accompanying drawings.
As shown in FIG. 1, the present invention discloses a dynamic battlefield environment real-time path planning system, which comprises.
The system comprises a map reconstruction module, a visual perception module, an intelligent decision module and a UI interaction module;
each module of the invention is decoupled, and the parameters among different modules are mutually called and connected with information through the UI interaction module. The UI interface module is a part for interacting with a user and comprises several functions of building a project, loading the project, detecting a target and training a model, and in the newly built project, the user can determine the name of the project, the position of the project, a picture sequence, a label category and a used detection algorithm. A load item is an item previously created by a load user. And finishing the result of the constructed image by target detection. Training the model to retrain the model using the corrected labels. And right clicking the display path planning list in the interface, wherein the UI interface supports the user to modify the incorrect prediction result, and the corrected label data is used for automatically updating the parameter weight of the model again, so that the precision of the model is continuously improved in the use process, and each module of the system can be independently called and tested by virtue of the modular design, and new modules can be conveniently and continuously added.
The map reconstruction module takes an image shot by the unmanned aerial vehicle as input to generate a real-time digital earth surface model;
the visual perception module is used for extracting geographic entities existing in images shot by the unmanned aerial vehicle, generating a grid map from the real-time digital earth surface model, and setting barrier marks on the grid map;
the intelligent decision module is combined with a DSAS algorithm to realize that an optimal path which can bypass the barrier is searched from the beginning of the environment to the end point, and real-time path planning is realized.
Specifically, the map reconstruction module includes: the visual SLAM takes aerial images shot by an unmanned aerial vehicle on line as input, generates 3D point cloud for each frame of image in real time, generates local DSM in real time by combining the 3D point cloud with texture information of the image, and fuses the local DSM to the global DSM to generate a real-time digital earth surface model.
Wherein, the vision SLAM system includes: extracting, tracking, drawing and looping the features;
the feature extraction is used for extracting feature points in the aerial image and calculating a descriptor;
tracking and completing feature matching by calculating the distance between feature point descriptors of a current frame and a previous frame, and recovering the pose change of the current frame relative to the previous frame through the feature matching;
establishing a map, searching more feature matches in a current frame and previous frames, recovering more 3D points through triangulation, and performing local BA optimization on the current frame, a commonly observed associated frame and commonly observed map points to obtain more accurate camera pose and 3D map point coordinates;
the loop is used for detecting whether the camera returns to a place which is previously visited or not, and a new constraint is added according to the detected loop to correct the corresponding camera pose and the 3D map point coordinate.
Specifically, the unmanned aerial vehicle carries a camera to acquire aerial images at the speed of 30frames/s and transmits the aerial images to the ground station in real time. To recover the pose of the camera and the 3D structure of the environment from the aerial images, feature points in the images, which are images of landmark points in the environment, are extracted, and from the perspective of the image, the feature points are representative parts of the images.
After the feature points are detected, whether the two feature points in the two frames are images formed by the unified landmark points is judged through feature point matching, so that the position information of the feature points is not enough to be simply detected, and other information is required to be extracted for matching between the feature points. The feature points generally consist of two parts: key points (Key-points) containing information such as position, size, direction, score, etc.; descriptor (Descriptor), pixel information around the keypoint.
In order to keep the rotational invariance, when the feature point descriptor is calculated, the direction information of the key points is added, the detection of the feature points is actually the detection of the key points, and common detection methods include Harris key points, FAST key points, SIFT key points and SURF key points. The FAST keypoint algorithm is simple, and if there are N consecutive pixels in 16 pixels on a circle with a radius of 3 and a pixel p as a center, the point is a keypoint, where N is usually 12 or 9, if the N is greater than or less than a certain threshold of the value of the p pixel. The FAST key points have no direction, the FAST algorithm is improved in the ORB characteristics, and an image pyramid is respectively constructed to realize the acquisition of angle information by a scale invariance method and a gray scale centroid method. In this embodiment, a Speeduprobubustfeature (SURF) key point detection method is used, and since the construction of the DoG pyramid in the SIFT key point detection and the search of the extremum are time-consuming, SURF improves the extremum detection method of SIFT, and the position of the extremum point is located by using the local maximum of the hessian matrix determinant, thereby greatly accelerating the detection speed of the key point.
The detection method of the descriptor in the embodiment uses a SIFT feature descriptor, the SIFT feature descriptor is selected by taking a feature point as a center, as shown in fig. 2, a coordinate axis is rotated by an angle of a principal direction of the feature point in a nearby neighborhood, an 8 × 8 window is taken by taking a key point as the center, every 4 × 4 pixel is taken as a window, a gradient amplitude and a gradient direction of each pixel are obtained, an arrow direction represents the gradient direction of the pixel, a length represents the gradient amplitude, then a gaussian window is used for carrying out weighting operation on the gradient amplitude, finally gradient histograms in 8 directions are drawn on each 4 × 4 small block, an accumulated value in each gradient direction is calculated, vector information in 8 directions is formed, and finally the descriptor is composed of floating point type vectors in 4 × 4 × 8.
The SIFT features have invariance such as scale, rotation, brightness and the like, and are extremely stable feature points, but the calculation speed is much slower than SURF, the real-time performance of the system cannot be guaranteed by using the SIFT features, but the SIFT can completely reach the real-time speed under the acceleration of the GPU.
The feature matching is used for matching feature points of the same road sign on different images. The matching of the feature points is obtained by calculating the distances of different feature point descriptors, and the feature points of a certain feature point on the first frame image, which are matched on the second frame image, are calculated. The SIFT feature point descriptors are 128-dimensional floating point type vectors, and the distance between the two feature point descriptors is calculated to be the vector distance between the two feature point descriptors. When the number of feature points is large, the speed of calculating the distance one by one is slow, and real-time matching cannot be achieved. In order to achieve real-time matching speed, camera motion can be predicted according to a motion model to obtain initial camera pose change, and epipolar search is performed according to the initial pose change to accelerate matching feature points. Bag of words (BOW) or fast nearest neighbor (FLANN) algorithms may also be employed to speed up matching. Map initialization when no map point cloud is known, the map point cloud is generated by using the first two frames. The basic matrix and the homography matrix can be calculated by utilizing the matched characteristic points between the two frames, and the pose transformation between the two frames can be calculated by further decomposing the basic matrix and the homography matrix. Due to the existence of noise and mismatching, when the basis matrix and the homography matrix are calculated, RANSAC (random sample consensus) algorithm is adopted, the homography matrix and the basis matrix are calculated at the same time, and the matrix with small projection error is selected to decompose the pose. In order to ensure the quality of the initialization map, a certain distance and a certain rotation angle are required between two frames. And when the number of matched interior points between the two frames is less, judging that the initialization is failed, and reselecting the two frames for initialization. And calculating the distance between the two frames according to the conversion of the GPS coordinates between the two frames to a ground-fixed earth-center rectangular coordinate system (Earth holder Earth fixed) and taking the distance as the initial scale of the map.
Map initialization is to recover 3D map points by using 2D-2D epipolar geometry, and after map points exist, the camera pose and the 3D map points can be recovered by solving the problem of 3D-2D PnP. Although the algorithms such as PnP can solve the pose of the camera and acquire the depth of the feature point by triangulation, the pose and depth information currently solved are both rough values due to the existence of noise. In order to obtain more accurate pose and depth values, the current pose and depth values can be used as initial values, and the accurate pose and depth values are further obtained by using a nonlinear optimization method.
The mathematical expression for imaging of a known pinhole camera is:
Figure GDA0002929923620000071
pi [ Xi, yi, zi ] is the three-dimensional coordinate of the first map point, pi = [ ui, vi ] is the pixel coordinate of the projection of the ith map point on the current frame, si is the depth value corresponding to Pi, K is the camera internal reference of the current frame, and T is the pose of the current frame. T can be written in the form of lie algebra:
Figure GDA0002929923620000072
in practice, due to the presence of noise, an error ei exists between the true projected pixel coordinate ui of Pi and the theoretical value Pi. The true pixel coordinates with noise are here recorded as the observed values. When the error satisfies the gaussian distribution:
Figure GDA0002929923620000081
a least squares problem can be constructed, namely:
Figure GDA0002929923620000082
let the derivative of ei with respect to ξ be Jacobian J (ξ). The most important of the gradient descent method in the nonlinear optimization is the derivation of J (ξ). Here the world coordinate system is first down-converted to the camera coordinate system as follows:
C i =exp(ξ^)P i =TP i =[X′ i ,Y′ i ,Z′ i ] T
the camera model may be represented as:
Figure GDA0002929923620000083
here, [ u ', v' ] T is a specific value of the predicted value of the pixel coordinate. Elimination of si from si = Z' gives:
Figure GDA0002929923620000084
Figure GDA0002929923620000085
in the formula of the formula error term, ci is taken as an intermediate variable, and a chain derivation method is used to obtain:
Figure GDA0002929923620000086
wherein:
Figure GDA0002929923620000091
the partial derivative of Ci with respect to xi is equivalent to derivation of lie algebra, and can be completed by constructing a disturbance model. Multiplying the disturbance quantity delta xi by xi left, and then solving the derivative of the increment of TPi to the disturbance item delta xi:
Figure GDA0002929923620000092
the partial derivative of the rearrangement error term with respect to xi can be derived as:
Figure GDA0002929923620000093
the negative sign in front of the jacobian matrix here is due to the fact that the error in the derivation process is defined by the observed value minus the predicted value. Of course, if the error is defined as a predicted minus observed value, the negative sign of the foregoing may be omitted. Carrying out cholesky decomposition on sigma-1 in the formula, namely:
-1 =LL T =L T L
bringing the above formula into available:
Figure GDA0002929923620000094
the formula for the least squares method can then be written as:
Figure GDA0002929923620000095
when iterative optimization is performed, a gaussian optimization algorithm is taken as an example here for simplicity. From the foregoing theoretical description, it can be seen that the idea of Gaussian is to perform a first order Taylor expansion on f (x). Taylor expansion is performed for the error term:
Figure GDA0002929923620000101
where J (ξ) is the derivative of e (ξ) with respect to ξ (Jacobian), which can be:
Figure GDA0002929923620000102
the single two-norm is expanded to obtain:
Figure GDA0002929923620000103
finding the derivative of the above equation with respect to ξ equal zero can give:
Figure GDA0002929923620000104
an incremental equation can be obtained:
Figure GDA0002929923620000105
the increment delta xi can be solved by an increment equation, so that iterative optimization is realized. When the Gaussian Newton method or Levenberg-Marquardt method is used again, the loss function is L (ui-1/siexp (ξ) Pi), and the Jacobian is LJ (ξ).
The key frame is a frame selected from the local normal frames as a representative of the local frame. Trying to capture the same image when the camera is stationary. Needless to say, if all images are taken as the objects of the back-end optimization, unnecessary calculations are added. Therefore, the key frame's primary role is to face a compromise between back-end optimization computational power and accuracy. The key frame extraction generally considers the following aspects:
whether the number of frames temporally distant from a previous key frame is enough;
whether the distance in space from the previous key frame is sufficiently far;
the observation quality is controlled to a certain range with the common observation point of the last key frame. Too many co-observations indicate a higher overlap rate, and too few will reduce the tracking accuracy.
In this embodiment, in addition to considering the above-mentioned 3 points, special requirements are made on the rotation of the aerial image. Mainly because unmanned aerial vehicle (camera) yaw angle changes when (around the z axle) rotatory, the image of taking photo by plane that shoots changes by a short time. Therefore, when the pitch angle and the yaw angle of the unmanned aerial vehicle (camera) exceed a certain angle, the key frames are added, and when only the yaw angle changes, the key frames are not added.
If the current frame tracks the previous key frame, the tracking loss phenomenon can occur due to the reasons that the detected feature points are few, the matched feature points are few, or the calculated pose deviation is large and the like. Relocation after a loss is crucial in order to improve the redundancy of the whole system. In the system herein, when GPS is available, the GPS is used to find several frames closer to the current frame as candidate frames. And then finding a key frame with better feature matching from the candidate frames, and solving the PnP again to track and calculate the pose. When the GPS is not available, a bag of words (BOW) is used to find candidate frames, and a keyframe with better feature matching is also found from the candidate frames for tracking.
The bag of words was originally used for text classification, representing text as feature vectors. Later, word bags were used in computer vision for image classification and image retrieval, etc. Taking the SIFT features as an example, firstly extracting a large number of SIFT features from an image, then classifying the features layer by layer according to the distance of the 128 floating point type descriptors of the features by using a clustering method such as Kmeans and the like, and taking the class at the bottommost layer as a word in a dictionary. And outputting the frequency of the word in each image statistical dictionary as a vector, and completing the conversion from the image to the vector. By comparing the vector distance between different images, the similarity between the images can be approximately calculated. A few frames with a small vector distance from the current frame may be candidate frames for relocation.
In the map initialization process, the pose is obtained by utilizing two frames of epipolar constraints, and then an initial map point cloud is generated through triangulation. In the tracking process, the pose of a new frame (current frame) can be obtained by continuously solving the PnP through the matched feature points of the current frame and the map point cloud and optimizing the PnP. However, as the camera is far away from the area where the map point cloud is initialized, the feature points of the new frame are less and less matched with the map point cloud, and finally, the PnP cannot be solved and the tracking is lost. Therefore, to ensure the continuity of the system, new map points are continuously added after the current frame is calculated. Since the map point creation and key frame addition conditions are relaxed, a very strict screening mechanism is used to remove redundant key frames and map points which are not easy to track.
And in the tracking process, one part of all the feature points of the current frame is used for matching the feature points of the previous key frame, and then the PnP is solved by using the map points corresponding to the feature points of the previous key frame and the feature points of the current frame. So that the feature points matching the previous key frame already have corresponding map points, and new map points are generated from these feature points not matching the map points. Feature points that do not have matching map points match feature points in keyframes that have a common view relationship with the current frame (except for the last keyframe utilized in the tracking process). And calculating epipolar lines of the feature points on the matched key frames by using an epipolar constraint equation, and judging whether the matching is correct or not by calculating whether the matched feature points are near the epipolar lines. If the matched feature point is more than a certain threshold away from the epipolar line, a mismatch is identified and re-matched in other keyframes. Since the current frame needs to be matched with the features of a plurality of co-view key frames, the matching time is long due to large calculation amount. To speed up matching, using the bag of words introduced above, matching is only done in features that have the same word label in the co-view keyframes. After the matched feature points are obtained, new map points can be calculated by utilizing triangulation.
Because the adding condition of the map points and the key frames is looser, the redundant key frames and the map points with low confidence coefficient are deleted, so that the calculation amount of optimization can be reduced. Deletion judgment conditions of key frames: 90% of the observed map points for that key frame are observed simultaneously by more than three other key frames. Deleting conditions of map points: less than three keyframes are observed for the map point at the same time.
So far, only one-to-one constraint is considered for the pose of the current frame and the coordinate estimation of the new map point. The pose estimation of the current frame is obtained by calculating and solving PnP and optimizing projection errors through projecting map points to the current frame. The new map point is obtained by triangulation of feature points matched between the current frame and the common frame. In fact, after the map is updated, any map point can be observed by multiple frames together. Thus, the projection error of all map points onto the keyframe where they are observed can be jointly optimized. The optimization process is similar to BA, constructing a projection with least squares adding multiple frames compared to equations 2-100:
Figure GDA0002929923620000131
where χ ij represents whether the ith map point is visible in the j cameras. If it is seen that its value is 1, otherwise it is 0.θ = (ξ 1.. ξ M, P1.. PN) is the variable to be optimized, θ ∈ R9M +3N.
The incremental equation for the Levenberg-Marquardt optimization method is as follows:
[J(θ) T J(θ)+λI]δ0=-J(θ) T (n-f(0))
the Jacobian matrix J (theta) is a sparse matrix, as shown in FIG. 3 (a), which shows the Jacobian matrix when only 4 map points are observed by 3 key frames at the same time. Can be decomposed into two parts:
J(θ)=[J C ,J P ]
each small block in the JC is a derivative of a projection error relative to the pose of each key frame, and a matrix of 2 multiplied by 6 is obtained by an equation 2-98. In JP, each small block is a derivative of the projection error with respect to each map point, and is a 2 × 3 matrix, and the calculation formula can be obtained by a chain rule and then by substituting the above formula:
Figure GDA0002929923620000132
wherein
Figure GDA0002929923620000133
The derivative to Pi is then R. In addition, each small block and the left side of the reprojection error are all decomposed into an L matrix by the information matrix Σ in the above equation. The value of [ J (theta) TJ (theta) + Lambda I in the formula can be obtained from the Jacobian matrix J (theta)]The matrix is shown in FIG. 3 (b). The matrix is a symmetric sparse matrix, and can be partitioned into the following blocks according to the definition of a Jacobian matrix:
Figure GDA0002929923620000134
thus, the equation for optimization can be:
Figure GDA0002929923620000141
and (3) performing simultaneous multiplication Schur compensation on the left and right sides of the formula to realize marginalization:
Figure GDA0002929923620000142
the first equation in the equation set of the above formula is a linear equation, and all the camera pose increments δ C can be directly solved. All map point increments δ P can be solved by substituting δ C into the second expression of the system of equations. And the accurate camera pose and map point coordinates can be continuously and iteratively estimated by obtaining the increment of the optimization variable of each step. In order to avoid excessive calculation amount in the mapping part, only local optimization is carried out, namely, only the poses of the current frame and the key frames with the common-view relation and the coordinates of map points observed by the poses are optimized. In addition, GPS information is fused in the local optimization to be used as a constraint item so as to obtain a more accurate camera pose. The cost function is obtained by adding GPS constraint in the optimization method, and more accurate camera pose and point cloud coordinates can be obtained by optimizing the cost function.
Figure GDA0002929923620000143
And (4) optimizing all map points and key frames (global optimization) by opening an independent thread after detecting and correcting loop 2.5 and every certain frame number.
In the tracking process, the estimation of the pose of the camera is a gradual recursion process, namely the pose of the previous key frame is tracked to solve the pose of the current frame. In fact, each frame pose solution has errors, and the errors are accumulated as the tracking is performed. One effective way to eliminate the accumulated error is to perform loop detection. The loop detection is used to determine whether the camera has returned to a previous position, and if loop is detected, the constraint between the current frame and the loop frame can be added to the back-end optimization to correct the error.
When loop detection is actually performed, the most direct method is to match all previous key frames with the current frame, and the key frame with the best match can be used as a loop frame. However, with the continuous addition of key frames, the matching speed is slower and the calculation amount is larger. In order to reduce the calculation amount, when the GPS is available, the GPS is used to find several key frames closer to the current frame as candidate frames, and the current frame is matched with the several candidate frames, and the best matched frame is taken as a loop back frame. The bag of words approach is used to speed up the matching when GPS is not available. The method of the word bag is to search words corresponding to the descriptors of the picture feature points in the dictionary tree. And counting the number of the words to form a bag-of-words vector. The similarity between two pictures can be estimated by comparing the distance of the bag-of-word vectors between the images. And calculating the distance between the current frame and the bag-of-word vectors of all key frames, searching the frames with the minimum distance as candidate frames, and finally taking the best matched candidate frame as a loop back frame.
Today most drones provide more accurate GPS information and therefore methods for detecting loops using GPS are well suited. However, in the method using bags, when the environment itself has similar scenes, it is easy to detect that the wrong loop leads to optimizing the destructive result. Therefore, it is required to design a more rigorous selection strategy for loop candidate frames: 1) Loop detection is performed every 10 frames to prevent redundant calculation; 2) Calculating the maximum distance Dm of the word bag vectors of the current frame and all the connected key frames, wherein the distance between the word bag vectors of the candidate frame and the current frame is smaller than Dm; 3) And if 3 continuous frames continuously detect the candidate frame, the loop is considered to be detected. And matching the 3 rd frame with the loop candidate frame, and searching the key frame with the best matching as a loop frame. The matching constraint relationship of the current frame and the loop frame is brought into global optimization to correct the accumulated error.
The visual perception module and the intelligent decision module are combined to use a target detection algorithm to extract geographic entities existing in images, depth level features of the geographic entities are extracted by using a pre-trained AlexNet convolutional neural network, dimension reduction processing is carried out on feature vectors of each geographic entity, a grid map is generated by using a real-time digital earth surface model, barrier marks are arranged at corresponding positions in the grid map according to the types of the ground objects, starting from the initial place of the environment is achieved by combining a DSAS algorithm, an optimal path capable of bypassing the barriers is searched to reach the end point, and real-time path planning is achieved.
The invention also provides a dynamic battlefield environment real-time path planning method, which comprises the following steps:
as shown in fig. 4, an image shot by an unmanned aerial vehicle is used as an input to generate a real-time digital earth surface model, a target detection algorithm is used to extract geographic entities existing in the image, a pre-trained AlexNet convolutional neural network is used to extract depth level features of the geographic entities, dimension reduction processing is performed on feature vectors of each geographic entity to generate a grid map, corresponding positions in the grid map are provided with barrier marks according to types of the ground objects, starting from an initial place of the environment and combining with a DSAS algorithm, an optimal path capable of bypassing the barriers is searched to reach a terminal point, and real-time path planning is achieved.
Specifically, after the real-time digital earth surface model is generated, the geographic entities existing in the image are extracted by using a target detection algorithm, and in consideration of the problem of computational efficiency, the embodiment adopts an entity detection method based on Yolo, but the range of the detected entities is restricted, only the relatively static geographic entities are detected, such as the geographic entities of houses, bridges and the like, and the volatile entities of pedestrians, vehicles and the like are not detected. After the geographic entities are extracted, the depth level features of the geographic entities are extracted by using a pre-trained AlexNet convolutional neural network, because the dimension of the feature vector of each geographic entity is too high and greatly influences the calculation efficiency, dimension reduction needs to be carried out on the feature vector, dimension reduction is carried out by adopting a Gaussian random mapping method, barrier marks are arranged at corresponding positions in a grid map according to the types of the ground objects, and therefore data are provided for subsequent path planning.
After a grid map of path planning is obtained, a Dynamic Sparse A Search (DSAS) algorithm is adopted to Search an optimal path which can bypass an obstacle from the beginning of the environment to reach an end point, and real-time path planning is realized.
In the DSAS algorithm, each node of the flight path contains pointers pointing to a father node and all child nodes of the node except coordinate information representing the space position of the node, when a new obstacle is detected in front of the current flight path, the DSAS algorithm deletes the flight path node threatening the coverage area and all descendant nodes of the node, reinserts the father node into an OPEN table, then calculates the cost value of a new expansion node according to the method, and continues the searching process until a new optimal flight path of the aircraft is obtained.
The method comprises the steps of using a visual SLAM to take an aerial image shot by an unmanned aerial vehicle as input, generating a 3D point cloud for each frame of aerial image in real time, removing noise points of the 3D point cloud by adopting iterative filtering, performing Delaunay triangulation on feature points of the 3D point cloud after the noise points are filtered to generate a 2D mesh, projecting the generated 2D mesh into the 3D point cloud after the noise points are filtered to generate a 3D mesh, fusing texture information of the aerial image into the 3D mesh to generate a local DSM, segmenting the local DSM into tiles, and fusing the tiles to the global DSM through a multi-band algorithm to generate the real-time digital earth surface model.
The geographic entities present in the images are extracted using a target detection algorithm, which in this embodiment is Youonlylookone (YOLO). Different from the two-stage algorithm, the YOLO removes a candidate box generation part (a region suggestion stage), the whole process directly extracts features, candidate box regression and classification are completed in a model without branches, and the detection speed is improved by about 10 times compared with the two-stage algorithm.
Specifically, as shown in fig. 5, the input picture is divided into S × S grids, and each grid is responsible for detecting all objects falling into its own region. Each grid predicts 5 values, and the coordinates (x, y) of the center position of the predicted frame of the object, the width and height (w, h) of the object, and the confidence that the object belongs to a certain class are respectively predicted. Wherein, the confidence degree refers to the confidence degree that the region contains the object and the accuracy of the prediction box. If the prediction box contains the detection object, then Pr (object) is 1. Otherwise, pr (object) is 0. In the prediction phase, the probability of a class is the product of the class conditional probability and the confidence of the prediction:
Figure GDA0002929923620000171
the detailed flow chart of the network is shown in fig. 6, and the network includes 24 convolutional layers, 2 fully-connected layers, and 5 pooling layers. The output of the network is a 7 x 30 dimensional vector: 7 × 7 represents a grid that divides the input into 7 × 7; 30 includes object type and width, position, and confidence information of each of the two prediction frames, and the size of the input image is not changeable because of the inclusion of the full link layer, and must be 448 × 448.
Finally, it should be noted that: although the present invention has been described in detail with reference to the foregoing embodiments, it will be apparent to those skilled in the art that changes may be made in the embodiments and/or equivalents thereof without departing from the spirit and scope of the invention. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the scope of the claims of the present invention.

Claims (5)

1. A dynamic battlefield environment real-time path planning system, comprising: the map reconstruction module, the visual perception module, the intelligent decision module and the UI interaction module;
the UI interaction module realizes mutual calling and information connection of parameters among different modules;
the map reconstruction module includes: the visual SLAM takes aerial images shot by an unmanned aerial vehicle on line as input, generates 3D point cloud for each frame of image in real time, generates local DSMs in real time by combining the 3D point cloud with texture information of the image, and fuses the local DSMs to a global DSM to generate a real-time digital earth surface model;
the visual perception module is used for extracting geographic entities existing in images shot by the unmanned aerial vehicle, generating a grid map from the real-time digital earth surface model, and setting barrier marks on the grid map;
the intelligent decision module is combined with a DSAS algorithm to realize that an optimal path capable of bypassing the barrier is searched from the beginning of the environment to reach the end point, so that real-time path planning is realized;
the visual SLAM system comprises feature extraction, tracking, drawing and looping;
the feature extraction is used for extracting feature points in the aerial image and calculating a descriptor;
the tracking completes feature matching by calculating the distance between feature point descriptors of a current frame and a previous frame, and restores the pose change of the current frame relative to the previous frame through the feature matching, wherein the obtaining of the pose of the previous frame comprises:
the mathematical expression for the imaging of a pinhole camera is:
Figure FDA0003859871700000011
wherein Pi [ Xi, yi, zi ] is the three-dimensional coordinate of the ith map point, pi = [ ui, vi ] is the pixel coordinate of the projection of the ith map point on the current frame, si is the depth value corresponding to Pi, K is the camera parameter of the current frame, and T is the pose of the current frame, where T can be written in the form of a lie algebra:
Figure FDA0003859871700000012
in practice, due to the existence of noise, an error ei exists between the pixel coordinate ui of Pi after real projection and the theoretical value Pi, the real pixel coordinate with noise is recorded as an observed value, and when the error satisfies Gaussian distribution:
Figure FDA0003859871700000021
a least squares problem can be constructed, namely:
Figure FDA0003859871700000022
noting the derivative of ei with respect to ξ as Jacobian J (ξ), the world coordinate system is first down-converted to camera coordinate system as:
C i =exp(ξ^)P i =TP i =[X i ′,Y i ′,Z′ i ] T
the camera model may be represented as:
Figure FDA0003859871700000023
here, [ u ', v ' ] T is a specific value of the predicted value of the pixel coordinate, and can be obtained by removing si from si = Z ':
Figure FDA0003859871700000024
Figure FDA0003859871700000025
in the formula of the formula error term, ci is taken as an intermediate variable, and a chain derivation method is used to obtain:
Figure FDA0003859871700000026
wherein:
Figure FDA0003859871700000031
the partial derivative of Ci about xi is equivalent to derivation of a lie algebra, a disturbance model can be constructed to complete the derivation, the disturbance quantity delta xi is multiplied by xi left, and then the derivative of the increment of TPi to a disturbance item delta xi is calculated:
Figure FDA0003859871700000032
the partial derivative of the rearrangement error term with respect to ξ can be found as:
Figure FDA0003859871700000033
the negative sign in front of the Jacobian matrix is defined by subtracting the predicted value from the observed value due to the error in the derivation process, and cholesky decomposition is carried out on sigma-1 in the formula, namely:
-1 =LL T =L T L
bringing the above formula into available:
Figure FDA0003859871700000034
the formula for the least squares method can then be written as:
Figure FDA0003859871700000035
when performing iterative optimization, taking a gaussian optimization algorithm as an example here for simplicity, taylor expansion is performed for the error term:
Figure FDA0003859871700000036
where J (ξ) is the derivative of e (ξ) with respect to ξ (Jacobian), which can be:
Figure FDA0003859871700000041
the single two-norm is expanded to obtain:
Figure FDA0003859871700000042
finding the derivative of the above equation with respect to ξ equal zero can give:
Figure FDA0003859871700000043
an incremental equation can be obtained:
Figure FDA0003859871700000044
the increment delta xi can be solved by an increment equation, so that iterative optimization is realized;
searching more feature matches in the current frame and the previous frames, recovering more 3D points through triangulation, and performing local BA optimization on the current frame, the commonly observed associated frame and the commonly observed map point to acquire more accurate camera pose and 3D map point coordinates;
the loop is used for detecting whether the camera returns to a place which is previously visited or not, and a new constraint is added according to the detected loop to correct the corresponding camera pose and the 3D map point coordinate.
2. A dynamic battlefield environment real-time path planning method is characterized in that: the method comprises the following steps:
the method comprises the steps of generating a real-time digital earth surface model by taking an image shot by an unmanned aerial vehicle as input, extracting geographic entities existing in the image by using a target detection algorithm, extracting depth level features of the geographic entities by using a pre-trained AlexNet convolutional neural network, performing dimension reduction processing on feature vectors of each geographic entity to generate a grid map, setting barrier marks at corresponding positions in the grid map according to the types of the ground objects, starting from an initial place of an environment by combining a DSAS algorithm, searching for an optimal path capable of bypassing the barriers to reach an end point, and realizing real-time path planning;
the method comprises the steps that a real-time digital earth surface model is generated by taking an image shot by an unmanned aerial vehicle as input, wherein the image shot by the unmanned aerial vehicle is taken as input by using a visual SLAM, and the visual SLAM comprises feature extraction, tracking, drawing and looping;
the feature extraction is used for extracting feature points in the aerial image and calculating a descriptor;
the tracking completes feature matching by calculating the distance between feature point descriptors of a current frame and a previous frame, and restores the pose change of the current frame relative to the previous frame through the feature matching, wherein the obtaining of the pose of the previous frame comprises:
the mathematical expression for the imaging of a pinhole camera is:
Figure FDA0003859871700000051
wherein Pi [ Xi, yi, zi ] is the three-dimensional coordinate of the ith map point, pi = [ ui, vi ] is the pixel coordinate of the projection of the ith map point on the current frame, si is the depth value corresponding to Pi, K is the camera parameter of the current frame, and T is the pose of the current frame, where T can be written in the form of a lie algebra:
Figure FDA0003859871700000052
in practice, due to the existence of noise, an error ei exists between the pixel coordinate ui of Pi after real projection and the theoretical value Pi, the real pixel coordinate with noise is recorded as an observed value, and when the error satisfies Gaussian distribution:
Figure FDA0003859871700000053
a least squares problem can be constructed, namely:
Figure FDA0003859871700000061
noting the derivative of ei with respect to ξ as Jacobian J (ξ), the world coordinate system is first down-converted to camera coordinate system as:
C i =exp(ξ^)P i =TP i =[X i ′,Y i ′,Z i ′] T
the camera model may be represented as:
Figure FDA0003859871700000062
here, [ u ', v ' ] T is a specific value of the predicted value of the pixel coordinate, and can be obtained by removing si from si = Z ':
Figure FDA0003859871700000063
Figure FDA0003859871700000064
in the formula of the formula error term, ci is taken as an intermediate variable, and a chain derivative method is used to obtain:
Figure FDA0003859871700000065
wherein:
Figure FDA0003859871700000066
the partial derivative of Ci about xi is equivalent to derivation of a lie algebra, a disturbance model can be constructed to complete the derivation, the disturbance quantity delta xi is multiplied by xi on the left side, and then the derivative of the increment of TPi to a disturbance item delta xi is calculated:
Figure FDA0003859871700000067
the partial derivative of the rearrangement error term with respect to xi can be derived as:
Figure FDA0003859871700000071
the negative sign in front of the Jacobian matrix is defined by subtracting the predicted value from the observed value due to the error in the derivation process, and cholesky decomposition is carried out on sigma-1 in the formula, namely:
-1 =LL T =L T L
bringing the above formula into available:
Figure FDA0003859871700000072
the formula for the least squares method can then be written as:
Figure FDA0003859871700000073
when performing iterative optimization, taking the gaussian optimization algorithm as an example here for simplicity, a taylor expansion is performed for the error term:
Figure FDA0003859871700000074
where J (ξ) is the derivative of e (ξ) with respect to ξ (Jacobian), which can be:
Figure FDA0003859871700000075
the single two-norm is expanded to:
Figure FDA0003859871700000081
finding the derivative of the above equation with respect to ξ equal zero yields:
Figure FDA0003859871700000082
an incremental equation can be obtained:
Figure FDA0003859871700000083
the increment delta xi can be solved by an increment equation, so that iterative optimization is realized;
generating a 3D point cloud for each frame of aerial image in real time, removing noise points of the 3D point cloud by iterative filtering, executing Delaunay triangulation on feature points of the 3D point cloud after the noise points are filtered to generate a 2D mesh, projecting the generated 2D mesh into the 3D point cloud after the noise points are filtered to generate a 3D mesh, fusing texture information of the aerial image into the 3D mesh to generate a local DSM, segmenting the local DSM into tiles, and fusing the tiles to the global DSM by a multiband algorithm to generate a real-time digital earth surface model.
3. The method of claim 2, wherein the dynamic battlefield environment real-time path planning method comprises the following steps: the method for extracting the geographic entities existing in the image by using the target detection algorithm comprises the steps of dividing the input image into S × S grids, wherein each grid is responsible for detecting all objects falling into the own region, 5 values are predicted by each grid, the coordinates (x, y) of the center position of a prediction frame of each predicted object, the width and height (w, h) of each object and the confidence coefficient of each object belonging to a certain class are respectively predicted, if the prediction frame contains a detection object, pr is 1, otherwise, pr is 0, and in the prediction stage, the probability of each class is the product of the conditional probability of each class and the confidence coefficient of each prediction:
Figure FDA0003859871700000084
4. the method of claim 3, wherein the dynamic battlefield environment real-time path planning method comprises the following steps: the geographic entity comprises a relatively stationary geographic entity.
5. The dynamic battlefield environment real-time path planning method according to claim 3, wherein: in the DSAS algorithm, besides coordinate information representing the space position of the unmanned aerial vehicle, pointers pointing to a father node and all child nodes of the unmanned aerial vehicle are also included, when a new obstacle is detected in front of the current track, the DSAS algorithm deletes the track node threatening the coverage area and all descendant nodes of the track node, reinserts the father node into an OPEN table, then calculates the cost value of a new extended node, and continues the searching process until a new optimal track of the unmanned aerial vehicle is obtained.
CN202011373402.8A 2020-11-30 2020-11-30 Dynamic battlefield environment real-time path planning system and method Active CN112509056B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011373402.8A CN112509056B (en) 2020-11-30 2020-11-30 Dynamic battlefield environment real-time path planning system and method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011373402.8A CN112509056B (en) 2020-11-30 2020-11-30 Dynamic battlefield environment real-time path planning system and method

Publications (2)

Publication Number Publication Date
CN112509056A CN112509056A (en) 2021-03-16
CN112509056B true CN112509056B (en) 2022-12-20

Family

ID=74968697

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011373402.8A Active CN112509056B (en) 2020-11-30 2020-11-30 Dynamic battlefield environment real-time path planning system and method

Country Status (1)

Country Link
CN (1) CN112509056B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113296514B (en) * 2021-05-24 2022-09-27 南开大学 Local path optimization method and system based on sparse banded structure
CN115107904B (en) * 2022-07-12 2023-09-05 武汉理工大学 Reconfigurable wheel-track universal distributed driving unmanned vehicle and control method thereof
CN116842127B (en) * 2023-08-31 2023-12-05 中国人民解放军海军航空大学 Self-adaptive auxiliary decision-making intelligent method and system based on multi-source dynamic data
CN117191048B (en) * 2023-11-07 2024-01-05 北京四象爱数科技有限公司 Emergency path planning method, equipment and medium based on three-dimensional stereopair

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105843223A (en) * 2016-03-23 2016-08-10 东南大学 Mobile robot three-dimensional mapping and obstacle avoidance method based on space bag of words model
CN108827306A (en) * 2018-05-31 2018-11-16 北京林业大学 A kind of unmanned plane SLAM navigation methods and systems based on Multi-sensor Fusion
CN109242862A (en) * 2018-09-08 2019-01-18 西北工业大学 A kind of real-time digital surface model generation method
CN109828578A (en) * 2019-02-22 2019-05-31 南京天创电子技术有限公司 A kind of instrument crusing robot optimal route planing method based on YOLOv3
CN109959377A (en) * 2017-12-25 2019-07-02 北京东方兴华科技发展有限责任公司 A kind of robot navigation's positioning system and method
CN110675307A (en) * 2019-08-19 2020-01-10 杭州电子科技大学 Implementation method of 3D sparse point cloud to 2D grid map based on VSLAM
CN110673632A (en) * 2019-09-27 2020-01-10 中国船舶重工集团公司第七0九研究所 Unmanned aerial vehicle autonomous obstacle avoidance method and device based on visual SLAM
CN111367318A (en) * 2020-03-31 2020-07-03 华东理工大学 Dynamic obstacle environment navigation method and device based on visual semantic information

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105843223A (en) * 2016-03-23 2016-08-10 东南大学 Mobile robot three-dimensional mapping and obstacle avoidance method based on space bag of words model
CN109959377A (en) * 2017-12-25 2019-07-02 北京东方兴华科技发展有限责任公司 A kind of robot navigation's positioning system and method
CN108827306A (en) * 2018-05-31 2018-11-16 北京林业大学 A kind of unmanned plane SLAM navigation methods and systems based on Multi-sensor Fusion
CN109242862A (en) * 2018-09-08 2019-01-18 西北工业大学 A kind of real-time digital surface model generation method
CN109828578A (en) * 2019-02-22 2019-05-31 南京天创电子技术有限公司 A kind of instrument crusing robot optimal route planing method based on YOLOv3
CN110675307A (en) * 2019-08-19 2020-01-10 杭州电子科技大学 Implementation method of 3D sparse point cloud to 2D grid map based on VSLAM
CN110673632A (en) * 2019-09-27 2020-01-10 中国船舶重工集团公司第七0九研究所 Unmanned aerial vehicle autonomous obstacle avoidance method and device based on visual SLAM
CN111367318A (en) * 2020-03-31 2020-07-03 华东理工大学 Dynamic obstacle environment navigation method and device based on visual semantic information

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于DSAS 算法的直升机贴地飞行在线航迹规划;陈洁等;《飞机力学》;20160430;第86-89页 *

Also Published As

Publication number Publication date
CN112509056A (en) 2021-03-16

Similar Documents

Publication Publication Date Title
CN112509056B (en) Dynamic battlefield environment real-time path planning system and method
CN108242079B (en) VSLAM method based on multi-feature visual odometer and graph optimization model
CN109631855B (en) ORB-SLAM-based high-precision vehicle positioning method
CN113345018B (en) Laser monocular vision fusion positioning mapping method in dynamic scene
US10269148B2 (en) Real-time image undistortion for incremental 3D reconstruction
Liu et al. Seqlpd: Sequence matching enhanced loop-closure detection based on large-scale point cloud description for self-driving vehicles
CN104200523B (en) A kind of large scene three-dimensional rebuilding method for merging additional information
US20180315232A1 (en) Real-time incremental 3d reconstruction of sensor data
CN110827415A (en) All-weather unknown environment unmanned autonomous working platform
CN113506318B (en) Three-dimensional target perception method under vehicle-mounted edge scene
Ding et al. Vehicle pose and shape estimation through multiple monocular vision
CN113593017A (en) Method, device and equipment for constructing surface three-dimensional model of strip mine and storage medium
Li et al. Review of vision-based Simultaneous Localization and Mapping
Agostinho et al. A practical survey on visual odometry for autonomous driving in challenging scenarios and conditions
CN114088081B (en) Map construction method for accurate positioning based on multistage joint optimization
Ding et al. Persistent stereo visual localization on cross-modal invariant map
Saleem et al. Neural network-based recent research developments in SLAM for autonomous ground vehicles: A review
Liu et al. Comparison of 2D image models in segmentation performance for 3D laser point clouds
Gökçe et al. Recognition of dynamic objects from UGVs using Interconnected Neuralnetwork-based Computer Vision system
Lu et al. Vision-based localization methods under GPS-denied conditions
Voodarla et al. S-BEV: Semantic birds-eye view representation for weather and lighting invariant 3-dof localization
Xu et al. Uav image geo-localization by point-line-patch feature matching and iclk optimization
Zhang et al. GPS-assisted aerial image stitching based on optimization algorithm
Li et al. RF-LOAM: Robust and Fast LiDAR Odometry and Mapping in Urban Dynamic Environment
Li et al. Intelligent vehicle localization and navigation based on intersection fingerprint roadmap (IRM) in underground parking lots

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant