CN111783838A - Point cloud characteristic space representation method for laser SLAM - Google Patents

Point cloud characteristic space representation method for laser SLAM Download PDF

Info

Publication number
CN111783838A
CN111783838A CN202010504793.6A CN202010504793A CN111783838A CN 111783838 A CN111783838 A CN 111783838A CN 202010504793 A CN202010504793 A CN 202010504793A CN 111783838 A CN111783838 A CN 111783838A
Authority
CN
China
Prior art keywords
point cloud
network
global description
point
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010504793.6A
Other languages
Chinese (zh)
Inventor
莫凌飞
索传哲
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Southeast University
Original Assignee
Southeast University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Southeast University filed Critical Southeast University
Priority to CN202010504793.6A priority Critical patent/CN111783838A/en
Publication of CN111783838A publication Critical patent/CN111783838A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)

Abstract

The invention discloses a point cloud feature space characterization method for laser SLAM, which is used for establishing bidirectional mapping between a point cloud space and a point cloud feature space based on a deep learning network, wherein the feature space realizes key problems of closed-loop detection, relocation and the like in the laser SLAM and map compression storage and transmission. The method comprises the steps of realizing global description feature extraction and compression reconstruction of large scene point clouds through a neural network of a self-encoder structure, designing a characteristic space of the point clouds formed by extracting global description features through the encoder network, giving similarity measurement of scenes according to the distance in the characteristic space, judging whether two or more scene structures are similar, and realizing closed-loop detection and relocation of laser SLAM; reconstructing the point cloud through a designed decoder network, reconstructing an original point cloud from global description features extracted from an encoder network, and realizing compressed storage and low-bandwidth transmission of a point cloud map; the constructed encoder network does not need to be trained in advance according to the point cloud map, and has strong generalization capability.

Description

Point cloud characteristic space representation method for laser SLAM
Technical Field
The invention relates to the technical field of laser point cloud mapping and autonomous navigation, in particular to a point cloud feature space characterization method for laser SLAM.
Background
How to enable the mobile robot to better understand and perceive the surrounding environment and achieve flexible and reliable high-level autonomous navigation, the artificial intelligence technology promoted by deep learning brings high-speed development for environment perception understanding and autonomous navigation of the mobile robot, and particularly, the application of laser radar enables the robot to directly perceive a three-dimensional environment, but also brings challenges for point cloud data processing. Synchronous positioning And map building (SLAM) is one of the basic And key technologies for realizing autonomous navigation positioning of a mobile robot, And aims to establish a local map And determine the position of the robot in the map when the robot enters an unknown environment. The SLAM mainly comprises four parts, namely a front-end odometer, rear-end pose graph optimization, closed-loop detection and map construction, wherein the closed-loop detection is one of key components, the problem that whether a mobile robot returns to a previously visited environment is solved, a closed-loop scene is accurately detected, long-time accumulated errors are corrected for the SLAM system, and the establishment of globally consistent pose estimation and a map is very important.
The method is characterized in that how to make a robot recognize the environment is an important problem, closed loop detection is also a scene recognition problem in nature, and environment recognition and relocation are carried out through sensor data such as a camera, a laser radar and the like. Scene recognition is always a hotspot and difficult problem in the field of pattern recognition, and a camera needs to overcome adverse effects of factors such as illumination, visual angle and the like on scene recognition and extract useful information in the scene recognition. Compared with a camera, the laser radar can overcome the influence caused by illumination, seasons, weather and the like, but objects in laser point cloud scene data are various and comprise a large number of dynamic scenes, even if the same scene has larger difference under various external influence factors at different time, the scene identification problem is very complex, and the key of the scene identification lies in how to find out a global description feature which can completely represent the structural information and semantic information of the same scene, in particular to a feature characterization method for large-range laser point cloud.
The traditional environment recognition algorithm bag-of-words model (BoW), FV (Fisher vector), VLAD and the like based on point cloud usually depend on global, offline and high-resolution maps, a codebook needs to be trained in advance according to the maps to realize high-precision positioning, a scene recognition method based on 3D segmentation matching is provided by Segmatch based on deep learning, local matching is carried out through segmented scene blocks, the reliability of the local matching and the scene recognition is ensured through geometric consistency check, but for dynamic scene application, the static assumption of fast scene cannot be met. The PointNet and NetVLAD are combined, the PointNet and the NetVLAD are used for feature learning, the NetVLAD is used for feature aggregation, global description features can be learned, the PCAN based on point cloud deep learning is also used, the contribution degrees of different local features are considered on the basis of the PointNet VLAD, a self-attention mechanism is introduced, the contribution weights of the different local features are learned through the network during feature aggregation, but the local geometric features and semantic context features, point cloud neighborhood relations and feature space distribution are not considered in the method, and the effect in large-scene point cloud scene recognition feature coding is poor.
Disclosure of Invention
The technical problem to be solved by the invention is to provide a point cloud feature space characterization method for laser SLAM, which can extract point cloud global description features representing scene structure information and semantic information from complex large-scene laser point clouds and restore and reconstruct original point clouds from the features, and is used for closed-loop detection, relocation, compression storage and transmission of the laser SLAM.
In order to solve the technical problem, the invention provides a point cloud feature space characterization method for a laser SLAM, which comprises the following steps:
s1, filtering and downsampling the point cloud samples in the point cloud data set, and constructing a training point cloud sample pair according to the similarity of the point cloud scenes, wherein the training point cloud sample pair comprises a positive sample pposAnd negative sample pneg
s2, constructing an encoder network, inputting the training point cloud sample pair in the step s1, and obtaining global description characteristics f (P) of the input point cloud P through a point cloud self-adaptive neighborhood characteristic extraction module, a dynamic graph network aggregation module and a global description characteristic generation module;
s3, constructing a decoder network, inputting the global description characteristics generated in the step s2, generating a strategy through layeringGenerating refined and densified reconstructed point cloud from contour of key point
Figure BDA0002526140900000021
s4, taking sample pairs of the training data set in s1, and training the encoder network of s2 to obtain an encoder network model and parameters capable of measuring point cloud scene similarity;
s5, taking sample pairs of the training data set in s1 and corresponding global description features generated by an s2 encoder network, and training the decoder network of s3 to obtain a decoder network model and parameters capable of restoring and reconstructing original point clouds from the global description features;
s6, in the actual SLAM operation, taking the encoder network model and parameters in s4, extracting global description characteristics of the point cloud key frame, carrying out NN matching on the point cloud key frame and the global description characteristics of the historical key frame according to Euclidean distance, and taking the point cloud key frame which meets the matching conditions as a closed-loop candidate frame or a repositioning candidate position;
s7, in the actual SLAM operation, storing or transmitting the global description characteristics of the point cloud key frame extracted in s6, and restoring the global description characteristics to reconstruct the original SLAM point cloud map through the decoder network model and parameters in s 5;
s8, in the actual SLAM operation, the global description features extracted in s6 are taken, NN matching is conducted on the global description features and the global description features of the historical key frames of the multiple collaborative mapping robots according to Euclidean distances, and the multi-robot laser SLAM collaborative mapping task is conducted in the point cloud feature space through the point cloud feature space characterization method.
Preferably, in step s1, the similarity of the training sample pair is determined by the position coordinates of the samples in the map, the position interval is within 10m and is regarded as a positive sample with similar structure, the distance between 10m and 50 m is regarded as a negative sample selection range, the negative sample is randomly selected, and the randomly selected sample with the distance outside 50 m is regarded as a very dissimilar extreme negative sample.
Preferably, in step s2, the point cloud adaptive neighborhood feature extraction module of the encoder network uses the range [ k ]min,kmax]Constructing each point p in a point cloudiK NN neighborhood of
Figure BDA0002526140900000031
Eigenvalues λ of covariance matrix Σ using neighborhood point clouds1≥λ2≥λ3Computing linearity of point cloud neighborhood not less than 0
Figure BDA0002526140900000032
Planarity
Figure BDA0002526140900000033
And scattering property
Figure BDA0002526140900000034
Is characterized by and measures an equation E through Shannon information entropykAdaptive determination of optimal neighborhood size
Figure BDA0002526140900000036
Extracting neighborhood characteristics as characteristics of points through a multilayer perceptron characteristic extraction network;
wherein:
Figure BDA0002526140900000035
v is a feature vector matrix, Ek=-LλlnLλ-PλlnPλ-SλlnSλ
Preferably, in step s2, the dynamic graph network aggregation module is composed of two-part aggregation:
(a) and (3) feature space dynamic graph network aggregation: performing k-fixed kNN on all points in the feature space in the step (3) to construct a dynamic graph, performing feature aggregation through a graph network, and updating the features of the points; the characteristic aggregation process of the graph network comprises the following steps: f. ofi'=ρ(φΘ((fi,fj))=ρ(φΘ(fi‖(fj-fi))),φΘThe method is characterized in that the method is a multilayer perceptron MLP, and rho is maximum pooling;
(b) physical space dynamic graph network aggregation: all click-throughs in the physical space of the point cloudConstructing a dynamic graph by the kNN of the line fixed k, and updating the characteristics of the points by utilizing the characteristics of the points updated in the step (a) through a graph network; center point piThe constructed dynamic graph G (V, E) is composed of nodes V and edges E, and points piIs characterized in thatiAnd the node V is the feature f of the neighborhood pointjAnd the edge E is the relation (f) between the central point and the characteristics of the neighborhood pointsi,fj) Wherein: (f)i,fj)=fi‖(fj-fi) And | is characterized by splicing according to the channels.
Preferably, in step s2, the global description feature generation module generates the global description feature through the NetVLAD Layer.
Preferably, in step s3, the hierarchical generation strategy includes two parts:
(a) the key point outline generating network comprises the following steps: generating a rough key point outline of points through a three-layer fully-connected network and Reshape;
(b) refining dense point cloud generating network, namely splicing η×η points of 2D plane mesh of each point in the key points generated in the step (a)
Figure BDA0002526140900000041
Splicing the global description characteristics generated in the step s2 according to channels, and generating dense reconstruction point cloud through MLP;
wherein:
Figure BDA0002526140900000042
η is the number of dots on the edge of the planar grid and ξ is the size of the planar grid dots.
Preferably, in step s4, the encoder network is trained using a metric learning method, and a quadruple loss function is used
Figure BDA0002526140900000044
Semi-supervised training leads to input samples and positive samples pposGlobal of each otherEuclidean distance of descriptive featuresposCloser, with negative example pnegBetween the global descriptor feature of the Euclidean distancenegFarther, extremely negative sample
Figure BDA0002526140900000045
For the zooming-in for preventing the error of the global description characteristic distance between the negative sample pairs, the Euclidean distance between the extreme negative sample and the negative sample is
Figure BDA0002526140900000046
Wherein:
Figure BDA0002526140900000047
α is a hyperparameter.
Preferably, in step s5, the decoder network is trained in a supervised manner, using the loss function
Figure BDA0002526140900000048
Constraining Euclidean distance between the reconstructed point cloud and the truth point cloud, wherein
Figure BDA0002526140900000049
Applying the keypoint contours generated in step s2 (a) for constraining the physical spatial distribution of the point cloud,
Figure BDA00025261409000000410
applied to dense reconstructed point clouds;
wherein:
Figure BDA00025261409000000411
psi (p) is
Figure BDA00025261409000000412
The corresponding point of p in (1) is,
Figure BDA00025261409000000413
Figure BDA0002526140900000051
preferably, in steps s6 and s8, NN matching of global description features of the point cloud key frames is determined by sequence matching of consecutive N current point cloud frames and consecutive N historical point cloud frames, and if the euclidean distance of the global description features between consecutive sequence matching key frames is smaller than a set threshold and the gradient change compared with the feature distance of other historical key frames is larger than the set threshold, it may be regarded as a closed-loop detection candidate sequence key frame or a relocation candidate position; if a plurality of robots provide current continuous point cloud frames and historical key frames, establishing a map for the plurality of robots cooperatively; and performing a laser SLAM closed-loop task, a relocation task and a multi-robot cooperative task in the point cloud feature space by the point cloud feature space characterization method.
Preferably, in step s7, the encoder network in s4 is used to extract global description features from the keyframes in the laser point cloud map as compressed codes for storage or communication transmission, and then the decoder network in s5 is used to restore the storage features or transmission and reception features to reconstruct the original laser point cloud map.
The invention has the beneficial effects that: (1) an encoder network is designed to extract information of geometric structural features and semantic context features of large-scene laser point cloud data, point cloud scene information is fully mined, global description features are generated to form a feature space of point cloud, and similarity measurement of scenes is given according to the distance in the feature space and used for judging whether two or more scene structures are similar or not, and closed-loop detection and relocation of laser SLAM can be realized in the point cloud feature space; (2) reconstructing the point cloud through a designed decoder network, reconstructing an original point cloud from global description features extracted from an encoder network, and realizing compressed storage and low-bandwidth transmission of a point cloud map; (3) the constructed encoder network does not need to be trained in advance according to the point cloud map, and has strong generalization capability.
Drawings
FIG. 1 is a schematic flow chart of the method of the present invention.
Fig. 2 is a schematic diagram of a network model of an encoder according to the present invention.
Fig. 3 is a schematic diagram of a decoder network model structure according to the present invention.
Detailed Description
As shown in fig. 1, a method for characterizing a point cloud feature space of a laser SLAM includes the following steps:
step 1: preprocessing point cloud samples in the point cloud data set such as filtering and downsampling, and constructing training point cloud sample pairs according to similarity of point cloud scenes, wherein the training point cloud sample pairs comprise positive samples pposAnd negative sample pneg
The similarity of the constructed training sample pairs is determined by the position coordinate assistance judgment of the samples in the map, the position interval is within 10m and is regarded as a positive sample with similar structure, the distance between 10m and 50 m is regarded as a negative sample selection range, a negative sample is randomly selected, and the randomly selected samples beyond 50 m are regarded as extremely dissimilar polar and negative samples.
The method specifically comprises the following steps: an Oxford RobotCar dataset is chosen, the point cloud range is set to [ x: + -30 m, y: + -20 m, z:0-10m ] using a straight-through filter, and the point cloud is downsampled to 4096 points using a random voxel filter and normalized to [ -1 ].
Step 2: the encoder network is constructed by the network structure design shown in fig. 2, the training point cloud sample pair in the step 1 is input, and the global description feature f (P) of the input point cloud P is obtained by the point cloud self-adaptive neighborhood feature extraction module, the dynamic graph network aggregation module and the global description feature generation module.
Point cloud adaptive neighborhood feature extraction module for encoder networks by range kmin,kmax]Constructing each point p in a point cloudiK NN neighborhood of
Figure BDA0002526140900000061
Eigenvalues λ of covariance matrix Σ using neighborhood point clouds1≥λ2≥λ3Computing linearity of point cloud neighborhood not less than 0
Figure BDA0002526140900000062
Planarity
Figure BDA0002526140900000063
And scattering property
Figure BDA0002526140900000064
Is characterized by and measures an equation E through Shannon information entropykAdaptive determination of optimal neighborhood size
Figure BDA0002526140900000065
Extracting neighborhood characteristics as characteristics of points through an MLP (multi-layer perceptron) characteristic extraction network.
Wherein:
Figure BDA0002526140900000066
v is a feature vector matrix
Ek=-LλlnLλ-PλlnPλ-SλlnSλ
The method specifically comprises the following steps: setting a point cloud self-adaptive neighborhood characteristic range [20-100], setting the step length as 10, and calculating and determining the size of the optimal neighborhood.
The dynamic graph network aggregation module is composed of two parts of aggregation:
(a) and (3) feature space dynamic graph network aggregation: constructing a dynamic graph by performing kNN of a fixed k on all points of the feature space of claim 3, performing feature aggregation through a graph network, and updating the features of the points;
(b) physical space dynamic graph network aggregation: and (3) performing k-fixed kNN on all points in the point cloud physical space to construct a dynamic graph, and performing feature update by using the updated features of the points in the step (a) through a graph network to update the features of the points. Center point piThe constructed dynamic graph G (V, E) is composed of nodes V and edges E, and points piIs characterized in thatiAnd the node V is the feature f of the neighborhood pointjAnd the edge E is the relation (f) between the central point and the characteristics of the neighborhood pointsi,fj)。
Wherein: (f)i,fj)=fi‖(fj-fi),II, splicing according to channels by using characteristics;
the characteristic aggregation process of the graph network comprises the following steps:
fi'=ρ(φΘ((fi,fj))=ρ(φΘ(fi‖(fj-fi))),φΘfor MLP (multilayer perceptron), ρ is the maximum pooling.
The method specifically comprises the following steps: dynamic graph construction kNN, phi using a fixed k of 20ΘMLP in two layers (64,64)
Preferably, the global description feature generation module generates the global description feature through a NetVLAD Layer.
The method specifically comprises the following steps: the NetVLAD Layer selects the central feature dimension D as 256 and selects K as 64 central features.
And the global description feature generation module generates a global description feature through the NetVLAD Layer.
And step 3: constructing a decoder network by designing a network structure as shown in FIG. 3, inputting the global description features generated in step 2, generating key point outlines and generating refined and densified reconstructed point clouds by a hierarchical generation strategy
Figure BDA0002526140900000075
The hierarchical generation strategy comprises two parts:
(a) the key point outline generating network comprises the following steps: generating a rough key point outline of points through a three-layer fully-connected network and Reshape;
(b) refining dense point cloud generating network, namely splicing η×η points of 2D plane mesh of each point in the key points generated in the step (a)
Figure BDA0002526140900000071
And splicing the global description characteristics generated in the step 2 according to channels, and generating dense reconstruction point cloud through MLP (multi-layer perceptron).
Wherein:
Figure BDA0002526140900000072
Figure BDA0002526140900000073
η is the number of dots on the edge of the planar grid and ξ is the size of the planar grid dots.
The method specifically comprises the following steps: the keypoint profile generation network generated a rough profile containing 4096 keypoints and a 2 × 2 point planar network was stitched with a size ξ of 0.05.
And 4, step 4: and (3) taking a sample pair of the training data set in the step (1), and training the encoder network in the step (2) to obtain an encoder network model and parameters capable of measuring the similarity of the point cloud scene.
Training an encoder network using a metric learning method, using a quadruple loss function
Figure BDA0002526140900000074
Semi-supervised training leads to input samples and positive samples pposBetween global descriptor featureposCloser, with negative example pnegBetween the global descriptor feature of the Euclidean distancenegFarther, extremely negative sample
Figure BDA0002526140900000081
For the zooming-in for preventing the error of the global description characteristic distance between the negative sample pairs, the Euclidean distance between the extreme negative sample and the negative sample is
Figure BDA0002526140900000082
Wherein:
Figure BDA0002526140900000083
α is a hyperparameter.
The method specifically comprises the following steps: setting the hyper-parameter alpha to 0.5 and beta to 0.2, and selecting 2 positive samples, 18 negative samples and 1 randomly selected extreme negative sample during training.
And 5: and (3) taking the sample pairs of the training data set in the step (1) and the corresponding global description features generated by the encoder network in the step (2), and training the decoder network in the step (3) to obtain a decoder network model and parameters which can restore and reconstruct the original point cloud from the global description features.
In step 5, the decoder network is trained in a supervised manner, using the loss function
Figure BDA0002526140900000084
Constraining Euclidean distance between the reconstructed point cloud and the truth point cloud, wherein
Figure BDA0002526140900000085
Applied to the generated key point contour for constraining the physical spatial distribution of the point cloud,
Figure BDA0002526140900000086
applied to dense reconstructed point clouds.
Wherein:
Figure BDA0002526140900000087
psi (p) is
Figure BDA0002526140900000088
Corresponding point of (3)
Figure BDA0002526140900000089
Figure BDA00025261409000000810
Step 6: in the actual SLAM operation, the encoder network model and parameters in step 4 are taken, global description features are extracted from the point cloud key frames, NN matching is carried out on the point cloud key frames and the global description features of the historical key frames according to Euclidean distances, and the point cloud key frames which meet matching conditions are used as closed-loop candidate frames or relocation candidate positions.
In step 6, the NN matching of the global description features of the point cloud key frames is determined by sequence matching of N consecutive current point cloud frames and N consecutive historical point cloud frames, and if the euclidean distance of the global description features between consecutive sequence matching key frames is smaller than a set threshold and the gradient change compared with the feature distance of other historical key frames is larger than the set threshold, the NN matching may be regarded as a closed-loop detection candidate sequence key frame or a relocation candidate position. And performing laser SLAM closed-loop task and repositioning task in the point cloud characteristic space by the point cloud characteristic space characterization method.
And 7: in the actual SLAM operation, the global description features of the point cloud key frame extracted in the step 6 are stored or transmitted, and the global description features can be restored to reconstruct the original SLAM point cloud map through the decoder network model and the parameters in the step 5.
And (4) extracting global description features from the key frames in the laser point cloud map by using the encoder network in the step (4) as compression codes for storage or communication transmission, and restoring the storage features or transmission and reception features by using the decoder network in the step (5) to reconstruct the original laser point cloud map.
And 8: in the actual SLAM operation, the global description features extracted in s6 are taken, NN matching is carried out on the global description features and the global description features of the historical key frames of the multiple collaborative mapping robots according to Euclidean distances, and the multi-robot laser SLAM collaborative mapping task is carried out in the point cloud feature space through the point cloud feature space characterization method.
In step 8, if a plurality of robots provide current continuous point cloud frames and historical key frames, multi-robot collaborative map building is performed, NN matching is performed on global description characteristics of all point cloud historical key frames, and multi-robot collaborative map building can be achieved. And performing a laser SLAM multi-robot cooperative task in the characteristic space of the point cloud by the point cloud characteristic space characterization method.
The invention has the beneficial effects that: (1) an encoder network is designed to extract information of geometric structural features and semantic context features of large-scene laser point cloud data, scene recognition is realized by the point cloud feature space representation method in a feature retrieval mode, and the average recall rate in an Oxford RobotCar data set scene recognition task reaches 93.35%; (2) reconstructing the point cloud through a designed decoder network, reconstructing an original point cloud from global description features extracted from an encoder network, and realizing compressed storage and low-bandwidth transmission of a point cloud map; (3) the constructed encoder network does not need to be trained in advance according to the point cloud map, has strong generalization capability, and can be directly applied to various data sets such as KITTI without training.

Claims (10)

1. A point cloud feature space characterization method for laser SLAM is characterized by comprising the following steps:
s1, filtering and downsampling the point cloud samples in the point cloud data set, and constructing a training point cloud sample pair according to the similarity of the point cloud scenes, wherein the training point cloud sample pair comprises a positive sample pposAnd negative sample pneg
s2, constructing an encoder network, inputting the training point cloud sample pair in the step s1, and obtaining global description characteristics f (P) of the input point cloud P through a point cloud self-adaptive neighborhood characteristic extraction module, a dynamic graph network aggregation module and a global description characteristic generation module;
s3, constructing a decoder network, inputting the global description characteristics generated in the step s2, generating key point outlines to generate refined and densified reconstructed point clouds through a layering generation strategy
Figure FDA0002526140890000011
s4, taking sample pairs of the training data set in s1, and training the encoder network of s2 to obtain an encoder network model and parameters capable of measuring point cloud scene similarity;
s5, taking sample pairs of the training data set in s1 and corresponding global description features generated by an s2 encoder network, and training the decoder network of s3 to obtain a decoder network model and parameters capable of restoring and reconstructing original point clouds from the global description features;
s6, in the actual SLAM operation, taking the encoder network model and parameters in s4, extracting global description characteristics of the point cloud key frame, carrying out NN matching on the point cloud key frame and the global description characteristics of the historical key frame according to Euclidean distance, and taking the point cloud key frame which meets the matching conditions as a closed-loop candidate frame or a repositioning candidate position;
s7, in the actual SLAM operation, storing or transmitting the global description characteristics of the point cloud key frame extracted in s6, and restoring the global description characteristics to reconstruct the original SLAM point cloud map through the decoder network model and parameters in s 5;
s8, in the actual SLAM operation, the global description features extracted in s6 are taken, NN matching is conducted on the global description features and the global description features of the historical key frames of the multiple collaborative mapping robots according to Euclidean distances, and the multi-robot laser SLAM collaborative mapping task is conducted in the point cloud feature space through the point cloud feature space characterization method.
2. The method as claimed in claim 1, wherein in step s1, the similarity of the training sample pair is determined by the position coordinates of the samples in the map, the position interval is within 10m and is regarded as a positive sample with similar structure, the distance between 10m and 50 m is regarded as a negative sample selection range, the negative sample is randomly selected, and the randomly selected samples beyond 50 m are regarded as polar and negative samples with very dissimilar structure.
3. The method of claim 1, wherein in step s2, the point cloud adaptive neighborhood feature extraction module of the encoder network uses the range [ k ] as the point cloud feature extraction modulemin,kmax]Constructing each point p in a point cloudiK NN neighborhood of NiUsing the eigenvalues λ of the covariance matrix Σ of the neighborhood point cloud1≥λ2≥λ3Computing linearity of point cloud neighborhood not less than 0
Figure FDA0002526140890000021
Planarity
Figure FDA0002526140890000022
And scattering property
Figure FDA0002526140890000023
Is characterized by and measures an equation E through Shannon information entropykAdaptive determination of optimal neighborhood size
Figure FDA0002526140890000024
Extracting neighborhood characteristics as characteristics of points through a multilayer perceptron characteristic extraction network;
wherein:
Figure FDA0002526140890000025
v is a feature vector matrix, Ek=-Lλln Lλ-Pλln Pλ-SλlnSλ
4. The method of claim 1, wherein in step s2, the dynamic graph network aggregation module is composed of two parts:
(a) and (3) feature space dynamic graph network aggregation: performing k-fixed kNN on all points in the feature space in the step (3) to construct a dynamic graph, performing feature aggregation through a graph network, and updating the features of the points; the characteristic aggregation process of the graph network comprises the following steps: f. ofi'=ρ(φΘ((fi,fj))=ρ(φΘ(fi‖(fj-fi))),φΘThe method is characterized in that the method is a multilayer perceptron MLP, and rho is maximum pooling;
(b) physical space dynamic graph network aggregation: performing k-fixed kNN on all points in the point cloud physical space to construct a dynamic graph, and performing feature update by using the updated characteristics of the points in the step (a) through a graph network to update the characteristics of the points; center point piThe constructed dynamic graph G (V, E) is composed of nodes V and edges E, and points piIs characterized in thatiAnd the node V is the feature f of the neighborhood pointjAnd the edge E is the relation (f) between the central point and the characteristics of the neighborhood pointsi,fj) Wherein: (f)i,fj)=fi‖(fj-fi) And | is characterized by splicing according to the channels.
5. The method of claim 1, wherein in step s2, the global description feature generation module generates the global description feature through a NetVLAD Layer.
6. The method of claim 1, wherein in step s3, the hierarchical generation strategy comprises two parts:
(a) the key point outline generating network comprises the following steps: generating a rough key point outline of points through a three-layer fully-connected network and Reshape;
(b) refining dense point cloud generating network, namely splicing η×η points of 2D plane mesh of each point in the key points generated in the step (a)
Figure FDA0002526140890000031
Splicing the global description characteristics generated in the step s2 according to channels, and generating dense reconstruction point cloud through MLP;
wherein:
Figure FDA0002526140890000032
Figure FDA0002526140890000033
η is the number of dots on the edge of the planar grid and ξ is the size of the planar grid dots.
7. The method of claim 1, wherein in step s4, the encoder network is trained using a metric learning method, and a quadruple loss function is used to characterize the point cloud feature space
Figure FDA0002526140890000034
Semi-supervised training leads to input samples and positive samples pposBetween global descriptor featureposCloser, with negative example pnegBetween the global description feature of the Euclidean distanceSeparation devicenegFarther, extremely negative sample
Figure FDA00025261408900000314
For the zooming-in for preventing the error of the global description characteristic distance between the negative sample pairs, the Euclidean distance between the extreme negative sample and the negative sample is
Figure FDA00025261408900000313
Wherein:
Figure FDA0002526140890000035
α is a hyperparameter.
8. The method of claim 1, wherein in step s5, the decoder network is trained in a supervised manner using a loss function
Figure FDA0002526140890000036
Constraining Euclidean distance between the reconstructed point cloud and the truth point cloud, wherein
Figure FDA0002526140890000037
Applying the keypoint contours generated in step s2 (a) for constraining the physical spatial distribution of the point cloud,
Figure FDA0002526140890000038
applied to dense reconstructed point clouds;
wherein:
Figure FDA0002526140890000039
psi (p) is
Figure FDA00025261408900000310
The corresponding point of p in (1) is,
Figure FDA00025261408900000311
Figure FDA00025261408900000312
9. the method according to claim 1, wherein in steps s6 and s8, NN matching of global description features of point cloud key frames is determined by sequence matching of consecutive N current point cloud frames and consecutive N historical point cloud frames, and if euclidean distance of global description features between consecutive sequence matching key frames is smaller than a set threshold and gradient change compared with other historical key frame feature distances is larger than a set threshold, it is considered as a closed-loop detection candidate sequence key frame or a relocation candidate position; if a plurality of robots provide current continuous point cloud frames and historical key frames, establishing a map for the plurality of robots cooperatively; and performing a laser SLAM closed-loop task, a relocation task and a multi-robot cooperative task in the point cloud feature space by the point cloud feature space characterization method.
10. The method as claimed in claim 1, wherein in step s7, the encoder network of s4 is used to extract global description features from key frames in the laser point cloud map as compressed codes for storage or communication transmission, and the decoder network of s5 is used to restore the storage features or transmission reception features to reconstruct the original laser point cloud map.
CN202010504793.6A 2020-06-05 2020-06-05 Point cloud characteristic space representation method for laser SLAM Pending CN111783838A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010504793.6A CN111783838A (en) 2020-06-05 2020-06-05 Point cloud characteristic space representation method for laser SLAM

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010504793.6A CN111783838A (en) 2020-06-05 2020-06-05 Point cloud characteristic space representation method for laser SLAM

Publications (1)

Publication Number Publication Date
CN111783838A true CN111783838A (en) 2020-10-16

Family

ID=72754632

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010504793.6A Pending CN111783838A (en) 2020-06-05 2020-06-05 Point cloud characteristic space representation method for laser SLAM

Country Status (1)

Country Link
CN (1) CN111783838A (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112581515A (en) * 2020-11-13 2021-03-30 上海交通大学 Outdoor scene point cloud registration method based on graph neural network
CN113139996A (en) * 2021-05-06 2021-07-20 南京大学 Point cloud registration method and system based on three-dimensional point cloud geometric feature learning
CN113219493A (en) * 2021-04-26 2021-08-06 中山大学 End-to-end point cloud data compression method based on three-dimensional laser radar sensor
CN113343765A (en) * 2021-05-11 2021-09-03 武汉大学 Scene retrieval method and system based on point cloud rigid registration
CN113486963A (en) * 2021-07-12 2021-10-08 厦门大学 Density self-adaptive point cloud end-to-end sampling method
CN113538675A (en) * 2021-06-30 2021-10-22 同济人工智能研究院(苏州)有限公司 Neural network for calculating attention weight for laser point cloud and training method
CN113674416A (en) * 2021-08-26 2021-11-19 中国电子科技集团公司信息科学研究院 Three-dimensional map construction method and device, electronic equipment and storage medium
CN113807184A (en) * 2021-08-17 2021-12-17 北京百度网讯科技有限公司 Obstacle detection method and device, electronic equipment and automatic driving vehicle
CN113810736A (en) * 2021-08-26 2021-12-17 北京邮电大学 AI-driven real-time point cloud video transmission method and system
CN113808268A (en) * 2021-09-17 2021-12-17 浙江大学 Low-bandwidth globally consistent multi-machine dense graph building method
CN115578426A (en) * 2022-10-25 2023-01-06 哈尔滨工业大学 Indoor service robot repositioning method based on dense feature matching
CN116222577A (en) * 2023-04-27 2023-06-06 苏州浪潮智能科技有限公司 Closed loop detection method, training method, system, electronic equipment and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102592281A (en) * 2012-01-16 2012-07-18 北方工业大学 Image matching method
CN110243370A (en) * 2019-05-16 2019-09-17 西安理工大学 A kind of three-dimensional semantic map constructing method of the indoor environment based on deep learning

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102592281A (en) * 2012-01-16 2012-07-18 北方工业大学 Image matching method
CN110243370A (en) * 2019-05-16 2019-09-17 西安理工大学 A kind of three-dimensional semantic map constructing method of the indoor environment based on deep learning

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
CHUANZHE SUO等: "LPD-AE: Latent Space Representation of Large-Scale 3D Point Cloud", 《IEEE ACCESS》 *

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112581515A (en) * 2020-11-13 2021-03-30 上海交通大学 Outdoor scene point cloud registration method based on graph neural network
CN112581515B (en) * 2020-11-13 2022-12-13 上海交通大学 Outdoor scene point cloud registration method based on graph neural network
CN113219493A (en) * 2021-04-26 2021-08-06 中山大学 End-to-end point cloud data compression method based on three-dimensional laser radar sensor
CN113219493B (en) * 2021-04-26 2023-08-25 中山大学 End-to-end cloud data compression method based on three-dimensional laser radar sensor
CN113139996A (en) * 2021-05-06 2021-07-20 南京大学 Point cloud registration method and system based on three-dimensional point cloud geometric feature learning
CN113139996B (en) * 2021-05-06 2024-02-06 南京大学 Point cloud registration method and system based on three-dimensional point cloud geometric feature learning
CN113343765B (en) * 2021-05-11 2022-07-22 武汉大学 Scene retrieval method and system based on point cloud rigid registration
CN113343765A (en) * 2021-05-11 2021-09-03 武汉大学 Scene retrieval method and system based on point cloud rigid registration
CN113538675A (en) * 2021-06-30 2021-10-22 同济人工智能研究院(苏州)有限公司 Neural network for calculating attention weight for laser point cloud and training method
CN113486963B (en) * 2021-07-12 2023-07-07 厦门大学 Point cloud end-to-end sampling method with self-adaptive density
CN113486963A (en) * 2021-07-12 2021-10-08 厦门大学 Density self-adaptive point cloud end-to-end sampling method
CN113807184A (en) * 2021-08-17 2021-12-17 北京百度网讯科技有限公司 Obstacle detection method and device, electronic equipment and automatic driving vehicle
CN113810736A (en) * 2021-08-26 2021-12-17 北京邮电大学 AI-driven real-time point cloud video transmission method and system
CN113674416A (en) * 2021-08-26 2021-11-19 中国电子科技集团公司信息科学研究院 Three-dimensional map construction method and device, electronic equipment and storage medium
CN113674416B (en) * 2021-08-26 2024-04-26 中国电子科技集团公司信息科学研究院 Three-dimensional map construction method and device, electronic equipment and storage medium
CN113808268A (en) * 2021-09-17 2021-12-17 浙江大学 Low-bandwidth globally consistent multi-machine dense graph building method
CN113808268B (en) * 2021-09-17 2023-09-26 浙江大学 Low-bandwidth globally consistent multi-machine dense mapping method
CN115578426A (en) * 2022-10-25 2023-01-06 哈尔滨工业大学 Indoor service robot repositioning method based on dense feature matching
CN115578426B (en) * 2022-10-25 2023-08-18 哈尔滨工业大学 Indoor service robot repositioning method based on dense feature matching
CN116222577A (en) * 2023-04-27 2023-06-06 苏州浪潮智能科技有限公司 Closed loop detection method, training method, system, electronic equipment and storage medium

Similar Documents

Publication Publication Date Title
CN111783838A (en) Point cloud characteristic space representation method for laser SLAM
CN112529015B (en) Three-dimensional point cloud processing method, device and equipment based on geometric unwrapping
CN109377530B (en) Binocular depth estimation method based on depth neural network
CN110688905B (en) Three-dimensional object detection and tracking method based on key frame
CN108537837A (en) A kind of method and relevant apparatus of depth information determination
CN111462210B (en) Monocular line feature map construction method based on epipolar constraint
CN113705631B (en) 3D point cloud target detection method based on graph convolution
CN113191387A (en) Cultural relic fragment point cloud classification method combining unsupervised learning and data self-enhancement
CN111753789A (en) Robot vision SLAM closed loop detection method based on stack type combined self-encoder
CN110348299B (en) Method for recognizing three-dimensional object
CN112560865B (en) Semantic segmentation method for point cloud under outdoor large scene
CN110335299B (en) Monocular depth estimation system implementation method based on countermeasure network
CN111210382A (en) Image processing method, image processing device, computer equipment and storage medium
CN112329662B (en) Multi-view saliency estimation method based on unsupervised learning
CN113313176A (en) Point cloud analysis method based on dynamic graph convolution neural network
Fu et al. Pt-flownet: Scene flow estimation on point clouds with point transformer
CN115330935A (en) Three-dimensional reconstruction method and system based on deep learning
Deng et al. Neslam: Neural implicit mapping and self-supervised feature tracking with depth completion and denoising
CN115830375A (en) Point cloud classification method and device
CN117078753A (en) Progressive feature distribution sampling 6D pose estimation method and system based on camera
CN115578574A (en) Three-dimensional point cloud completion method based on deep learning and topology perception
Feng et al. Mesh reconstruction from aerial images for outdoor terrain mapping using joint 2d-3d learning
CN111860668B (en) Point cloud identification method for depth convolution network of original 3D point cloud processing
CN114155406A (en) Pose estimation method based on region-level feature fusion
KR20240115727A (en) Apparatus and method for performing visual localization

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20201016