CN115994933A - Partial point cloud registration method based on consistency learning - Google Patents

Partial point cloud registration method based on consistency learning Download PDF

Info

Publication number
CN115994933A
CN115994933A CN202310095301.6A CN202310095301A CN115994933A CN 115994933 A CN115994933 A CN 115994933A CN 202310095301 A CN202310095301 A CN 202310095301A CN 115994933 A CN115994933 A CN 115994933A
Authority
CN
China
Prior art keywords
point cloud
point
feature
corresponding relation
partial
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310095301.6A
Other languages
Chinese (zh)
Inventor
秦红星
谭博元
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing University of Post and Telecommunications
Original Assignee
Chongqing University of Post and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing University of Post and Telecommunications filed Critical Chongqing University of Post and Telecommunications
Priority to CN202310095301.6A priority Critical patent/CN115994933A/en
Publication of CN115994933A publication Critical patent/CN115994933A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/30Computing systems specially adapted for manufacturing

Landscapes

  • Image Analysis (AREA)

Abstract

The invention relates to a partial point cloud registration method based on consistency learning, which belongs to the field of computer vision and comprises the following steps: s1: collecting a source point cloud P, a target point cloud Q and real transformation from an actual scene, constructing a data set, preprocessing the data set, and dividing the data set into a training set and a testing set; s2: constructing a partial point cloud registration model, wherein the partial point cloud registration model comprises a point consistency learning module, a correspondence consistency learning module and an estimated rigidity transformation module; the point consistency learning module firstly extracts the local characteristics of each point, then estimates the corresponding overlapping score, and finally removes the points in the non-overlapping area according to the overlapping score of each point; s3: training the partial point cloud registration model by using a training set, and testing by using a testing set; s4: and initializing the source point cloud and the target point cloud to be registered, inputting a trained partial point cloud registration model, and outputting rigid transformation between the source point cloud and the target point cloud.

Description

Partial point cloud registration method based on consistency learning
Technical Field
The invention belongs to the field of computer vision, and relates to a partial point cloud registration method based on consistency learning.
Background
Point cloud registration is one of the fundamental tasks in the field of computer vision, which is widely used in the fields of simultaneous localization and mapping (SLAM), autopilot, 3D reconstruction, and the like. The goal of point cloud registration is to try to find a rigid transformation (rotation matrix and translation vector) between two 3D point clouds (source point cloud and target point cloud) to align them. Unfortunately, point cloud registration in partially overlapping scenes becomes more challenging due to the negative impact of non-overlapping areas of the point clouds.
In recent years, many approaches have been proposed to solve the point cloud registration problem. Iterative Closest Point (ICP) is one of the most classical algorithms, which iteratively performs the two steps of finding the closest point in another point cloud and solving the optimal transformation using SVD and updating the source point cloud state until the algorithm converges. However, ICP tends to fall into a local optimum and is sensitive to noise, limiting its application. For this reason, go-ICP, symmetric-ICP, generalized-ICP and the like have been proposed to address these challenges. In recent years, deep learning-based methods have made a great breakthrough. The depth nearest point (DCP) uses the nearest neighbor of the feature space instead of the nearest neighbor of the euclidean space. IDAM uses a convolutional neural network to directly predict the mapping matrix. Deep gmr uses a gaussian mixture model to estimate the optimal transformation. However, none of these methods assumes registration in a partially overlapping scene and therefore has poor performance. PRNet solves part-to-part point cloud registration by extracting keypoints from the point cloud and using them to construct a mapping matrix. RPMNet uses sink horn regularization to compute the mapping matrix. Deep bbs improves registration performance in partially overlapping scenarios by using soft mutual neighbor constraints.
Most of the current methods based on deep learning divide point cloud registration into three parts: extracting the characteristics of the points, constructing corresponding relations based on characteristic matching, and estimating transformation parameters. However, the negative impact of non-overlapping regions in partially overlapping scenes can result in an inability to learn effective feature space.
Most of the current methods cannot effectively identify and remove the wrong correspondence. When there is an excessive proportion of wrong correspondence, the estimated transformation parameters and the actual transformation are excessively different, and the registration is failed.
Disclosure of Invention
In view of the above, the present invention aims to provide a partial point cloud registration method based on consistency learning, which predicts an overlap score for each point by using a consistency constraint of a point hierarchy according to a negative effect of a non-overlapping region in a partially overlapped scene, and removes the effect of the non-overlapping region point according to the overlap score; aiming at the problem of excessive error corresponding relations, the error corresponding relations are identified by exploring local to global context information between the corresponding relations, and are removed.
In order to achieve the above purpose, the present invention provides the following technical solutions:
a partial point cloud registration method based on consistency learning comprises the following steps:
s1: collecting a source point cloud P, a target point cloud Q and real transformation from an actual scene, constructing a data set, preprocessing the data set, and dividing the data set into a training set and a testing set;
s2: constructing a partial point cloud registration model, wherein the partial point cloud registration model comprises a point consistency learning module, a correspondence consistency learning module and an estimated rigidity transformation module; the point consistency learning module firstly extracts the local characteristics of each point, then estimates the corresponding overlapping score, and finally removes the points in the non-overlapping area according to the overlapping score of each point;
s3: training the partial point cloud registration model by using a training set, and testing by using a testing set;
s4: and initializing the source point cloud and the target point cloud to be registered, inputting a trained partial point cloud registration model, and outputting rigid transformation between the source point cloud and the target point cloud.
Further, preprocessing the data set, including removing outliers in the point cloud, voxel downsampling, and regularization.
Further, the specific calculation steps of the point consistency learning module are as follows:
a1: using a graph neural network from a point p in a source point cloud i E P and Point q in the target Point cloud j Extracting local characteristics of points from the E Q;
a2: predicting the overlapping score of each point;
a3: points with overlapping scores less than 70% of the point cloud size are considered as points of a non-overlapping region, and the points of the non-overlapping region are removed.
Further, the step A1 specifically includes the following steps:
a11: for each point p i E P, firstly, performing k neighbor search to obtain k neighbor point set thereof
Figure BDA0004071476200000021
A12: each point p i The difference between the 3D coordinates of e P and the 3D coordinates of its neighborhood points constitutes the initial input feature:
Figure BDA0004071476200000022
a13: using a neural network f g Will be
Figure BDA0004071476200000023
Mapping to a high-dimensional feature space, the resulting feature is denoted +.>
Figure BDA0004071476200000024
Figure BDA0004071476200000025
Wherein, the graph neural network f g (. Cndot.) the interior is divided into four layers, each expressed as:
Figure BDA0004071476200000026
in the method, in the process of the invention,
Figure BDA0004071476200000027
is an output feature and is also an input feature of the next layer; MLP is a multi-layer perceptron; maxPooling represents a max pooling operation; />
Figure BDA0004071476200000028
Is->
Figure BDA0004071476200000029
K neighbor sets of (a); []Representing stitching along the channel direction of the feature;
a14: the same method is adopted for the point q in the target point cloud j Performing feature extraction on the E Q, wherein the obtained features are as follows
Figure BDA00040714762000000210
Figure BDA00040714762000000211
Wherein->
Figure BDA00040714762000000212
Forming an initial input characteristic for the difference between the 3D coordinates of each point in the target point cloud Q and the 3D coordinates of the neighborhood points;
a15: local feature point-to-point with self-attention and cross-attention modules
Figure BDA0004071476200000031
And->
Figure BDA0004071476200000032
Performing characteristic enhancement, and marking the enhanced characteristics as +.>
Figure BDA0004071476200000033
And->
Figure BDA0004071476200000034
Further, the step A2 specifically includes the following steps:
a21: for the extracted
Figure BDA0004071476200000035
And->
Figure BDA0004071476200000036
Applying a max-pooling operation to obtain global features G of an entire point cloud P =maxpooling(C P ) And G Q =maxpooling(C Q );
A22: the global characteristics of two point clouds and the characteristics of each point are spliced and input into a multi-layer perceptron to predict the overlapping score of each point
Figure BDA0004071476200000037
And
Figure BDA0004071476200000038
Figure BDA0004071476200000039
wherein MLP is a multi-layer perceptron, the output dimension is N multiplied by 1, N is the number of points in the point cloud; delta (·) is the repeat operation along the characteristic channel direction; [ x, y, z]Representing stitching of features x, y and z in the channel dimension; o (O) P And O Q Representing overlapping sets of scores for point cloud P and point cloud Q, respectively.
Further, the specific calculation steps of the correspondence consistency learning module are as follows:
b1: defining a mapping matrix M:
Figure BDA00040714762000000310
Figure BDA00040714762000000311
wherein M is ij For the elements of the ith row and jth column in the mapping matrix M;
Figure BDA00040714762000000312
and->
Figure BDA00040714762000000313
Respectively point p i Sum point q j The characteristics after strengthening by the self-attention and cross-attention modules; />
Figure BDA00040714762000000314
Representing warpThe inner product of the eigenvector amplified by the exponential function can be regarded as an intermediate variable of the calculation.
For one point P in the source point cloud P r Its corresponding point in the target point cloud Q is
Figure BDA00040714762000000315
Thereby obtaining a set of initial corresponding relations
Figure BDA00040714762000000316
B2: searching local to global context information between the corresponding relations to identify the wrong corresponding relation;
b3: and removing the wrong corresponding relation.
Further, the step B2 specifically includes:
b21: first, an initial corresponding relation feature set H is calculated C
Figure BDA00040714762000000317
Figure BDA00040714762000000318
In the method, in the process of the invention,
Figure BDA00040714762000000319
representing the correspondence c r =(p r ,q r ) Initial characteristics of e C; [ (i. ], i]The method comprises the steps of representing that features are spliced in a channel dimension; />
Figure BDA00040714762000000320
Correspondence feature representing distance perception +.>
Figure BDA00040714762000000321
Representation->
Figure BDA00040714762000000322
Is the j-th neighbor,/, of>
Figure BDA00040714762000000323
Is->
Figure BDA00040714762000000324
Is the modulo length of the vector;
b22: inputting the initial corresponding relation characteristic into a multi-scale attention module to explore local to global context information; the multi-scale attention module comprises three layers of identical operations, each layer operating as follows:
features to be characterized
Figure BDA0004071476200000041
Input to a neural network to explore local consistency, and output is characterized as Y r (1)
Will local feature Y r (1) Input into an inner product attention module, output feature F r (1) Then calculate the offset feature OF r (1) =Y r (1) -F r (1) Then the offset feature OF r (1) Inputting into a multi-layer perceptron to obtain
Figure BDA0004071476200000042
Figure BDA0004071476200000043
The final characteristic is->
Figure BDA0004071476200000044
B23: the three-layer characteristics are maximally pooled, the obtained characteristics are all spliced together and input into a multi-layer perceptron to predict the weight W=MLP ([ F) of each corresponding relation (1) ,F (2) ,F (3) ]) The method comprises the steps of carrying out a first treatment on the surface of the Wherein MLP is a multi-layer perceptron which ultimately outputsThe dimension is |C| multiplied by 1, and |C| is the size of the corresponding relation set C.
In step B3, whether a certain corresponding relationship is to be reserved is determined according to the weight W, and the final corresponding relationship set is recorded as
Figure BDA0004071476200000045
The corresponding weight is->
Figure BDA0004071476200000046
Further, the calculation steps of the estimated rigid transformation module are as follows:
according to the obtained corresponding relation set
Figure BDA0004071476200000047
Weight +.>
Figure BDA0004071476200000048
The final transformation is obtained by solving the following objective function: />
Figure BDA0004071476200000049
Wherein (R) e ,t e ) For the final solved transformation parameters, where R e For rotating matrix, t e Is a translation vector;
Figure BDA00040714762000000410
is of corresponding relationship (p) i ,q i ) Corresponding weights;
the objective function is calculated by a weighted singular value decomposition method.
Further, in the step S3, when the partial point cloud registration model is trained by using a training set, the adopted loss function is divided into three parts: mapping matrix loss, point prediction loss and corresponding relation prediction loss, wherein the point prediction loss and the corresponding relation prediction loss are binary cross entropy loss;
mapping matrix loss L 1 The method comprises the following steps:
Figure BDA00040714762000000411
j * =argmin j ||R * p i +t * -q j || 2
wherein M is a mapping matrix; a, a i Is a parameter that determines whether the current item is calculated, if it is satisfied
Figure BDA00040714762000000412
Then a i =1; otherwise it is a i =0;R * And t * The real rotation matrix and the translation vector are respectively; p is p i And q j One point of the source point cloud P and the target point cloud Q respectively; j (j) * Representing point p i The number of the true corresponding point in the target point cloud Q;
point prediction loss L 21 And L 22 The method comprises the following steps:
Figure BDA0004071476200000051
Figure BDA0004071476200000052
wherein N is the size of the point cloud;
Figure BDA0004071476200000053
and->
Figure BDA0004071476200000054
Respectively the predicted point p i Sum point q j Is a superposition fraction of (2); />
Figure BDA0004071476200000055
For point p i True overlap score, if p i In the overlapping region->
Figure BDA0004071476200000056
Otherwise->
Figure BDA0004071476200000057
For point q i True overlap score, if q i In the overlapping region->
Figure BDA0004071476200000058
Otherwise->
Figure BDA0004071476200000059
Corresponding relation prediction loss L 3 The method comprises the following steps:
Figure BDA00040714762000000510
wherein w is i E W is the weight of the predicted ith correspondence,
Figure BDA00040714762000000511
is the weight of the real i-th corresponding relation;
the final loss is
L=L 1 +L 21 +L 22 +L 3
Where L represents the final loss.
The invention has the beneficial effects that: in the point consistency learning stage, the invention radically solves the negative influence of a non-overlapping area by predicting the overlapping area of the source point cloud and the target point cloud, converts partial-to-partial point cloud registration into approximately complete-to-complete point cloud registration, and greatly reduces task difficulty. In the correspondence consistency learning stage, the method acquires the initial correspondence through the similarity between the features, and then identifies the mismatching through exploring the local-global context information between the initial correspondences, thereby realizing accurate point cloud registration.
Additional advantages, objects, and features of the invention will be set forth in part in the description which follows and in part will become apparent to those having ordinary skill in the art upon examination of the following or may be learned from practice of the invention. The objects and other advantages of the invention may be realized and obtained by means of the instrumentalities and combinations particularly pointed out in the specification.
Drawings
For the purpose of making the objects, technical solutions and advantages of the present invention more apparent, the present invention will be described in the following preferred detail with reference to the accompanying drawings, in which:
FIG. 1 is a flow chart of a partial point cloud registration method based on consistency learning;
fig. 2 is a registration result diagram.
Detailed Description
Other advantages and effects of the present invention will become apparent to those skilled in the art from the following disclosure, which describes the embodiments of the present invention with reference to specific examples. The invention may be practiced or carried out in other embodiments that depart from the specific details, and the details of the present description may be modified or varied from the spirit and scope of the present invention. It should be noted that the illustrations provided in the following embodiments merely illustrate the basic idea of the present invention by way of illustration, and the following embodiments and features in the embodiments may be combined with each other without conflict.
Wherein the drawings are for illustrative purposes only and are shown in schematic, non-physical, and not intended to limit the invention; for the purpose of better illustrating embodiments of the invention, certain elements of the drawings may be omitted, enlarged or reduced and do not represent the size of the actual product; it will be appreciated by those skilled in the art that certain well-known structures in the drawings and descriptions thereof may be omitted.
The same or similar reference numbers in the drawings of embodiments of the invention correspond to the same or similar components; in the description of the present invention, it should be understood that, if there are terms such as "upper", "lower", "left", "right", "front", "rear", etc., that indicate an azimuth or a positional relationship based on the azimuth or the positional relationship shown in the drawings, it is only for convenience of describing the present invention and simplifying the description, but not for indicating or suggesting that the referred device or element must have a specific azimuth, be constructed and operated in a specific azimuth, so that the terms describing the positional relationship in the drawings are merely for exemplary illustration and should not be construed as limiting the present invention, and that the specific meaning of the above terms may be understood by those of ordinary skill in the art according to the specific circumstances.
As shown in fig. 1, the present invention provides a partial point cloud registration method based on consistency learning, wherein given a source point cloud P and a target point cloud Q, rigid transformations R and t between the two point clouds are found through three stages: point consistency learning, correspondence consistency learning, and estimation rigidity transformation R, t.
Point consistency learning phase:
the point consistency learning first extracts the local features of each point, then estimates the corresponding overlap score, and finally removes the points of the non-overlapping region according to the overlap score of each point.
First from p using a graph neural network (Graph Neural Network) i E P and q j The local characteristics of the points are extracted from the E Q, and the characteristics can capture the local information of the points and pay attention to the local fine granularity structure in the neighborhood range. Specifically, for each point p i E P, firstly, performing k neighbor search to obtain k neighbor point set thereof
Figure BDA0004071476200000061
The difference between the 3D coordinates of each point and the 3D coordinates of its neighboring points is then used to form the initial input feature +.>
Figure BDA0004071476200000062
Then use a neural network f g Will->
Figure BDA0004071476200000063
Mapping to a high-dimensional feature space, the resulting feature is denoted +.>
Figure BDA0004071476200000064
Point q in a target point cloud j E Q also uses similar operations for feature extraction, the resulting feature is +.>
Figure BDA0004071476200000065
Local features of dots->
Figure BDA0004071476200000066
And->
Figure BDA0004071476200000067
Similarity between them expresses the point p i And q j Is a local neighborhood of the object. However, in order to better predict the overlapping score of each point and then construct a reliable correspondence, it is necessary to strengthen the extracted local features. Thus, we use the self-attention and cross-attention modules for feature enhancement. The reinforced features are respectively marked as +.>
Figure BDA0004071476200000068
And->
Figure BDA0004071476200000069
To predict the overlap score of each point, i.e., the probability of each point being in the overlap region, for the extracted
Figure BDA0004071476200000071
And
Figure BDA0004071476200000072
applying a max-pooling operation to obtain global features G of an entire point cloud P =maxpooling(C P ) And G Q =maxpooling(C Q ). Then the global features of the two point clouds and the features of each point are spliced and input into a multi-layer perceptron (MLP) to predict the overlapping score of each point +.>
Figure BDA0004071476200000073
And->
Figure BDA0004071476200000074
Figure BDA0004071476200000075
Points with overlapping scores less than 70% of the point cloud size are regarded as points of the non-overlapping area, and finally the points are removed, and the rest points participate in subsequent operations.
Correspondence consistency learning stage:
through the point consistency learning phase, points located in non-overlapping areas have been removed. The rest point clouds are still marked as P and Q, and the corresponding characteristics are still marked as C P And C Q . In order to obtain the mapping relationship between two point clouds, the inner product of the feature vectors is used to measure the similarity of the features. Defining a mapping matrix M, wherein
Figure BDA0004071476200000076
For one point P in the source point cloud P r Its corresponding point in the target point cloud Q is +.>
Figure BDA0004071476200000077
Thus an initial set of correspondences is obtained>
Figure BDA0004071476200000078
However, there is some mismatch in the initial correspondence C obtained, which would seriously affect the final registration result if it could not be removed. Therefore, the invention identifies the wrong corresponding relation by exploring the local to global context information between the corresponding relations, then removes the wrong corresponding relation, and finally solves the rigid transformation by using the residual correct corresponding relation. For each corresponding relation, first calculating initial corresponding relation characteristic
Figure BDA0004071476200000079
Figure BDA00040714762000000710
The feature is then input into aLocal to global context information is explored in a plurality of multi-scale attention modules. The multi-scale attention module has three layers, and since the operations between the different layers are identical, we will only describe the operation of the first layer in the following. Features to be characterized
Figure BDA00040714762000000711
Input to a neural network to explore local consistency, and output is characterized as Y r (1) . Thereafter Y is taken r (1) Input to an offset non-local (offsetnon) module explores global consistency. In particular, local feature Y r (1) Input into an inner product attention module, output feature F r (1) . Then calculate the offset feature OF r (1) =Y r (1) -F r (1) Then the offset feature OF r (1) Inputting into a multi-layer perceptron to obtain +.>
Figure BDA00040714762000000712
The final characteristics are that
Figure BDA00040714762000000713
Finally, we apply maximum pooling to the three-layer features, splice the features together, input them into a multi-layer perceptron to predict the weights w=mlps ([ F) for each correspondence (1) ,F (2) ,F (3) ]). And determining whether a certain corresponding relation is to be reserved according to the weight W. In the implementation, the first 50% of corresponding relation is reserved for subsequent solving transformation according to the size of the weight W, and the final corresponding relation set is recorded as +.>
Figure BDA00040714762000000714
The corresponding weight is->
Figure BDA00040714762000000715
Estimating a rigid transformation:
according to the obtained corresponding relation set
Figure BDA00040714762000000716
Weight +.>
Figure BDA00040714762000000717
The final transformation may be obtained by solving the following objective function: />
Figure BDA0004071476200000081
The objective function may be solved by Weighted singular value decomposition (Weighted SVD).
Training part:
training data, namely source point cloud, target point cloud and real transformation are prepared according to an actual scene: if there are a large number of outliers in the point cloud, an operation of removing outliers is required to improve the quality of the point cloud, and using a radius outlier removal (remote_radius_outlier) function of the Open3D tool, points less than 30 points in a given sphere within a radius of 0.05m are deleted. Voxel downsampling (voxel down sample) was performed using Open3D with a voxel size set to 0.025. Then randomly sampling 1024 points from the downsampled source point cloud and target point cloud respectively, and marking the sampled points as P epsilon R 1024×3 And Q.epsilon.R 1024×3 . Finally regularizing P and Q:
Figure BDA0004071476200000082
in 3D space, a random direction transformation T (comprising 0 ° to 360 ° rotations and translations within plus or minus 0.05 unit distances) is generated, which is applied to the target point cloud Q.
Inputting the point clouds P and Q into the partial point cloud registration model of the invention, recording the generated mapping matrix M and overlapping the score O P And O Q And a weight W of the correspondence. The loss value is calculated from these data and the actual data, and the update parameters are back-propagated. The above description is of the operation performed for one epoch.
The total epoch in the training part is set to be 1000, the optimizer used is Adam, and the initial learning rate is 0.0001,batch size and is set to be 8. The loss function used is divided into three parts: mapping matrix loss, point prediction loss and correspondence prediction loss. Wherein the point prediction loss and the correspondence prediction loss are both binary cross entropy losses.
(1) Mapping matrix loss L 1 The method comprises the following steps:
Figure BDA0004071476200000083
j * =argmin j ||R * p i +t * -q j || 2
wherein M is a mapping matrix; a, a i Is a parameter that determines whether the current item is calculated, if it is satisfied
Figure BDA0004071476200000084
Then a i =1; otherwise it is a i =0;R * And t * The real rotation matrix and the translation vector are respectively; p is p i And q j One point in the source point cloud P and the target point cloud Q, respectively. j (j) * Representing point p i And the number of the real corresponding point in the target point cloud Q.
(2) Point prediction loss L 21 And L 22 The method comprises the following steps:
Figure BDA0004071476200000085
Figure BDA0004071476200000086
wherein N is the size of the point cloud;
Figure BDA0004071476200000087
and->
Figure BDA0004071476200000088
Respectively the predicted point p i Sum point q j Is a superposition fraction of (2); />
Figure BDA0004071476200000089
For point p i True overlap score, if p i In the overlapping region->
Figure BDA0004071476200000091
Otherwise->
Figure BDA0004071476200000092
For point q i True overlap score, if q i In the overlapping region->
Figure BDA0004071476200000093
Otherwise->
Figure BDA0004071476200000094
/>
(3) Corresponding relation prediction loss L 3 The method comprises the following steps:
Figure BDA0004071476200000095
wherein w is i E W is the weight of the predicted ith correspondence,
Figure BDA0004071476200000096
is the weight of the true i-th correspondence.
The final loss L is
L=L 1 +L 21 +L 22 +L 3
Testing and practical application:
1. preparing a source point cloud and a target point cloud to be registered, and preprocessing the source point cloud and the target point cloud: outliers are removed, voxel downsampling and regularization.
2. Loading a trained model for estimating the transformation, and recording the model as f θ Where θ represents the pre-training parameters.
3. Inputting the preprocessed point clouds P and Q into a trained model to solve the transformation (R, t) =f θ (P,Q)。
4. The last step finds the transformation (R, t) of the regularized space, and therefore requires mapping back to the non-normalized original space to get the final result:
Figure BDA0004071476200000097
Figure BDA0004071476200000098
as shown in fig. 2, the registration result chart of the present embodiment corresponds to the results of unseen shapes, unseen categories and unseen shapes with noise, respectively, in the first three rows. Inputs are the input source point cloud and target point cloud. GT is the registration result of group-trunk.
Finally, it is noted that the above embodiments are only for illustrating the technical solution of the present invention and not for limiting the same, and although the present invention has been described in detail with reference to the preferred embodiments, it should be understood by those skilled in the art that modifications and equivalents may be made thereto without departing from the spirit and scope of the present invention, which is intended to be covered by the claims of the present invention.

Claims (10)

1. A partial point cloud registration method based on consistency learning is characterized by comprising the following steps: the method comprises the following steps:
s1: collecting a source point cloud P, a target point cloud Q and real transformation from an actual scene, constructing a data set, preprocessing the data set, and dividing the data set into a training set and a testing set;
s2: constructing a partial point cloud registration model, wherein the partial point cloud registration model comprises a point consistency learning module, a correspondence consistency learning module and an estimated rigidity transformation module; the point consistency learning module firstly extracts the local characteristics of each point, then estimates the corresponding overlapping score, and finally removes the points in the non-overlapping area according to the overlapping score of each point;
s3: training the partial point cloud registration model by using a training set, and testing by using a testing set;
s4: and initializing the source point cloud and the target point cloud to be registered, inputting a trained partial point cloud registration model, and outputting rigid transformation between the source point cloud and the target point cloud.
2. The consistency learning-based partial point cloud registration method as claimed in claim 1, wherein: preprocessing the data set, including removing outliers in the point cloud, performing voxel downsampling and regularization.
3. The consistency learning-based partial point cloud registration method as claimed in claim 1, wherein: the specific calculation steps of the point consistency learning module are as follows:
a1: using a graph neural network from a point p in a source point cloud i E P and Point q in the target Point cloud j Extracting local characteristics of points from the E Q;
a2: predicting the overlapping score of each point;
a3: points with overlapping scores less than 70% of the point cloud size are considered as points of a non-overlapping region, and the points of the non-overlapping region are removed.
4. The consistency learning-based partial point cloud registration method as claimed in claim 3, wherein: the step A1 specifically comprises the following steps:
a11: for each point p i E P, firstly, performing k neighbor search to obtain k neighbor point set thereof
Figure FDA0004071476190000011
A12: each point p i The difference between the 3D coordinates of e P and the 3D coordinates of its neighborhood points constitutes the initial input feature:
Figure FDA0004071476190000012
A13:using a neural network f g Will be
Figure FDA00040714761900000112
Mapping to a high-dimensional feature space, the resulting feature is denoted +.>
Figure FDA0004071476190000013
Figure FDA0004071476190000014
Wherein, the graph neural network f g (. Cndot.) the interior is divided into four layers, each expressed as:
Figure FDA0004071476190000015
in the method, in the process of the invention,
Figure FDA0004071476190000016
is an output feature and is also an input feature of the next layer; MLP is a multi-layer perceptron; maxPooling represents a max pooling operation; />
Figure FDA0004071476190000017
Is->
Figure FDA0004071476190000018
K neighbor sets of (a); []Representing stitching along the channel direction of the feature;
a14: the same method is adopted for the point q in the target point cloud j Performing feature extraction on the E Q, wherein the obtained features are as follows
Figure FDA0004071476190000019
Figure FDA00040714761900000110
Wherein->
Figure FDA00040714761900000111
Forming an initial input characteristic for the difference between the 3D coordinates of each point in the target point cloud Q and the 3D coordinates of the neighborhood points;
a15: local feature point-to-point with self-attention and cross-attention modules
Figure FDA0004071476190000021
And->
Figure FDA0004071476190000022
Performing characteristic enhancement, and marking the enhanced characteristics as +.>
Figure FDA0004071476190000023
And->
Figure FDA0004071476190000024
5. The consistency learning-based partial point cloud registration method as claimed in claim 3, wherein: the step A2 specifically comprises the following steps:
a21: for the extracted
Figure FDA0004071476190000025
And->
Figure FDA0004071476190000026
Applying a max-pooling operation to obtain global features G of an entire point cloud P =maxpooling(C P ) And G Q =maxpooling(C Q );
A22: the global characteristics of two point clouds and the characteristics of each point are spliced and input into a multi-layer perceptron to predict the overlapping score of each point
Figure FDA0004071476190000027
And
Figure FDA0004071476190000028
Figure FDA0004071476190000029
wherein MLP is a multi-layer perceptron, the output dimension is N multiplied by 1, N is the number of points in the point cloud; delta (·) is the repeat operation along the characteristic channel direction; [ x, y, z]Representing stitching of features x, y and z in the channel dimension; o (O) P And O Q Representing overlapping sets of scores for point cloud P and point cloud Q, respectively.
6. The consistency learning-based partial point cloud registration method as claimed in claim 1, wherein: the specific calculation steps of the correspondence consistency learning module are as follows:
b1: defining a mapping matrix M:
Figure FDA00040714761900000210
Figure FDA00040714761900000211
wherein M is ij For the elements of the ith row and jth column in the mapping matrix M;
Figure FDA00040714761900000212
and->
Figure FDA00040714761900000213
Respectively point p i Sum point q j The characteristics after strengthening by the self-attention and cross-attention modules; />
Figure FDA00040714761900000214
Representing the inner product of the eigenvectors amplified by the exponential function, taken as the calculation M ij Is an intermediate variable of (a);
for one point P in the source point cloud P r Its corresponding point in the target point cloud Q is
Figure FDA00040714761900000215
Thereby obtaining a set of initial corresponding relations
Figure FDA00040714761900000216
B2: searching local to global context information between the corresponding relations to identify the wrong corresponding relation;
b3: and removing the wrong corresponding relation.
7. The consistency learning-based partial point cloud registration method as claimed in claim 6, wherein: the step B2 specifically includes:
b21: first, an initial corresponding relation feature set H is calculated C
Figure FDA00040714761900000217
Figure FDA0004071476190000031
In the method, in the process of the invention,
Figure FDA0004071476190000032
representing the correspondence c r =(p r ,q r ) Initial characteristics of e C; [ (i. ], i]The method comprises the steps of representing that features are spliced in a channel dimension; />
Figure FDA0004071476190000033
Correspondence feature representing distance perception +.>
Figure FDA0004071476190000034
Representation->
Figure FDA0004071476190000035
Is the j-th neighbor,/, of>
Figure FDA0004071476190000036
Is that
Figure FDA0004071476190000037
Is the modulo length of the vector;
b22: inputting the initial corresponding relation characteristic into a multi-scale attention module to explore local to global context information; the multi-scale attention module comprises three layers of identical operations, each layer operating as follows:
features to be characterized
Figure FDA0004071476190000038
Input to a neural network to explore local consistency, and output is characterized as Y r (1)
Will local feature Y r (1) Input into an inner product attention module, output feature F r (1) Then calculate the offset feature OF r (1) =Y r (1) -F r (1) Then the offset feature OF r (1) Inputting into a multi-layer perceptron to obtain
Figure FDA0004071476190000039
Figure FDA00040714761900000310
The final characteristic is->
Figure FDA00040714761900000311
/>
B23: maximizing pooling of three layers of features, stitching all obtained features together and inputting the features into a multi-layer perceptionWeight w=mlp ([ F) for predicting each correspondence in the machine (1) ,F (2) ,F (3) ]) The method comprises the steps of carrying out a first treatment on the surface of the The MLP is a multi-layer perceptron, the dimension of the final output of the MLP is |C|x 1, and |C| is the size of the corresponding relation set C.
8. The consistency learning-based partial point cloud registration method as claimed in claim 6, wherein: in the step B3, it is determined whether a certain corresponding relation is to be reserved according to the weight W, and the final corresponding relation set is recorded as
Figure FDA00040714761900000312
The corresponding weight is->
Figure FDA00040714761900000313
9. The consistency learning-based partial point cloud registration method as claimed in claim 1, wherein: the calculation steps of the estimated rigid transformation module are as follows:
according to the obtained corresponding relation set
Figure FDA00040714761900000314
Weight +.>
Figure FDA00040714761900000315
The final transformation is obtained by solving the following objective function:
Figure FDA00040714761900000316
wherein (R) e ,t e ) For the final solved transformation parameters, where R e For rotating matrix, t e Is a translation vector;
Figure FDA00040714761900000317
is of corresponding relationship (p) i ,q i ) Corresponding weights;
the objective function is calculated by a weighted singular value decomposition method.
10. The consistency learning-based partial point cloud registration method as claimed in claim 1, wherein: in the step S3, when the training set is used to train the partial point cloud registration model, the adopted loss function is divided into three parts: mapping matrix loss, point prediction loss and corresponding relation prediction loss, wherein the point prediction loss and the corresponding relation prediction loss are binary cross entropy loss;
mapping matrix loss L 1 The method comprises the following steps:
Figure FDA0004071476190000041
j * =argmin j ||R * p i +t * -M|| 2
wherein M is a mapping matrix; a, a i Is a parameter that determines whether the current item is calculated, if it is satisfied
Figure FDA0004071476190000042
Then a i =1; otherwise it is a i =0;R * And t * The real rotation matrix and the translation vector are respectively; p is p i And q j One point of the source point cloud P and the target point cloud Q respectively; j (j) * Representing point p i The number of the true corresponding point in the target point cloud Q;
point prediction loss L 21 And L 22 The method comprises the following steps:
Figure FDA0004071476190000043
Figure FDA0004071476190000044
wherein N is the size of the point cloud;
Figure FDA0004071476190000045
and->
Figure FDA0004071476190000046
Respectively the predicted point p i Sum point q j Is a superposition fraction of (2); />
Figure FDA0004071476190000047
For point p i True overlap score, if p i In the overlapping region->
Figure FDA0004071476190000048
Otherwise->
Figure FDA0004071476190000049
Figure FDA00040714761900000410
For point q i True overlap score, if q i In the overlapping region->
Figure FDA00040714761900000411
Otherwise->
Figure FDA00040714761900000412
Corresponding relation prediction loss L 3 The method comprises the following steps:
Figure FDA00040714761900000413
wherein w is i E W is the weight of the predicted ith correspondence,
Figure FDA00040714761900000414
is the weight of the real i-th corresponding relation;
the final loss is
L=L 1 +L 21 +L 22 +L 3
Where L represents the final loss.
CN202310095301.6A 2023-02-10 2023-02-10 Partial point cloud registration method based on consistency learning Pending CN115994933A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310095301.6A CN115994933A (en) 2023-02-10 2023-02-10 Partial point cloud registration method based on consistency learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310095301.6A CN115994933A (en) 2023-02-10 2023-02-10 Partial point cloud registration method based on consistency learning

Publications (1)

Publication Number Publication Date
CN115994933A true CN115994933A (en) 2023-04-21

Family

ID=85991840

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310095301.6A Pending CN115994933A (en) 2023-02-10 2023-02-10 Partial point cloud registration method based on consistency learning

Country Status (1)

Country Link
CN (1) CN115994933A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116523979A (en) * 2023-04-24 2023-08-01 北京长木谷医疗科技股份有限公司 Point cloud registration method and device based on deep learning and electronic equipment

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116523979A (en) * 2023-04-24 2023-08-01 北京长木谷医疗科技股份有限公司 Point cloud registration method and device based on deep learning and electronic equipment
CN116523979B (en) * 2023-04-24 2024-01-30 北京长木谷医疗科技股份有限公司 Point cloud registration method and device based on deep learning and electronic equipment

Similar Documents

Publication Publication Date Title
Zhao et al. Knowledge as priors: Cross-modal knowledge generalization for datasets without superior knowledge
US11670071B2 (en) Fine-grained image recognition
Laskar et al. Camera relocalization by computing pairwise relative poses using convolutional neural network
CN109977757B (en) Multi-modal head posture estimation method based on mixed depth regression network
CN112288011B (en) Image matching method based on self-attention deep neural network
CN106991388B (en) Key point positioning method
WO2019178702A1 (en) Systems and methods for polygon object annotation and a method of training an object annotation system
Heinrich et al. Highly accurate and memory efficient unsupervised learning-based discrete CT registration using 2.5 D displacement search
Lin et al. Visual object recognition and pose estimation based on a deep semantic segmentation network
CN113705769A (en) Neural network training method and device
CN111368759B (en) Monocular vision-based mobile robot semantic map construction system
CN109840518B (en) Visual tracking method combining classification and domain adaptation
CN114445461A (en) Visible light infrared target tracking training method and device based on non-paired data
CN111414875A (en) Three-dimensional point cloud head attitude estimation system based on depth regression forest
CN115994933A (en) Partial point cloud registration method based on consistency learning
CN107798329B (en) CNN-based adaptive particle filter target tracking method
Kowdiki et al. Adaptive hough transform with optimized deep learning followed by dynamic time warping for hand gesture recognition
Hutchcroft et al. CoVisPose: Co-visibility pose transformer for wide-baseline relative pose estimation in 360∘ indoor panoramas
Kim et al. Self-supervised keypoint detection based on multi-layer random forest regressor
Kang et al. Yolo-6d+: single shot 6d pose estimation using privileged silhouette information
Felton et al. Deep metric learning for visual servoing: when pose and image meet in latent space
WO2022236647A1 (en) Methods, devices, and computer readable media for training a keypoint estimation network using cgan-based data augmentation
CN114066844A (en) Pneumonia X-ray image analysis model and method based on attention superposition and feature fusion
Slimani et al. RoCNet++: Triangle-based descriptor for accurate and robust point cloud registration
CN113139991A (en) 3D point cloud registration method based on overlapping region mask prediction

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination